var/home/core/zuul-output/0000755000175000017500000000000015067353520014533 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015067401035015472 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006545373115067401024017712 0ustar rootrootOct 02 01:41:01 crc systemd[1]: Starting Kubernetes Kubelet... Oct 02 01:41:01 crc restorecon[4726]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:01 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:41:02 crc restorecon[4726]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 02 01:41:02 crc restorecon[4726]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 02 01:41:03 crc kubenswrapper[4775]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 01:41:03 crc kubenswrapper[4775]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 02 01:41:03 crc kubenswrapper[4775]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 01:41:03 crc kubenswrapper[4775]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 01:41:03 crc kubenswrapper[4775]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 02 01:41:03 crc kubenswrapper[4775]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.489245 4775 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493757 4775 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493785 4775 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493793 4775 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493801 4775 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493809 4775 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493816 4775 feature_gate.go:330] unrecognized feature gate: Example Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493822 4775 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493831 4775 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493840 4775 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493847 4775 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493854 4775 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493861 4775 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493868 4775 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493875 4775 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493881 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493888 4775 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493894 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493904 4775 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493911 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493917 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493923 4775 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493929 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493935 4775 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493941 4775 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493948 4775 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493986 4775 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.493994 4775 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494000 4775 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494006 4775 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494013 4775 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494019 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494026 4775 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494032 4775 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494040 4775 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494046 4775 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494055 4775 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494061 4775 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494070 4775 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494079 4775 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494086 4775 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494095 4775 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494104 4775 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494111 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494117 4775 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494124 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494131 4775 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494138 4775 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494146 4775 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494153 4775 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494160 4775 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494166 4775 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494173 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494181 4775 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494187 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494193 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494199 4775 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494206 4775 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494212 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494218 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494224 4775 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494230 4775 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494236 4775 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494243 4775 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494249 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494255 4775 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494261 4775 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494271 4775 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494279 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494287 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494294 4775 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.494301 4775 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495196 4775 flags.go:64] FLAG: --address="0.0.0.0" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495220 4775 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495241 4775 flags.go:64] FLAG: --anonymous-auth="true" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495251 4775 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495261 4775 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495268 4775 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495279 4775 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495289 4775 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495297 4775 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495304 4775 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495313 4775 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495321 4775 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495328 4775 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495335 4775 flags.go:64] FLAG: --cgroup-root="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495345 4775 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495353 4775 flags.go:64] FLAG: --client-ca-file="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495360 4775 flags.go:64] FLAG: --cloud-config="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495368 4775 flags.go:64] FLAG: --cloud-provider="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495376 4775 flags.go:64] FLAG: --cluster-dns="[]" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495386 4775 flags.go:64] FLAG: --cluster-domain="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495414 4775 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495422 4775 flags.go:64] FLAG: --config-dir="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495429 4775 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495438 4775 flags.go:64] FLAG: --container-log-max-files="5" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495449 4775 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495457 4775 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495465 4775 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495472 4775 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495479 4775 flags.go:64] FLAG: --contention-profiling="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495486 4775 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495493 4775 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495501 4775 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495508 4775 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495518 4775 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495525 4775 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495533 4775 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495540 4775 flags.go:64] FLAG: --enable-load-reader="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495547 4775 flags.go:64] FLAG: --enable-server="true" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495554 4775 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495563 4775 flags.go:64] FLAG: --event-burst="100" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495571 4775 flags.go:64] FLAG: --event-qps="50" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495578 4775 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495585 4775 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495593 4775 flags.go:64] FLAG: --eviction-hard="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495603 4775 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495610 4775 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495617 4775 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495626 4775 flags.go:64] FLAG: --eviction-soft="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495633 4775 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495640 4775 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495647 4775 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495654 4775 flags.go:64] FLAG: --experimental-mounter-path="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495662 4775 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495673 4775 flags.go:64] FLAG: --fail-swap-on="true" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495680 4775 flags.go:64] FLAG: --feature-gates="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495689 4775 flags.go:64] FLAG: --file-check-frequency="20s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495696 4775 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495703 4775 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495711 4775 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495719 4775 flags.go:64] FLAG: --healthz-port="10248" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495726 4775 flags.go:64] FLAG: --help="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495733 4775 flags.go:64] FLAG: --hostname-override="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495739 4775 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495747 4775 flags.go:64] FLAG: --http-check-frequency="20s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495753 4775 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495760 4775 flags.go:64] FLAG: --image-credential-provider-config="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495767 4775 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495774 4775 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495782 4775 flags.go:64] FLAG: --image-service-endpoint="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495788 4775 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495795 4775 flags.go:64] FLAG: --kube-api-burst="100" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495802 4775 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495811 4775 flags.go:64] FLAG: --kube-api-qps="50" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495819 4775 flags.go:64] FLAG: --kube-reserved="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495826 4775 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495833 4775 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495840 4775 flags.go:64] FLAG: --kubelet-cgroups="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495847 4775 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495854 4775 flags.go:64] FLAG: --lock-file="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495861 4775 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495868 4775 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495875 4775 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495887 4775 flags.go:64] FLAG: --log-json-split-stream="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495895 4775 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495902 4775 flags.go:64] FLAG: --log-text-split-stream="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495909 4775 flags.go:64] FLAG: --logging-format="text" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495916 4775 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495924 4775 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495931 4775 flags.go:64] FLAG: --manifest-url="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495940 4775 flags.go:64] FLAG: --manifest-url-header="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495973 4775 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495982 4775 flags.go:64] FLAG: --max-open-files="1000000" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.495992 4775 flags.go:64] FLAG: --max-pods="110" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496000 4775 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496008 4775 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496017 4775 flags.go:64] FLAG: --memory-manager-policy="None" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496024 4775 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496033 4775 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496040 4775 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496048 4775 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496066 4775 flags.go:64] FLAG: --node-status-max-images="50" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496073 4775 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496080 4775 flags.go:64] FLAG: --oom-score-adj="-999" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496087 4775 flags.go:64] FLAG: --pod-cidr="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496095 4775 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496109 4775 flags.go:64] FLAG: --pod-manifest-path="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496117 4775 flags.go:64] FLAG: --pod-max-pids="-1" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496125 4775 flags.go:64] FLAG: --pods-per-core="0" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496132 4775 flags.go:64] FLAG: --port="10250" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496139 4775 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496146 4775 flags.go:64] FLAG: --provider-id="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496153 4775 flags.go:64] FLAG: --qos-reserved="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496160 4775 flags.go:64] FLAG: --read-only-port="10255" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496167 4775 flags.go:64] FLAG: --register-node="true" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496174 4775 flags.go:64] FLAG: --register-schedulable="true" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496181 4775 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496195 4775 flags.go:64] FLAG: --registry-burst="10" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496203 4775 flags.go:64] FLAG: --registry-qps="5" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496210 4775 flags.go:64] FLAG: --reserved-cpus="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496216 4775 flags.go:64] FLAG: --reserved-memory="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496226 4775 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496233 4775 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496240 4775 flags.go:64] FLAG: --rotate-certificates="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496247 4775 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496254 4775 flags.go:64] FLAG: --runonce="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496263 4775 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496270 4775 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496278 4775 flags.go:64] FLAG: --seccomp-default="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496285 4775 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496297 4775 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496305 4775 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496313 4775 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496321 4775 flags.go:64] FLAG: --storage-driver-password="root" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496328 4775 flags.go:64] FLAG: --storage-driver-secure="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496336 4775 flags.go:64] FLAG: --storage-driver-table="stats" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496344 4775 flags.go:64] FLAG: --storage-driver-user="root" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496352 4775 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496360 4775 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496367 4775 flags.go:64] FLAG: --system-cgroups="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496374 4775 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496388 4775 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496395 4775 flags.go:64] FLAG: --tls-cert-file="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496402 4775 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496411 4775 flags.go:64] FLAG: --tls-min-version="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496418 4775 flags.go:64] FLAG: --tls-private-key-file="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496425 4775 flags.go:64] FLAG: --topology-manager-policy="none" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496432 4775 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496438 4775 flags.go:64] FLAG: --topology-manager-scope="container" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496446 4775 flags.go:64] FLAG: --v="2" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496456 4775 flags.go:64] FLAG: --version="false" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496465 4775 flags.go:64] FLAG: --vmodule="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496474 4775 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.496481 4775 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496641 4775 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496652 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496659 4775 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496667 4775 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496674 4775 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496680 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496686 4775 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496692 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496701 4775 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496708 4775 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496714 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496720 4775 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496726 4775 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496732 4775 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496739 4775 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496745 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496751 4775 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496757 4775 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496763 4775 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496769 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496775 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496781 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496787 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496793 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496800 4775 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496806 4775 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496812 4775 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496818 4775 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496825 4775 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496832 4775 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496838 4775 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496844 4775 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496849 4775 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496855 4775 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496861 4775 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496868 4775 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496874 4775 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496880 4775 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496887 4775 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496893 4775 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496902 4775 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496910 4775 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496917 4775 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496925 4775 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496933 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496940 4775 feature_gate.go:330] unrecognized feature gate: Example Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496947 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496978 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496985 4775 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496991 4775 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.496997 4775 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497003 4775 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497009 4775 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497015 4775 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497023 4775 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497029 4775 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497036 4775 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497042 4775 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497048 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497055 4775 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497062 4775 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497070 4775 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497078 4775 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497085 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497093 4775 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497100 4775 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497107 4775 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497114 4775 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497120 4775 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497126 4775 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.497132 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.497154 4775 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.510525 4775 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.510598 4775 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510710 4775 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510723 4775 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510730 4775 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510737 4775 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510742 4775 feature_gate.go:330] unrecognized feature gate: Example Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510747 4775 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510752 4775 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510756 4775 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510761 4775 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510766 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510772 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510777 4775 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510782 4775 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510790 4775 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510796 4775 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510801 4775 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510807 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510814 4775 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510820 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510829 4775 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510841 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510848 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510855 4775 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510861 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510868 4775 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510873 4775 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510878 4775 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510882 4775 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510886 4775 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510890 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510895 4775 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510899 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510903 4775 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510909 4775 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510914 4775 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510921 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510926 4775 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510931 4775 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510935 4775 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510940 4775 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510947 4775 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510972 4775 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510978 4775 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510983 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510988 4775 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.510994 4775 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511001 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511006 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511011 4775 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511017 4775 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511022 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511026 4775 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511031 4775 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511038 4775 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511045 4775 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511049 4775 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511054 4775 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511058 4775 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511062 4775 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511066 4775 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511070 4775 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511075 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511079 4775 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511084 4775 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511089 4775 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511093 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511114 4775 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511120 4775 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511125 4775 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511130 4775 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511135 4775 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.511144 4775 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511283 4775 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511295 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511301 4775 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511305 4775 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511311 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511315 4775 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511319 4775 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511324 4775 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511330 4775 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511336 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511340 4775 feature_gate.go:330] unrecognized feature gate: Example Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511345 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511349 4775 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511356 4775 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511363 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511367 4775 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511394 4775 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511399 4775 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511403 4775 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511409 4775 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511415 4775 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511419 4775 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511423 4775 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511428 4775 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511432 4775 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511437 4775 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511441 4775 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511445 4775 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511449 4775 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511454 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511459 4775 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511463 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511468 4775 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511471 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511475 4775 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511479 4775 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511482 4775 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511486 4775 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511489 4775 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511493 4775 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511496 4775 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511500 4775 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511504 4775 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511507 4775 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511511 4775 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511514 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511518 4775 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511522 4775 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511526 4775 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511530 4775 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511533 4775 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511538 4775 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511541 4775 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511545 4775 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511549 4775 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511552 4775 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511556 4775 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511560 4775 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511565 4775 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511569 4775 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511574 4775 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511580 4775 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511585 4775 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511588 4775 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511592 4775 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511595 4775 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511599 4775 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511602 4775 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511606 4775 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511610 4775 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.511614 4775 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.511621 4775 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.512660 4775 server.go:940] "Client rotation is on, will bootstrap in background" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.516625 4775 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.516708 4775 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.518733 4775 server.go:997] "Starting client certificate rotation" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.518751 4775 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.519044 4775 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-09 10:49:09.839925718 +0000 UTC Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.519172 4775 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 921h8m6.320758942s for next certificate rotation Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.555139 4775 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.560884 4775 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.588539 4775 log.go:25] "Validated CRI v1 runtime API" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.633692 4775 log.go:25] "Validated CRI v1 image API" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.636615 4775 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.643256 4775 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-02-01-35-48-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.643334 4775 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:41 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.663160 4775 manager.go:217] Machine: {Timestamp:2025-10-02 01:41:03.660333292 +0000 UTC m=+0.827077372 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:c12787e0-b782-423d-acab-1fed869cc978 BootID:5a28e63a-b617-4331-805d-ce489133ccf9 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:41 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:47:51:5c Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:47:51:5c Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:97:34:de Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:4e:78:23 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:4a:84:18 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e3:1f:7f Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:b6:c3:e0 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:56:d2:0b:40:83:c8 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:3a:a5:70:49:e4:bb Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.663487 4775 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.663756 4775 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.665226 4775 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.665857 4775 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.665913 4775 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.666248 4775 topology_manager.go:138] "Creating topology manager with none policy" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.666265 4775 container_manager_linux.go:303] "Creating device plugin manager" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.666805 4775 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.666839 4775 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.667124 4775 state_mem.go:36] "Initialized new in-memory state store" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.667725 4775 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.671319 4775 kubelet.go:418] "Attempting to sync node with API server" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.671346 4775 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.671377 4775 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.671396 4775 kubelet.go:324] "Adding apiserver pod source" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.671415 4775 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.682433 4775 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.683576 4775 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.684100 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.684166 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Oct 02 01:41:03 crc kubenswrapper[4775]: E1002 01:41:03.684294 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:41:03 crc kubenswrapper[4775]: E1002 01:41:03.684289 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.686864 4775 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.689289 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.689347 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.689367 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.689386 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.689415 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.689434 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.689453 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.689482 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.689506 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.689556 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.689583 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.689602 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.692006 4775 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.693173 4775 server.go:1280] "Started kubelet" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.693391 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.694225 4775 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.694268 4775 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.695289 4775 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 02 01:41:03 crc systemd[1]: Started Kubernetes Kubelet. Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.697083 4775 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.697153 4775 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.697875 4775 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.697919 4775 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.698117 4775 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 02 01:41:03 crc kubenswrapper[4775]: E1002 01:41:03.698542 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.698940 4775 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 21:56:31.009209597 +0000 UTC Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.699398 4775 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1508h15m27.309835449s for next certificate rotation Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.699758 4775 server.go:460] "Adding debug handlers to kubelet server" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.700302 4775 factory.go:55] Registering systemd factory Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.700345 4775 factory.go:221] Registration of the systemd container factory successfully Oct 02 01:41:03 crc kubenswrapper[4775]: E1002 01:41:03.700399 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="200ms" Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.700778 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Oct 02 01:41:03 crc kubenswrapper[4775]: E1002 01:41:03.701036 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.700852 4775 factory.go:153] Registering CRI-O factory Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.702204 4775 factory.go:221] Registration of the crio container factory successfully Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.702277 4775 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.702303 4775 factory.go:103] Registering Raw factory Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.702324 4775 manager.go:1196] Started watching for new ooms in manager Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.703507 4775 manager.go:319] Starting recovery of all containers Oct 02 01:41:03 crc kubenswrapper[4775]: E1002 01:41:03.703036 4775 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.47:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186a8906172b198f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-02 01:41:03.693109647 +0000 UTC m=+0.859853757,LastTimestamp:2025-10-02 01:41:03.693109647 +0000 UTC m=+0.859853757,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723284 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723464 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723497 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723530 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723558 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723587 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723616 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723643 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723672 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723710 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723736 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723764 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723790 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723820 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723848 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.723880 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724173 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724226 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724254 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724279 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724298 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724326 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724347 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724376 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724402 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724423 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724462 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724488 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724514 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724535 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724627 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724652 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724674 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724695 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724724 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724754 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724775 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724796 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724818 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724840 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724860 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724881 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724902 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724922 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724943 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.724992 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725018 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725084 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725109 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725134 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725156 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725177 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725215 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725238 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725261 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725401 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725425 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725446 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725467 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725488 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725508 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725530 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725552 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725573 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725593 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725617 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725641 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.725662 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728025 4775 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728099 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728136 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728167 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728195 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728217 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728238 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728260 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728291 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728313 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728338 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728359 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728380 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728399 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728429 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728458 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728479 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728502 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728522 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728546 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728567 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728590 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728619 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728652 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728705 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728734 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728759 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728780 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728801 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728828 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728859 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728888 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728918 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.728948 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729047 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729076 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729099 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729140 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729165 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729187 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729215 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729239 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729265 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729288 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729322 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729346 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729370 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729392 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729416 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729441 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729464 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729487 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729510 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729533 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729554 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729574 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729599 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729619 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729643 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729686 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729708 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729732 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729754 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729775 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729796 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729833 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729855 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729877 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729898 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729928 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.729949 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730000 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730023 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730049 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730070 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730093 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730113 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730134 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730155 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730177 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730200 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730220 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730244 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730265 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730290 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730319 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730341 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730369 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730392 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730414 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730468 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730490 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730520 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730540 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730561 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730581 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730601 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730621 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730642 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730665 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730696 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730720 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730747 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730776 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730804 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730833 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730869 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730899 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730928 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.730994 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731021 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731045 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731068 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731124 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731146 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731178 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731200 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731220 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731241 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731277 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731298 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731324 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731348 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731371 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731395 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731431 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731452 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731473 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731521 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731544 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731566 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731591 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731612 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731635 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731656 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731678 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731700 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731721 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731745 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731804 4775 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731824 4775 reconstruct.go:97] "Volume reconstruction finished" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.731839 4775 reconciler.go:26] "Reconciler: start to sync state" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.740310 4775 manager.go:324] Recovery completed Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.754745 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.757591 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.757658 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.757679 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.759685 4775 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.759712 4775 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.759732 4775 state_mem.go:36] "Initialized new in-memory state store" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.761724 4775 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.763837 4775 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.763927 4775 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.763989 4775 kubelet.go:2335] "Starting kubelet main sync loop" Oct 02 01:41:03 crc kubenswrapper[4775]: E1002 01:41:03.764089 4775 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 02 01:41:03 crc kubenswrapper[4775]: W1002 01:41:03.767434 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Oct 02 01:41:03 crc kubenswrapper[4775]: E1002 01:41:03.767693 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.779727 4775 policy_none.go:49] "None policy: Start" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.781591 4775 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.781647 4775 state_mem.go:35] "Initializing new in-memory state store" Oct 02 01:41:03 crc kubenswrapper[4775]: E1002 01:41:03.800302 4775 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.829386 4775 manager.go:334] "Starting Device Plugin manager" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.829486 4775 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.829527 4775 server.go:79] "Starting device plugin registration server" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.830329 4775 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.830365 4775 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.831128 4775 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.831275 4775 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.831299 4775 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 02 01:41:03 crc kubenswrapper[4775]: E1002 01:41:03.844421 4775 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.864734 4775 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.864856 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.866136 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.866175 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.866186 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.866366 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.866801 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.866869 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.867642 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.867743 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.867698 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.867798 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.867804 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.867921 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.868179 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.868232 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.868256 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.871131 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.871173 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.871190 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.871286 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.871331 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.871368 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.871457 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.871864 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.871931 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.872848 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.872898 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.872912 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.873013 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.873057 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.873203 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.874095 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.874291 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.874379 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.875297 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.875603 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.875730 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.875778 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.875984 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.876006 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.876365 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.876547 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.877725 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.877900 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.878070 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:03 crc kubenswrapper[4775]: E1002 01:41:03.902317 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="400ms" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.932149 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.933523 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.933568 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.933585 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.933627 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934081 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934146 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934199 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934245 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934283 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: E1002 01:41:03.934268 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.47:6443: connect: connection refused" node="crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934352 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934464 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934508 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934532 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934558 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934607 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934635 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934653 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934674 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:03 crc kubenswrapper[4775]: I1002 01:41:03.934696 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036432 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036515 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036554 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036587 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036635 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036667 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036654 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036694 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036741 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036696 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036832 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036853 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036835 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036853 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036891 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036781 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036922 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.036950 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.037002 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.037026 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.037016 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.037048 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.037097 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.037150 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.037188 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.037227 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.037285 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.037327 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.037314 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.037418 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.134712 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.136930 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.136996 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.137007 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.137037 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 01:41:04 crc kubenswrapper[4775]: E1002 01:41:04.137611 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.47:6443: connect: connection refused" node="crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.205149 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.228601 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.242585 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.278560 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: W1002 01:41:04.283939 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-a67a7455ecf433197cef478bda1ec7979a7b638d8108b5613d9d26763be608b8 WatchSource:0}: Error finding container a67a7455ecf433197cef478bda1ec7979a7b638d8108b5613d9d26763be608b8: Status 404 returned error can't find the container with id a67a7455ecf433197cef478bda1ec7979a7b638d8108b5613d9d26763be608b8 Oct 02 01:41:04 crc kubenswrapper[4775]: W1002 01:41:04.287967 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-d0429b41197d0135009a1f5f594f6b37670966fc01e05e95fa224b0770e1d08e WatchSource:0}: Error finding container d0429b41197d0135009a1f5f594f6b37670966fc01e05e95fa224b0770e1d08e: Status 404 returned error can't find the container with id d0429b41197d0135009a1f5f594f6b37670966fc01e05e95fa224b0770e1d08e Oct 02 01:41:04 crc kubenswrapper[4775]: W1002 01:41:04.288918 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-ec71c070501d2a885c9529d29f0ae8d5fad45b049ca82740755ef3a55dca6398 WatchSource:0}: Error finding container ec71c070501d2a885c9529d29f0ae8d5fad45b049ca82740755ef3a55dca6398: Status 404 returned error can't find the container with id ec71c070501d2a885c9529d29f0ae8d5fad45b049ca82740755ef3a55dca6398 Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.289371 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:04 crc kubenswrapper[4775]: E1002 01:41:04.304006 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="800ms" Oct 02 01:41:04 crc kubenswrapper[4775]: W1002 01:41:04.304548 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-3989b85fd55638d23aa736cfd157ac3a29512d9772fd1a4bb8b70508b42b17ec WatchSource:0}: Error finding container 3989b85fd55638d23aa736cfd157ac3a29512d9772fd1a4bb8b70508b42b17ec: Status 404 returned error can't find the container with id 3989b85fd55638d23aa736cfd157ac3a29512d9772fd1a4bb8b70508b42b17ec Oct 02 01:41:04 crc kubenswrapper[4775]: W1002 01:41:04.312308 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-48dead3f9ce75d5f83ca2aa24493dd60a5919c3c16cba1044a926deb3aae01c0 WatchSource:0}: Error finding container 48dead3f9ce75d5f83ca2aa24493dd60a5919c3c16cba1044a926deb3aae01c0: Status 404 returned error can't find the container with id 48dead3f9ce75d5f83ca2aa24493dd60a5919c3c16cba1044a926deb3aae01c0 Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.538045 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.539821 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.539864 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.539873 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.539905 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 01:41:04 crc kubenswrapper[4775]: E1002 01:41:04.540845 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.47:6443: connect: connection refused" node="crc" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.694456 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Oct 02 01:41:04 crc kubenswrapper[4775]: W1002 01:41:04.708486 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Oct 02 01:41:04 crc kubenswrapper[4775]: E1002 01:41:04.708601 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.770931 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"48dead3f9ce75d5f83ca2aa24493dd60a5919c3c16cba1044a926deb3aae01c0"} Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.772208 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3989b85fd55638d23aa736cfd157ac3a29512d9772fd1a4bb8b70508b42b17ec"} Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.773834 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ec71c070501d2a885c9529d29f0ae8d5fad45b049ca82740755ef3a55dca6398"} Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.775139 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d0429b41197d0135009a1f5f594f6b37670966fc01e05e95fa224b0770e1d08e"} Oct 02 01:41:04 crc kubenswrapper[4775]: I1002 01:41:04.776218 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a67a7455ecf433197cef478bda1ec7979a7b638d8108b5613d9d26763be608b8"} Oct 02 01:41:05 crc kubenswrapper[4775]: W1002 01:41:05.014251 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Oct 02 01:41:05 crc kubenswrapper[4775]: E1002 01:41:05.014483 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:41:05 crc kubenswrapper[4775]: W1002 01:41:05.067116 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Oct 02 01:41:05 crc kubenswrapper[4775]: E1002 01:41:05.067362 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:41:05 crc kubenswrapper[4775]: W1002 01:41:05.076082 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Oct 02 01:41:05 crc kubenswrapper[4775]: E1002 01:41:05.076197 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:41:05 crc kubenswrapper[4775]: E1002 01:41:05.105085 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="1.6s" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.341263 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.343296 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.343370 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.343390 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.343428 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 01:41:05 crc kubenswrapper[4775]: E1002 01:41:05.344056 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.47:6443: connect: connection refused" node="crc" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.695027 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.783472 4775 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="ce6578795fea6ce646a7bfcbd7599591e0e189c7fd8ba2088f6b959b96be7964" exitCode=0 Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.783623 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.783600 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"ce6578795fea6ce646a7bfcbd7599591e0e189c7fd8ba2088f6b959b96be7964"} Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.786153 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.786203 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.786222 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.789034 4775 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345" exitCode=0 Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.789222 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.789219 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345"} Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.791010 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.791052 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.791071 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.794664 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796"} Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.794718 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e"} Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.797036 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa" exitCode=0 Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.797164 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa"} Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.797374 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.798802 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.798883 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.798910 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.799818 4775 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423" exitCode=0 Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.799878 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423"} Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.800065 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.806595 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.807729 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.807760 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.807787 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.808633 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.808714 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:05 crc kubenswrapper[4775]: I1002 01:41:05.808741 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:06 crc kubenswrapper[4775]: W1002 01:41:06.545323 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Oct 02 01:41:06 crc kubenswrapper[4775]: E1002 01:41:06.545450 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.694491 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Oct 02 01:41:06 crc kubenswrapper[4775]: E1002 01:41:06.706940 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.47:6443: connect: connection refused" interval="3.2s" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.807838 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4"} Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.807881 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508"} Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.807892 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29"} Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.810325 4775 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666" exitCode=0 Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.810374 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666"} Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.810475 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.811391 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.811412 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.811426 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.813519 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"dfca4ef06649242b5de270bcec809ecc2fcee3f340129e6581efe764f2fecf52"} Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.813591 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.814824 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.814868 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.814885 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.816547 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7b0204478719ee799c2b7cff276ce9514d66b27c9eafa538ceec0e6478c0bf47"} Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.816571 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1a349484f3015205e6074e79651f9e3a242b3b2597b70e60231115430ee41cd5"} Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.816581 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4ef520bf5435b6a70fb920599ee3b72d11062641a24a364340ced83ff966eeab"} Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.816641 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.817320 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.817374 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.817394 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.820090 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899"} Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.820133 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9"} Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.820195 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.825209 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.825322 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.825423 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.944340 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.945765 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.945807 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.945818 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:06 crc kubenswrapper[4775]: I1002 01:41:06.945858 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 01:41:06 crc kubenswrapper[4775]: E1002 01:41:06.946342 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.47:6443: connect: connection refused" node="crc" Oct 02 01:41:07 crc kubenswrapper[4775]: W1002 01:41:07.152519 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.47:6443: connect: connection refused Oct 02 01:41:07 crc kubenswrapper[4775]: E1002 01:41:07.152638 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.47:6443: connect: connection refused" logger="UnhandledError" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.826910 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778"} Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.826986 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653"} Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.827062 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.828097 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.828117 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.828125 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.829702 4775 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8" exitCode=0 Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.829781 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.829771 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8"} Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.829881 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.829921 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.829896 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.830077 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.830844 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.830887 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.830919 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.831299 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.831361 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.831381 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.831326 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.831435 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.831443 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.831835 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.831891 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.831912 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.923707 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:07 crc kubenswrapper[4775]: I1002 01:41:07.975014 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.690380 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.701313 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.839591 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90"} Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.839648 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d"} Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.839661 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153"} Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.839676 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.839702 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.839767 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.839711 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.841291 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.841338 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.841355 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.842146 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.842168 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.842208 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.842226 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.842176 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:08 crc kubenswrapper[4775]: I1002 01:41:08.842274 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.654908 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.759008 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.845518 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7"} Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.845569 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0"} Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.845577 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.845611 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.845585 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.846644 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.846676 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.846689 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.846698 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.846722 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.846739 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.847322 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.847352 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:09 crc kubenswrapper[4775]: I1002 01:41:09.847360 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.147055 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.148707 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.148760 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.148776 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.148813 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.848499 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.848570 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.848703 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.850355 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.850369 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.850400 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.850405 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.850414 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.850411 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.850459 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.850477 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.850423 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.975448 4775 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 01:41:10 crc kubenswrapper[4775]: I1002 01:41:10.975602 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 01:41:12 crc kubenswrapper[4775]: I1002 01:41:12.205037 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:12 crc kubenswrapper[4775]: I1002 01:41:12.205287 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:12 crc kubenswrapper[4775]: I1002 01:41:12.206858 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:12 crc kubenswrapper[4775]: I1002 01:41:12.206907 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:12 crc kubenswrapper[4775]: I1002 01:41:12.206924 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:13 crc kubenswrapper[4775]: E1002 01:41:13.844544 4775 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 02 01:41:14 crc kubenswrapper[4775]: I1002 01:41:14.658295 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 02 01:41:14 crc kubenswrapper[4775]: I1002 01:41:14.658566 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:14 crc kubenswrapper[4775]: I1002 01:41:14.660267 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:14 crc kubenswrapper[4775]: I1002 01:41:14.660345 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:14 crc kubenswrapper[4775]: I1002 01:41:14.660366 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:15 crc kubenswrapper[4775]: I1002 01:41:15.863425 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:15 crc kubenswrapper[4775]: I1002 01:41:15.863673 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:15 crc kubenswrapper[4775]: I1002 01:41:15.865144 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:15 crc kubenswrapper[4775]: I1002 01:41:15.865210 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:15 crc kubenswrapper[4775]: I1002 01:41:15.865229 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:17 crc kubenswrapper[4775]: W1002 01:41:17.470186 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 02 01:41:17 crc kubenswrapper[4775]: I1002 01:41:17.471070 4775 trace.go:236] Trace[953534668]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 01:41:07.468) (total time: 10002ms): Oct 02 01:41:17 crc kubenswrapper[4775]: Trace[953534668]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (01:41:17.470) Oct 02 01:41:17 crc kubenswrapper[4775]: Trace[953534668]: [10.002161168s] [10.002161168s] END Oct 02 01:41:17 crc kubenswrapper[4775]: E1002 01:41:17.471350 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 02 01:41:17 crc kubenswrapper[4775]: W1002 01:41:17.481242 4775 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 02 01:41:17 crc kubenswrapper[4775]: I1002 01:41:17.481626 4775 trace.go:236] Trace[26283688]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 01:41:07.479) (total time: 10001ms): Oct 02 01:41:17 crc kubenswrapper[4775]: Trace[26283688]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (01:41:17.481) Oct 02 01:41:17 crc kubenswrapper[4775]: Trace[26283688]: [10.001928035s] [10.001928035s] END Oct 02 01:41:17 crc kubenswrapper[4775]: E1002 01:41:17.481883 4775 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 02 01:41:17 crc kubenswrapper[4775]: I1002 01:41:17.585530 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 02 01:41:17 crc kubenswrapper[4775]: I1002 01:41:17.585785 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:17 crc kubenswrapper[4775]: I1002 01:41:17.587271 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:17 crc kubenswrapper[4775]: I1002 01:41:17.587485 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:17 crc kubenswrapper[4775]: I1002 01:41:17.587676 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:17 crc kubenswrapper[4775]: I1002 01:41:17.695729 4775 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 02 01:41:18 crc kubenswrapper[4775]: I1002 01:41:18.571573 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 01:41:18 crc kubenswrapper[4775]: I1002 01:41:18.571710 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 01:41:18 crc kubenswrapper[4775]: I1002 01:41:18.576391 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 02 01:41:18 crc kubenswrapper[4775]: I1002 01:41:18.576490 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 02 01:41:19 crc kubenswrapper[4775]: I1002 01:41:19.763884 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]log ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]etcd ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/openshift.io-api-request-count-filter ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/openshift.io-startkubeinformers ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/generic-apiserver-start-informers ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/priority-and-fairness-config-consumer ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/priority-and-fairness-filter ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/start-apiextensions-informers ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/start-apiextensions-controllers ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/crd-informer-synced ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/start-system-namespaces-controller ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/start-cluster-authentication-info-controller ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/start-legacy-token-tracking-controller ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/start-service-ip-repair-controllers ok Oct 02 01:41:19 crc kubenswrapper[4775]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/priority-and-fairness-config-producer ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/bootstrap-controller ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/start-kube-aggregator-informers ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/apiservice-status-local-available-controller ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/apiservice-status-remote-available-controller ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/apiservice-registration-controller ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/apiservice-wait-for-first-sync ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/apiservice-discovery-controller ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/kube-apiserver-autoregistration ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]autoregister-completion ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/apiservice-openapi-controller ok Oct 02 01:41:19 crc kubenswrapper[4775]: [+]poststarthook/apiservice-openapiv3-controller ok Oct 02 01:41:19 crc kubenswrapper[4775]: livez check failed Oct 02 01:41:19 crc kubenswrapper[4775]: I1002 01:41:19.763988 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:41:20 crc kubenswrapper[4775]: I1002 01:41:20.976013 4775 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 01:41:20 crc kubenswrapper[4775]: I1002 01:41:20.976114 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.191698 4775 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.683509 4775 apiserver.go:52] "Watching apiserver" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.696257 4775 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.696569 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.697276 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.697545 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.697432 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.697813 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:22 crc kubenswrapper[4775]: E1002 01:41:22.698227 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:22 crc kubenswrapper[4775]: E1002 01:41:22.698322 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.698900 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.699031 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:22 crc kubenswrapper[4775]: E1002 01:41:22.699128 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.702720 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.703262 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.703257 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.703555 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.703465 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.703591 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.703994 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.706112 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.706434 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.742069 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.754912 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.773228 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.776920 4775 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.796748 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.799414 4775 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.825178 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.847304 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:22 crc kubenswrapper[4775]: I1002 01:41:22.865342 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.563504 4775 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.566382 4775 trace.go:236] Trace[1861996822]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 01:41:12.177) (total time: 11389ms): Oct 02 01:41:23 crc kubenswrapper[4775]: Trace[1861996822]: ---"Objects listed" error: 11389ms (01:41:23.566) Oct 02 01:41:23 crc kubenswrapper[4775]: Trace[1861996822]: [11.389104557s] [11.389104557s] END Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.566447 4775 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.566508 4775 trace.go:236] Trace[596765899]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (02-Oct-2025 01:41:11.309) (total time: 12257ms): Oct 02 01:41:23 crc kubenswrapper[4775]: Trace[596765899]: ---"Objects listed" error: 12256ms (01:41:23.566) Oct 02 01:41:23 crc kubenswrapper[4775]: Trace[596765899]: [12.257020313s] [12.257020313s] END Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.566542 4775 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.567326 4775 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.572366 4775 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.619435 4775 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38356->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.619559 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38356->192.168.126.11:17697: read: connection reset by peer" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668078 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668167 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668296 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668332 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668386 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668419 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668447 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668481 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668516 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668547 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668625 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668619 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668664 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668697 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668777 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668809 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668806 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668840 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668870 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668913 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668982 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.668902 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669023 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669069 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669108 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669143 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669150 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669141 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669190 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669225 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669264 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669301 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669334 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669351 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669367 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669402 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669435 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669468 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669502 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669536 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669569 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669677 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669719 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669752 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669366 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669374 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669787 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669821 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669853 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669889 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669943 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670000 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670030 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670097 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670132 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670163 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670195 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670227 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670259 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670294 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670327 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670359 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670391 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670426 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670458 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670489 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670529 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670562 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670592 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670627 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670662 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670696 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670731 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670766 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670801 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670889 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670944 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671010 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671045 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671076 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671109 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671164 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671199 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671241 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671283 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671330 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671365 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671402 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671584 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671617 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671652 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671692 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671725 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671763 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671796 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671829 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671862 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671894 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671927 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671998 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672033 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672072 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672142 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672298 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672347 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672383 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672416 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672450 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672484 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672519 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672554 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672589 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672624 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672659 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672696 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672733 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672767 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672801 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672837 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672872 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672904 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672938 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672996 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673033 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673072 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673107 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673141 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673179 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673215 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673248 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673280 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673315 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673348 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669390 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669597 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674081 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674170 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674237 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674264 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669652 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674292 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669666 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674299 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674322 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674350 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674379 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674412 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674440 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674466 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674489 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674513 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674539 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674562 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674584 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674606 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674630 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674652 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674674 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674743 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674768 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674790 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674818 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674841 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674866 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674895 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674921 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674943 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674991 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.675015 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.675039 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676169 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676210 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676236 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676269 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676300 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676327 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676357 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676385 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676411 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676450 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676522 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676559 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676583 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676607 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676632 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676661 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676694 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676721 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676745 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676771 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676795 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676820 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676849 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676874 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676904 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676929 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676969 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677031 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677063 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677088 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677116 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677139 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677164 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677188 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677212 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677235 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677258 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677284 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677308 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677328 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677378 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677407 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677427 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677450 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677470 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677496 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680353 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680417 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680444 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680472 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680510 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680541 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680571 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680597 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680696 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680715 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680730 4775 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680743 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680766 4775 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680780 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680795 4775 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680809 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680826 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680839 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680854 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680870 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680884 4775 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680898 4775 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669658 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669666 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.669743 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670028 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.670219 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671347 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671373 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671507 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.671890 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672021 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672250 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672441 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672595 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.672648 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673000 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673157 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673324 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673678 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.673924 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674018 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674031 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674179 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674352 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.674989 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.675025 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.675323 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.675438 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.675552 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.675829 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.675857 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.675879 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.675922 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.676795 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677151 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677340 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.677342 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.678003 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.678142 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.678263 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.678544 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.678813 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.679107 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.679231 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.679597 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.680440 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.681192 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.681412 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.681494 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.681662 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.681819 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.681761 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.682147 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.682598 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.685169 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.685498 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.685506 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.686065 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.686078 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.686719 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.686750 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.686733 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.686738 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.687001 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.687317 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.687258 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.687219 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.687575 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.687609 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.687896 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.688271 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.688628 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.688789 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.689026 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.689644 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.690009 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.690949 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.691075 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.691110 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.691639 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.691895 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.692600 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.693029 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.693752 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.694232 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.694256 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.694263 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.694337 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.694439 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.694672 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.694786 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.694814 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.694996 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.695076 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.695510 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.681289 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.696773 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.696776 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.697160 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.697336 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.697440 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.697517 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.698015 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.698313 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.698369 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.698388 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.698692 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:41:24.198650626 +0000 UTC m=+21.365394676 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.700865 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.699304 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.700107 4775 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.699334 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.700854 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.700878 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.701131 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.701233 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:24.201202115 +0000 UTC m=+21.367946345 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.701258 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:24.201249596 +0000 UTC m=+21.367993636 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.701727 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.702520 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.714592 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.701910 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.703002 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.703013 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.704371 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.706096 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.707497 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.710719 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.714077 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.715099 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.715114 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.715190 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:24.215172319 +0000 UTC m=+21.381916359 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.717098 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.717848 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.718134 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.720165 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.720598 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.720788 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.720806 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.720821 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:23 crc kubenswrapper[4775]: E1002 01:41:23.720877 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:24.220859862 +0000 UTC m=+21.387603912 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.724038 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.727703 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.728235 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.731981 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.735128 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.736270 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.736710 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.737376 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.737499 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.739072 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.739273 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.739505 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.739516 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.739539 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.739721 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.739853 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.739901 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.739939 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.739973 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.740139 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.739452 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.741675 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.741875 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.742109 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.744035 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.744112 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.746557 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.746682 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.746916 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.747356 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.747633 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.747887 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.748045 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.748255 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.748306 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.748587 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.748967 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.749036 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.749241 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.749539 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.749586 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.749707 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.749810 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.749853 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.749867 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.749881 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.750096 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.750179 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.750326 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.750351 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.750850 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.750850 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.750979 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.752394 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.752503 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.752649 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.752720 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.752734 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.753151 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.757359 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.764287 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.774089 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.774664 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.776414 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.777086 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.777586 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.778346 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.779002 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.779703 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.780891 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781556 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781605 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781657 4775 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781668 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781679 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781691 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781702 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781711 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781722 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781759 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781760 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781778 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781789 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781678 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781800 4775 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781821 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781831 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781841 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781849 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781858 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781868 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781876 4775 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781886 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781895 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781905 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781914 4775 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781925 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781935 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781944 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781970 4775 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781979 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781988 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.781998 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782007 4775 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782015 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782024 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782033 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782042 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782050 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782059 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782068 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782076 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782085 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782094 4775 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782102 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782110 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782118 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782127 4775 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782136 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782144 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782153 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782161 4775 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782169 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782178 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782186 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782195 4775 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782204 4775 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782212 4775 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782220 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782229 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782239 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782247 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782257 4775 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782266 4775 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782274 4775 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782285 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782293 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782302 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782311 4775 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782320 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782329 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782337 4775 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782346 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782354 4775 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782362 4775 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782372 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782381 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782390 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782398 4775 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782406 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782414 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782422 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782432 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782441 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782451 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782460 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782470 4775 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782479 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782489 4775 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782498 4775 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782506 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782515 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782523 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782531 4775 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782539 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782548 4775 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782600 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782615 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782628 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782638 4775 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782647 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782656 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782664 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782673 4775 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782682 4775 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782691 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782701 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782710 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782719 4775 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782728 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782737 4775 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782745 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782754 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782762 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782772 4775 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782780 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782780 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782789 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782798 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782807 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782815 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782824 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782833 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782843 4775 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782855 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782864 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782873 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782881 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782891 4775 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782899 4775 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782907 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782916 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782926 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782935 4775 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782943 4775 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782972 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782980 4775 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.782991 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783000 4775 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783008 4775 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783017 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783025 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783033 4775 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783042 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783050 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783060 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783068 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783077 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783085 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783093 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783102 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783111 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783119 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783128 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783136 4775 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783145 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783156 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783230 4775 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783242 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783252 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783261 4775 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783270 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783278 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783287 4775 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783296 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783304 4775 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783312 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783321 4775 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783330 4775 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783340 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783350 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783358 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783366 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783374 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783383 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783391 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783401 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783409 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783418 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783419 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783426 4775 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783435 4775 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783444 4775 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783452 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783461 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.783469 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.784490 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.785080 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.785717 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.786937 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.787547 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.791411 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.791727 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.792472 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.793079 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.794076 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.794540 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.795341 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.796258 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.797037 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.798359 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.799402 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.800499 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.801007 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.801918 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.802412 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.802852 4775 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.803323 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.804904 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.805487 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.806328 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.807771 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.808445 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.809335 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.810051 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.811103 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.811532 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.812509 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.813230 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.814174 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.814614 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.815461 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.815966 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.817579 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.818178 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.819066 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.819291 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.819678 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.820249 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.821239 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.821763 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.850646 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.871305 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.883540 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.884256 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.885758 4775 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778" exitCode=255 Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.885829 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778"} Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.886310 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.897684 4775 scope.go:117] "RemoveContainer" containerID="f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.899729 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.900906 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.918771 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.919872 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.935067 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.935181 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:23 crc kubenswrapper[4775]: W1002 01:41:23.937224 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-51de28ef9146bb4a7332ddcac54f3e861eaccec8f63269abac61483ed2c77f80 WatchSource:0}: Error finding container 51de28ef9146bb4a7332ddcac54f3e861eaccec8f63269abac61483ed2c77f80: Status 404 returned error can't find the container with id 51de28ef9146bb4a7332ddcac54f3e861eaccec8f63269abac61483ed2c77f80 Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.945470 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.946320 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.956867 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:23 crc kubenswrapper[4775]: W1002 01:41:23.963881 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-448b804cd59e046dc96b7051fe0f1ea75cc087b8d060dd6cec468bf7acf0002d WatchSource:0}: Error finding container 448b804cd59e046dc96b7051fe0f1ea75cc087b8d060dd6cec468bf7acf0002d: Status 404 returned error can't find the container with id 448b804cd59e046dc96b7051fe0f1ea75cc087b8d060dd6cec468bf7acf0002d Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.970725 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:23 crc kubenswrapper[4775]: W1002 01:41:23.982336 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-5358f1ac48f984a63b40daaae5a04b23d804e143826a890fe865598876882896 WatchSource:0}: Error finding container 5358f1ac48f984a63b40daaae5a04b23d804e143826a890fe865598876882896: Status 404 returned error can't find the container with id 5358f1ac48f984a63b40daaae5a04b23d804e143826a890fe865598876882896 Oct 02 01:41:23 crc kubenswrapper[4775]: I1002 01:41:23.992640 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.004816 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.018730 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.287784 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.287907 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.287936 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:41:25.287907853 +0000 UTC m=+22.454651893 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.287979 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.288015 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.288039 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.288082 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.288098 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.288104 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.288113 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.288159 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:25.288149879 +0000 UTC m=+22.454893919 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.288177 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:25.28816979 +0000 UTC m=+22.454913830 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.288199 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.288215 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.288248 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.288262 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.288293 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:25.288274092 +0000 UTC m=+22.455018132 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.288318 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:25.288303103 +0000 UTC m=+22.455047143 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.764231 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.764260 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.764327 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.764479 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.764662 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:24 crc kubenswrapper[4775]: E1002 01:41:24.764793 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.764892 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.776592 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.788981 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.815107 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.843275 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-9fxjb"] Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.843244 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.843636 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9fxjb" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.845472 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.846047 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.846690 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.861014 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.873377 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.888228 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.890016 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.892226 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c"} Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.892696 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.893396 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5358f1ac48f984a63b40daaae5a04b23d804e143826a890fe865598876882896"} Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.895479 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105"} Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.895541 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"448b804cd59e046dc96b7051fe0f1ea75cc087b8d060dd6cec468bf7acf0002d"} Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.897220 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.897371 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6"} Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.897421 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330"} Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.897443 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"51de28ef9146bb4a7332ddcac54f3e861eaccec8f63269abac61483ed2c77f80"} Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.910320 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.925684 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.934676 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.967530 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.979579 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.994195 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26cqh\" (UniqueName: \"kubernetes.io/projected/88740fe4-c496-4ab9-a518-cfdaeb7fee36-kube-api-access-26cqh\") pod \"node-resolver-9fxjb\" (UID: \"88740fe4-c496-4ab9-a518-cfdaeb7fee36\") " pod="openshift-dns/node-resolver-9fxjb" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.994314 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/88740fe4-c496-4ab9-a518-cfdaeb7fee36-hosts-file\") pod \"node-resolver-9fxjb\" (UID: \"88740fe4-c496-4ab9-a518-cfdaeb7fee36\") " pod="openshift-dns/node-resolver-9fxjb" Oct 02 01:41:24 crc kubenswrapper[4775]: I1002 01:41:24.996308 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.006715 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.019205 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.031860 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.042290 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.050406 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.057997 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.068433 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.078598 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.088464 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.094880 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26cqh\" (UniqueName: \"kubernetes.io/projected/88740fe4-c496-4ab9-a518-cfdaeb7fee36-kube-api-access-26cqh\") pod \"node-resolver-9fxjb\" (UID: \"88740fe4-c496-4ab9-a518-cfdaeb7fee36\") " pod="openshift-dns/node-resolver-9fxjb" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.094979 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/88740fe4-c496-4ab9-a518-cfdaeb7fee36-hosts-file\") pod \"node-resolver-9fxjb\" (UID: \"88740fe4-c496-4ab9-a518-cfdaeb7fee36\") " pod="openshift-dns/node-resolver-9fxjb" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.095051 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/88740fe4-c496-4ab9-a518-cfdaeb7fee36-hosts-file\") pod \"node-resolver-9fxjb\" (UID: \"88740fe4-c496-4ab9-a518-cfdaeb7fee36\") " pod="openshift-dns/node-resolver-9fxjb" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.097581 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.132156 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26cqh\" (UniqueName: \"kubernetes.io/projected/88740fe4-c496-4ab9-a518-cfdaeb7fee36-kube-api-access-26cqh\") pod \"node-resolver-9fxjb\" (UID: \"88740fe4-c496-4ab9-a518-cfdaeb7fee36\") " pod="openshift-dns/node-resolver-9fxjb" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.159800 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9fxjb" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.212713 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-ftrn8"] Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.213321 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-945lh"] Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.213487 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.213544 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:25 crc kubenswrapper[4775]: W1002 01:41:25.218740 4775 reflector.go:561] object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz": failed to list *v1.Secret: secrets "multus-ancillary-tools-dockercfg-vnmsz" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.218793 4775 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-vnmsz\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"multus-ancillary-tools-dockercfg-vnmsz\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 01:41:25 crc kubenswrapper[4775]: W1002 01:41:25.218748 4775 reflector.go:561] object-"openshift-machine-config-operator"/"proxy-tls": failed to list *v1.Secret: secrets "proxy-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Oct 02 01:41:25 crc kubenswrapper[4775]: W1002 01:41:25.218808 4775 reflector.go:561] object-"openshift-machine-config-operator"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Oct 02 01:41:25 crc kubenswrapper[4775]: W1002 01:41:25.218819 4775 reflector.go:561] object-"openshift-multus"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-multus": no relationship found between node 'crc' and this object Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.218845 4775 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.218848 4775 reflector.go:158] "Unhandled Error" err="object-\"openshift-multus\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-multus\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.218821 4775 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"proxy-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"proxy-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 01:41:25 crc kubenswrapper[4775]: W1002 01:41:25.218940 4775 reflector.go:561] object-"openshift-machine-config-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-config-operator": no relationship found between node 'crc' and this object Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.218977 4775 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-config-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.219362 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.219429 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.219643 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.219846 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.220790 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.228664 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-2xv98"] Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.229014 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.231005 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.233045 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.259221 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.285004 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.296917 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297050 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.297072 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:41:27.297053985 +0000 UTC m=+24.463798025 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297093 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-cnibin\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297135 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4ndm\" (UniqueName: \"kubernetes.io/projected/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-kube-api-access-m4ndm\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.297166 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297198 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.297213 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:27.297200989 +0000 UTC m=+24.463945029 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297232 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-system-cni-dir\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297258 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-cni-binary-copy\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297280 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297302 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c390c5b3-f65b-4f2e-9d03-bb09cf613134-rootfs\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297326 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxqq5\" (UniqueName: \"kubernetes.io/projected/c390c5b3-f65b-4f2e-9d03-bb09cf613134-kube-api-access-vxqq5\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297356 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297383 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c390c5b3-f65b-4f2e-9d03-bb09cf613134-mcd-auth-proxy-config\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297407 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-os-release\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297441 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297467 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c390c5b3-f65b-4f2e-9d03-bb09cf613134-proxy-tls\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.297492 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.297577 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.297591 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.297602 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.297582 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.297612 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.297618 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.297641 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:27.297631101 +0000 UTC m=+24.464375141 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.297673 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:27.297663801 +0000 UTC m=+24.464407841 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.297717 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:41:25 crc kubenswrapper[4775]: E1002 01:41:25.297742 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:27.297736333 +0000 UTC m=+24.464480373 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.299221 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.309606 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.318475 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.331768 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.346772 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.360809 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.368392 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.380483 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.397768 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.397845 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c390c5b3-f65b-4f2e-9d03-bb09cf613134-mcd-auth-proxy-config\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.397868 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-os-release\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.397894 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c390c5b3-f65b-4f2e-9d03-bb09cf613134-proxy-tls\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.397914 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-var-lib-cni-multus\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.397934 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.397975 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-run-k8s-cni-cncf-io\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.397995 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-var-lib-cni-bin\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398017 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-hostroot\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398032 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-multus-conf-dir\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398057 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c390c5b3-f65b-4f2e-9d03-bb09cf613134-rootfs\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398073 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-os-release\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398090 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ab7ffbf4-19df-4fff-aacc-344eab1d1089-cni-binary-copy\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398104 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-var-lib-kubelet\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398119 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ab7ffbf4-19df-4fff-aacc-344eab1d1089-multus-daemon-config\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398134 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-cnibin\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398148 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-multus-socket-dir-parent\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398162 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-multus-cni-dir\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398186 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-cnibin\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398200 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4ndm\" (UniqueName: \"kubernetes.io/projected/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-kube-api-access-m4ndm\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398216 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xw9j\" (UniqueName: \"kubernetes.io/projected/ab7ffbf4-19df-4fff-aacc-344eab1d1089-kube-api-access-4xw9j\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398238 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-system-cni-dir\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398256 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-run-multus-certs\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398271 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-cni-binary-copy\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398285 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398300 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-run-netns\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398314 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-etc-kubernetes\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398330 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxqq5\" (UniqueName: \"kubernetes.io/projected/c390c5b3-f65b-4f2e-9d03-bb09cf613134-kube-api-access-vxqq5\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398345 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-system-cni-dir\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398415 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-cnibin\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398215 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-os-release\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398592 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-system-cni-dir\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398790 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.398851 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c390c5b3-f65b-4f2e-9d03-bb09cf613134-rootfs\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.399022 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-cni-binary-copy\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.409025 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.434190 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.453568 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.464630 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.489372 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499656 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-var-lib-cni-multus\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499698 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-run-k8s-cni-cncf-io\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499718 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-var-lib-cni-bin\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499742 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-hostroot\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499758 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-multus-conf-dir\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499780 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ab7ffbf4-19df-4fff-aacc-344eab1d1089-cni-binary-copy\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499795 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-var-lib-kubelet\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499809 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ab7ffbf4-19df-4fff-aacc-344eab1d1089-multus-daemon-config\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499822 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-os-release\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499836 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-cnibin\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499850 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-multus-socket-dir-parent\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499868 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-multus-cni-dir\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499856 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-var-lib-cni-multus\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499906 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xw9j\" (UniqueName: \"kubernetes.io/projected/ab7ffbf4-19df-4fff-aacc-344eab1d1089-kube-api-access-4xw9j\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499923 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-run-multus-certs\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499943 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-run-netns\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499974 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-etc-kubernetes\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499981 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-var-lib-kubelet\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.499993 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-system-cni-dir\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.500014 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-run-k8s-cni-cncf-io\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.500045 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-var-lib-cni-bin\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.500078 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-hostroot\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.500111 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-multus-conf-dir\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.500139 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-system-cni-dir\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.500378 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-cnibin\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.500402 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-multus-socket-dir-parent\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.500440 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-run-multus-certs\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.500405 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-multus-cni-dir\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.500483 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-host-run-netns\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.500510 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-etc-kubernetes\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.500524 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ab7ffbf4-19df-4fff-aacc-344eab1d1089-os-release\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.500763 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ab7ffbf4-19df-4fff-aacc-344eab1d1089-multus-daemon-config\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.501069 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ab7ffbf4-19df-4fff-aacc-344eab1d1089-cni-binary-copy\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.502053 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.514062 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.526756 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.540832 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.553393 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.566466 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.625516 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zbgms"] Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.626295 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.629924 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.629996 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.630007 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.630142 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.630156 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.632453 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.635429 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.643619 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.658203 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.675966 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.691201 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.702647 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-openvswitch\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.702691 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-node-log\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.702710 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-slash\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.702730 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-systemd-units\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.702751 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7178a2dd-9182-400a-959d-c0c4181b6f18-ovn-node-metrics-cert\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.702784 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-env-overrides\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.702880 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-kubelet\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.702974 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-var-lib-openvswitch\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.703005 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-ovn\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.703027 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-cni-bin\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.703046 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.703156 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27cwc\" (UniqueName: \"kubernetes.io/projected/7178a2dd-9182-400a-959d-c0c4181b6f18-kube-api-access-27cwc\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.703327 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-log-socket\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.703446 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-ovnkube-config\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.703585 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-systemd\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.703652 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-etc-openvswitch\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.703727 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-cni-netd\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.703802 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-ovnkube-script-lib\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.703983 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-run-netns\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.704061 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-run-ovn-kubernetes\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.704113 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.720727 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.736517 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.759135 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.780261 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.796150 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805056 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-systemd\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805090 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-etc-openvswitch\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805119 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-cni-netd\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805137 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-ovnkube-script-lib\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805175 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-run-netns\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805206 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-run-ovn-kubernetes\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805230 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805259 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-openvswitch\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805248 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-systemd\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805299 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-cni-netd\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805328 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-run-ovn-kubernetes\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805625 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-run-netns\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805332 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-node-log\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805847 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805923 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-openvswitch\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805283 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-etc-openvswitch\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.805280 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-node-log\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.806406 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-ovnkube-script-lib\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.806464 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-systemd-units\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.806418 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-systemd-units\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.806570 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-slash\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.806664 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-env-overrides\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.806712 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7178a2dd-9182-400a-959d-c0c4181b6f18-ovn-node-metrics-cert\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.806754 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-kubelet\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.806803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-var-lib-openvswitch\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.806857 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-ovn\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.806905 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-cni-bin\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.806981 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27cwc\" (UniqueName: \"kubernetes.io/projected/7178a2dd-9182-400a-959d-c0c4181b6f18-kube-api-access-27cwc\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.807076 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-log-socket\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.807116 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-kubelet\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.807132 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-ovnkube-config\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.807187 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-slash\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.807604 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-var-lib-openvswitch\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.807671 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-ovn\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.807736 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-cni-bin\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.808364 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-ovnkube-config\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.808437 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-log-socket\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.808846 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-env-overrides\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.813243 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7178a2dd-9182-400a-959d-c0c4181b6f18-ovn-node-metrics-cert\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.826898 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.837228 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27cwc\" (UniqueName: \"kubernetes.io/projected/7178a2dd-9182-400a-959d-c0c4181b6f18-kube-api-access-27cwc\") pod \"ovnkube-node-zbgms\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.843110 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.901654 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9fxjb" event={"ID":"88740fe4-c496-4ab9-a518-cfdaeb7fee36","Type":"ContainerStarted","Data":"3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e"} Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.901737 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9fxjb" event={"ID":"88740fe4-c496-4ab9-a518-cfdaeb7fee36","Type":"ContainerStarted","Data":"da9a719bd6023d9284ab878dd40f089b04f2ecf1d565e4fa77f939706a2fa608"} Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.918206 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.935724 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.941887 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:25 crc kubenswrapper[4775]: W1002 01:41:25.957117 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7178a2dd_9182_400a_959d_c0c4181b6f18.slice/crio-96ee9704f7a21536812abcf8be7381ee439a206a18be456596d53a9a8bde4dee WatchSource:0}: Error finding container 96ee9704f7a21536812abcf8be7381ee439a206a18be456596d53a9a8bde4dee: Status 404 returned error can't find the container with id 96ee9704f7a21536812abcf8be7381ee439a206a18be456596d53a9a8bde4dee Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.958848 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.979822 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:25 crc kubenswrapper[4775]: I1002 01:41:25.997807 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:25Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.016187 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.031577 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.045277 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.067172 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.082105 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.098441 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.116128 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.359027 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.366395 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xw9j\" (UniqueName: \"kubernetes.io/projected/ab7ffbf4-19df-4fff-aacc-344eab1d1089-kube-api-access-4xw9j\") pod \"multus-2xv98\" (UID: \"ab7ffbf4-19df-4fff-aacc-344eab1d1089\") " pod="openshift-multus/multus-2xv98" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.380938 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4ndm\" (UniqueName: \"kubernetes.io/projected/9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a-kube-api-access-m4ndm\") pod \"multus-additional-cni-plugins-ftrn8\" (UID: \"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\") " pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:26 crc kubenswrapper[4775]: E1002 01:41:26.398099 4775 secret.go:188] Couldn't get secret openshift-machine-config-operator/proxy-tls: failed to sync secret cache: timed out waiting for the condition Oct 02 01:41:26 crc kubenswrapper[4775]: E1002 01:41:26.398205 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c390c5b3-f65b-4f2e-9d03-bb09cf613134-proxy-tls podName:c390c5b3-f65b-4f2e-9d03-bb09cf613134 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:26.898179883 +0000 UTC m=+24.064923963 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/c390c5b3-f65b-4f2e-9d03-bb09cf613134-proxy-tls") pod "machine-config-daemon-945lh" (UID: "c390c5b3-f65b-4f2e-9d03-bb09cf613134") : failed to sync secret cache: timed out waiting for the condition Oct 02 01:41:26 crc kubenswrapper[4775]: E1002 01:41:26.398098 4775 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Oct 02 01:41:26 crc kubenswrapper[4775]: E1002 01:41:26.398353 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c390c5b3-f65b-4f2e-9d03-bb09cf613134-mcd-auth-proxy-config podName:c390c5b3-f65b-4f2e-9d03-bb09cf613134 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:26.898322227 +0000 UTC m=+24.065066307 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "mcd-auth-proxy-config" (UniqueName: "kubernetes.io/configmap/c390c5b3-f65b-4f2e-9d03-bb09cf613134-mcd-auth-proxy-config") pod "machine-config-daemon-945lh" (UID: "c390c5b3-f65b-4f2e-9d03-bb09cf613134") : failed to sync configmap cache: timed out waiting for the condition Oct 02 01:41:26 crc kubenswrapper[4775]: E1002 01:41:26.425789 4775 projected.go:288] Couldn't get configMap openshift-machine-config-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 02 01:41:26 crc kubenswrapper[4775]: E1002 01:41:26.425839 4775 projected.go:194] Error preparing data for projected volume kube-api-access-vxqq5 for pod openshift-machine-config-operator/machine-config-daemon-945lh: failed to sync configmap cache: timed out waiting for the condition Oct 02 01:41:26 crc kubenswrapper[4775]: E1002 01:41:26.425920 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c390c5b3-f65b-4f2e-9d03-bb09cf613134-kube-api-access-vxqq5 podName:c390c5b3-f65b-4f2e-9d03-bb09cf613134 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:26.925896466 +0000 UTC m=+24.092640536 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-vxqq5" (UniqueName: "kubernetes.io/projected/c390c5b3-f65b-4f2e-9d03-bb09cf613134-kube-api-access-vxqq5") pod "machine-config-daemon-945lh" (UID: "c390c5b3-f65b-4f2e-9d03-bb09cf613134") : failed to sync configmap cache: timed out waiting for the condition Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.435268 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.438531 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.456465 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-2xv98" Oct 02 01:41:26 crc kubenswrapper[4775]: W1002 01:41:26.457586 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d24b3c6_c6ef_4a3c_aa18_635b09af6d7a.slice/crio-4dee49a7f375ca121abe25c34bae32115c5563999e27a497e0e0ffffdbaed9c9 WatchSource:0}: Error finding container 4dee49a7f375ca121abe25c34bae32115c5563999e27a497e0e0ffffdbaed9c9: Status 404 returned error can't find the container with id 4dee49a7f375ca121abe25c34bae32115c5563999e27a497e0e0ffffdbaed9c9 Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.610134 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.623185 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.765013 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:26 crc kubenswrapper[4775]: E1002 01:41:26.765360 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.765463 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.765867 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:26 crc kubenswrapper[4775]: E1002 01:41:26.765884 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:26 crc kubenswrapper[4775]: E1002 01:41:26.765945 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.769330 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.907246 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2"} Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.909115 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" event={"ID":"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a","Type":"ContainerStarted","Data":"3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b"} Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.909477 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" event={"ID":"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a","Type":"ContainerStarted","Data":"4dee49a7f375ca121abe25c34bae32115c5563999e27a497e0e0ffffdbaed9c9"} Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.911496 4775 generic.go:334] "Generic (PLEG): container finished" podID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerID="a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352" exitCode=0 Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.911568 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerDied","Data":"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352"} Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.911621 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerStarted","Data":"96ee9704f7a21536812abcf8be7381ee439a206a18be456596d53a9a8bde4dee"} Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.913285 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2xv98" event={"ID":"ab7ffbf4-19df-4fff-aacc-344eab1d1089","Type":"ContainerStarted","Data":"30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a"} Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.913392 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2xv98" event={"ID":"ab7ffbf4-19df-4fff-aacc-344eab1d1089","Type":"ContainerStarted","Data":"758ebaff021999ab35eaa8e165e5fdc2a4886fa900bdfd05bd13195f410b318c"} Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.917698 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c390c5b3-f65b-4f2e-9d03-bb09cf613134-mcd-auth-proxy-config\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.917781 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c390c5b3-f65b-4f2e-9d03-bb09cf613134-proxy-tls\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.921103 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c390c5b3-f65b-4f2e-9d03-bb09cf613134-mcd-auth-proxy-config\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.921834 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c390c5b3-f65b-4f2e-9d03-bb09cf613134-proxy-tls\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.940488 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.952419 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.979040 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:26 crc kubenswrapper[4775]: I1002 01:41:26.989827 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.005352 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.018832 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxqq5\" (UniqueName: \"kubernetes.io/projected/c390c5b3-f65b-4f2e-9d03-bb09cf613134-kube-api-access-vxqq5\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.023858 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxqq5\" (UniqueName: \"kubernetes.io/projected/c390c5b3-f65b-4f2e-9d03-bb09cf613134-kube-api-access-vxqq5\") pod \"machine-config-daemon-945lh\" (UID: \"c390c5b3-f65b-4f2e-9d03-bb09cf613134\") " pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.026646 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.036218 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.043904 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: W1002 01:41:27.049602 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc390c5b3_f65b_4f2e_9d03_bb09cf613134.slice/crio-2b16b4c7e43092784a4e0bace9b7f9cc42289c66b39964b16e8186a34bdef0a5 WatchSource:0}: Error finding container 2b16b4c7e43092784a4e0bace9b7f9cc42289c66b39964b16e8186a34bdef0a5: Status 404 returned error can't find the container with id 2b16b4c7e43092784a4e0bace9b7f9cc42289c66b39964b16e8186a34bdef0a5 Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.069880 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.092933 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.117461 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.163775 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.184304 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.199347 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.219579 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.233449 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.251342 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.273444 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.287698 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.304401 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.318852 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.321869 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.322004 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.322050 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.322094 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.322137 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:27 crc kubenswrapper[4775]: E1002 01:41:27.322239 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:41:27 crc kubenswrapper[4775]: E1002 01:41:27.322291 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:31.322274955 +0000 UTC m=+28.489019015 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:41:27 crc kubenswrapper[4775]: E1002 01:41:27.322362 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:41:31.322354397 +0000 UTC m=+28.489098447 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:41:27 crc kubenswrapper[4775]: E1002 01:41:27.322452 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:41:27 crc kubenswrapper[4775]: E1002 01:41:27.322474 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:41:27 crc kubenswrapper[4775]: E1002 01:41:27.322487 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:27 crc kubenswrapper[4775]: E1002 01:41:27.322516 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:31.322508211 +0000 UTC m=+28.489252261 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:27 crc kubenswrapper[4775]: E1002 01:41:27.322566 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:41:27 crc kubenswrapper[4775]: E1002 01:41:27.322593 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:31.322584873 +0000 UTC m=+28.489328923 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:41:27 crc kubenswrapper[4775]: E1002 01:41:27.322643 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:41:27 crc kubenswrapper[4775]: E1002 01:41:27.322662 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:41:27 crc kubenswrapper[4775]: E1002 01:41:27.322671 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:27 crc kubenswrapper[4775]: E1002 01:41:27.322707 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:31.322688606 +0000 UTC m=+28.489432656 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.330340 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.343040 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.361859 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.385507 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.549495 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-zrblh"] Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.550023 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zrblh" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.552170 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.554472 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.555114 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.555763 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.572332 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.590433 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.606765 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.619287 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.624870 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/520ea1b5-f5b6-4de5-93dc-dadd513c4a33-serviceca\") pod \"node-ca-zrblh\" (UID: \"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\") " pod="openshift-image-registry/node-ca-zrblh" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.624967 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wntq\" (UniqueName: \"kubernetes.io/projected/520ea1b5-f5b6-4de5-93dc-dadd513c4a33-kube-api-access-4wntq\") pod \"node-ca-zrblh\" (UID: \"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\") " pod="openshift-image-registry/node-ca-zrblh" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.625010 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/520ea1b5-f5b6-4de5-93dc-dadd513c4a33-host\") pod \"node-ca-zrblh\" (UID: \"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\") " pod="openshift-image-registry/node-ca-zrblh" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.630133 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.635810 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.642095 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.655046 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.671921 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.685608 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.704880 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.718977 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.725637 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wntq\" (UniqueName: \"kubernetes.io/projected/520ea1b5-f5b6-4de5-93dc-dadd513c4a33-kube-api-access-4wntq\") pod \"node-ca-zrblh\" (UID: \"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\") " pod="openshift-image-registry/node-ca-zrblh" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.725687 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/520ea1b5-f5b6-4de5-93dc-dadd513c4a33-host\") pod \"node-ca-zrblh\" (UID: \"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\") " pod="openshift-image-registry/node-ca-zrblh" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.725724 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/520ea1b5-f5b6-4de5-93dc-dadd513c4a33-serviceca\") pod \"node-ca-zrblh\" (UID: \"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\") " pod="openshift-image-registry/node-ca-zrblh" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.725837 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/520ea1b5-f5b6-4de5-93dc-dadd513c4a33-host\") pod \"node-ca-zrblh\" (UID: \"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\") " pod="openshift-image-registry/node-ca-zrblh" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.731825 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/520ea1b5-f5b6-4de5-93dc-dadd513c4a33-serviceca\") pod \"node-ca-zrblh\" (UID: \"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\") " pod="openshift-image-registry/node-ca-zrblh" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.733274 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.748330 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.749332 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wntq\" (UniqueName: \"kubernetes.io/projected/520ea1b5-f5b6-4de5-93dc-dadd513c4a33-kube-api-access-4wntq\") pod \"node-ca-zrblh\" (UID: \"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\") " pod="openshift-image-registry/node-ca-zrblh" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.760398 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.783710 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.797413 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.822146 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.856732 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.873986 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zrblh" Oct 02 01:41:27 crc kubenswrapper[4775]: W1002 01:41:27.891859 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod520ea1b5_f5b6_4de5_93dc_dadd513c4a33.slice/crio-b2fdab257e19434faa99b52e7ab4269d6bc7ee001b7a6a42bf93d183be569b37 WatchSource:0}: Error finding container b2fdab257e19434faa99b52e7ab4269d6bc7ee001b7a6a42bf93d183be569b37: Status 404 returned error can't find the container with id b2fdab257e19434faa99b52e7ab4269d6bc7ee001b7a6a42bf93d183be569b37 Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.901304 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.919451 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zrblh" event={"ID":"520ea1b5-f5b6-4de5-93dc-dadd513c4a33","Type":"ContainerStarted","Data":"b2fdab257e19434faa99b52e7ab4269d6bc7ee001b7a6a42bf93d183be569b37"} Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.920629 4775 generic.go:334] "Generic (PLEG): container finished" podID="9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a" containerID="3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b" exitCode=0 Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.920665 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" event={"ID":"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a","Type":"ContainerDied","Data":"3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b"} Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.922508 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78"} Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.922571 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb"} Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.922591 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"2b16b4c7e43092784a4e0bace9b7f9cc42289c66b39964b16e8186a34bdef0a5"} Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.926254 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerStarted","Data":"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352"} Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.926316 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerStarted","Data":"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780"} Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.926334 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerStarted","Data":"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc"} Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.926347 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerStarted","Data":"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654"} Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.926357 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerStarted","Data":"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf"} Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.939334 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.978854 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.979511 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.982893 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 02 01:41:27 crc kubenswrapper[4775]: I1002 01:41:27.994899 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.038295 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.086832 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.115906 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.159330 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.199777 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.241302 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.276262 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.317417 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.361852 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.402894 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.436866 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.483066 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.523159 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.563022 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.605786 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.643458 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.684229 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.728126 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.759539 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.764765 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.764855 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.764921 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:28 crc kubenswrapper[4775]: E1002 01:41:28.765031 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:28 crc kubenswrapper[4775]: E1002 01:41:28.765249 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:28 crc kubenswrapper[4775]: E1002 01:41:28.765398 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.802590 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.859807 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.877231 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.922660 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.933604 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zrblh" event={"ID":"520ea1b5-f5b6-4de5-93dc-dadd513c4a33","Type":"ContainerStarted","Data":"8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0"} Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.935847 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" event={"ID":"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a","Type":"ContainerStarted","Data":"89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9"} Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.941543 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerStarted","Data":"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b"} Oct 02 01:41:28 crc kubenswrapper[4775]: I1002 01:41:28.962236 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.007029 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.042646 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.084126 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.121597 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.162527 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.197841 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.240520 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.281788 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.321485 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.355549 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.394910 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.438242 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.481552 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.518227 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.560541 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.600085 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.645894 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.683871 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.719574 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.757365 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.797088 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.855273 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.881437 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.925837 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.950145 4775 generic.go:334] "Generic (PLEG): container finished" podID="9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a" containerID="89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9" exitCode=0 Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.950353 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" event={"ID":"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a","Type":"ContainerDied","Data":"89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9"} Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.964653 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:29Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.972693 4775 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.976049 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.976113 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.976140 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:29 crc kubenswrapper[4775]: I1002 01:41:29.976391 4775 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.003065 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.048329 4775 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.048628 4775 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.050595 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.050640 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.050657 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.050678 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.050698 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:30Z","lastTransitionTime":"2025-10-02T01:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:30 crc kubenswrapper[4775]: E1002 01:41:30.074563 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.078874 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.078901 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.078910 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.078926 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.078935 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:30Z","lastTransitionTime":"2025-10-02T01:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.079084 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: E1002 01:41:30.093871 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.107948 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.108035 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.108048 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.108073 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.108091 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:30Z","lastTransitionTime":"2025-10-02T01:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:30 crc kubenswrapper[4775]: E1002 01:41:30.122686 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.123133 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.127831 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.127876 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.127888 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.127939 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.127970 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:30Z","lastTransitionTime":"2025-10-02T01:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:30 crc kubenswrapper[4775]: E1002 01:41:30.139389 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.146312 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.146347 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.146355 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.146374 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.146384 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:30Z","lastTransitionTime":"2025-10-02T01:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.157457 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: E1002 01:41:30.158398 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: E1002 01:41:30.158507 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.160165 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.160198 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.160207 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.160228 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.160240 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:30Z","lastTransitionTime":"2025-10-02T01:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.196533 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.246761 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.263869 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.263920 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.263933 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.263975 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.263991 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:30Z","lastTransitionTime":"2025-10-02T01:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.284536 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.320887 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.360167 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.366931 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.366990 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.367006 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.367025 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.367038 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:30Z","lastTransitionTime":"2025-10-02T01:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.394234 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.436486 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.470539 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.470580 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.470593 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.470611 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.470625 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:30Z","lastTransitionTime":"2025-10-02T01:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.481928 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.527928 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.559639 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.574424 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.574494 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.574513 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.574541 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.574561 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:30Z","lastTransitionTime":"2025-10-02T01:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.602140 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.637127 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.677441 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.678006 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.678095 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.678125 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.678159 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.678187 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:30Z","lastTransitionTime":"2025-10-02T01:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.720116 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.759125 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.764353 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:30 crc kubenswrapper[4775]: E1002 01:41:30.764508 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.764639 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.764640 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:30 crc kubenswrapper[4775]: E1002 01:41:30.764904 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:30 crc kubenswrapper[4775]: E1002 01:41:30.765016 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.782227 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.782281 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.782293 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.782317 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.782331 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:30Z","lastTransitionTime":"2025-10-02T01:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.885173 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.885245 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.885330 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.885386 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.885461 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:30Z","lastTransitionTime":"2025-10-02T01:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.956255 4775 generic.go:334] "Generic (PLEG): container finished" podID="9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a" containerID="921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c" exitCode=0 Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.956321 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" event={"ID":"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a","Type":"ContainerDied","Data":"921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c"} Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.964558 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerStarted","Data":"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698"} Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.984753 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.990200 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.990248 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.990268 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.990296 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:30 crc kubenswrapper[4775]: I1002 01:41:30.990316 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:30Z","lastTransitionTime":"2025-10-02T01:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.000166 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:30Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.018755 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:31Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.035602 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:31Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.053654 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:31Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.078722 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:31Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.093307 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:31Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.093520 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.093545 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.093556 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.093573 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.093587 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:31Z","lastTransitionTime":"2025-10-02T01:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.107727 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:31Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.125230 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:31Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.160283 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:31Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.196753 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.196810 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.196821 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.196842 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.196857 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:31Z","lastTransitionTime":"2025-10-02T01:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.200676 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:31Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.239824 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:31Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.279267 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:31Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.300991 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.301078 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.301097 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.301125 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.301144 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:31Z","lastTransitionTime":"2025-10-02T01:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.320385 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:31Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.370206 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.370489 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:31 crc kubenswrapper[4775]: E1002 01:41:31.370567 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:41:39.370533676 +0000 UTC m=+36.537277746 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.370637 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:31 crc kubenswrapper[4775]: E1002 01:41:31.370690 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.370725 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:31 crc kubenswrapper[4775]: E1002 01:41:31.370819 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:39.370788443 +0000 UTC m=+36.537532683 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:41:31 crc kubenswrapper[4775]: E1002 01:41:31.370888 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:41:31 crc kubenswrapper[4775]: E1002 01:41:31.370920 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:41:31 crc kubenswrapper[4775]: E1002 01:41:31.370932 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:41:31 crc kubenswrapper[4775]: E1002 01:41:31.370973 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.370894 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:31 crc kubenswrapper[4775]: E1002 01:41:31.371020 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:39.370998669 +0000 UTC m=+36.537742949 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:41:31 crc kubenswrapper[4775]: E1002 01:41:31.371098 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:39.371079611 +0000 UTC m=+36.537823681 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:31 crc kubenswrapper[4775]: E1002 01:41:31.371040 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:41:31 crc kubenswrapper[4775]: E1002 01:41:31.371130 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:41:31 crc kubenswrapper[4775]: E1002 01:41:31.371145 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:31 crc kubenswrapper[4775]: E1002 01:41:31.371187 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:39.371174623 +0000 UTC m=+36.537918703 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.373627 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:31Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.403685 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.403750 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.403761 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.403779 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.403791 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:31Z","lastTransitionTime":"2025-10-02T01:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.507637 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.507718 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.507743 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.507777 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.507801 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:31Z","lastTransitionTime":"2025-10-02T01:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.611463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.611542 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.611562 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.611594 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.611617 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:31Z","lastTransitionTime":"2025-10-02T01:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.714998 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.715143 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.715168 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.715202 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.715225 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:31Z","lastTransitionTime":"2025-10-02T01:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.819170 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.819223 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.819241 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.819268 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.819286 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:31Z","lastTransitionTime":"2025-10-02T01:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.922234 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.922306 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.922329 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.922358 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.922382 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:31Z","lastTransitionTime":"2025-10-02T01:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.974648 4775 generic.go:334] "Generic (PLEG): container finished" podID="9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a" containerID="2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc" exitCode=0 Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.974712 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" event={"ID":"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a","Type":"ContainerDied","Data":"2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc"} Oct 02 01:41:31 crc kubenswrapper[4775]: I1002 01:41:31.996136 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:31Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.017567 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.025771 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.025839 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.025865 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.025895 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.025920 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:32Z","lastTransitionTime":"2025-10-02T01:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.051918 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.074827 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.097355 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.119614 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.129469 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.129553 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.129582 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.129617 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.129641 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:32Z","lastTransitionTime":"2025-10-02T01:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.146541 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.167361 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.187932 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.209209 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.226408 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.232594 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.232663 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.232683 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.232713 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.232749 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:32Z","lastTransitionTime":"2025-10-02T01:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.248308 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.265274 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.286823 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.312145 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.335604 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.335651 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.335669 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.335689 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.335703 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:32Z","lastTransitionTime":"2025-10-02T01:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.438765 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.438850 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.438868 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.438899 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.438921 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:32Z","lastTransitionTime":"2025-10-02T01:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.541473 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.541937 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.541991 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.542021 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.542039 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:32Z","lastTransitionTime":"2025-10-02T01:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.644890 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.645013 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.645041 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.645070 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.645093 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:32Z","lastTransitionTime":"2025-10-02T01:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.748658 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.748719 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.748742 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.748778 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.748803 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:32Z","lastTransitionTime":"2025-10-02T01:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.764530 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.764574 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:32 crc kubenswrapper[4775]: E1002 01:41:32.764698 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.764727 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:32 crc kubenswrapper[4775]: E1002 01:41:32.764903 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:32 crc kubenswrapper[4775]: E1002 01:41:32.765127 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.852200 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.852262 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.852285 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.852314 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.852337 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:32Z","lastTransitionTime":"2025-10-02T01:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.954991 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.955039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.955052 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.955078 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.955091 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:32Z","lastTransitionTime":"2025-10-02T01:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.983282 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" event={"ID":"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a","Type":"ContainerStarted","Data":"b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099"} Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.991274 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerStarted","Data":"3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c"} Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.991716 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.991778 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:32 crc kubenswrapper[4775]: I1002 01:41:32.991946 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.008063 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.022894 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.023418 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.028516 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.044016 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.058514 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.058582 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.058605 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.058632 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.058650 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:33Z","lastTransitionTime":"2025-10-02T01:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.071922 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.082416 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.099263 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.116638 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.137789 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.150112 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.161758 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.161799 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.161816 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.161842 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.161857 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:33Z","lastTransitionTime":"2025-10-02T01:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.166653 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.183408 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.199939 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.225836 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.241620 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.300285 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.300325 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.300340 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.300361 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.300376 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:33Z","lastTransitionTime":"2025-10-02T01:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.305478 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.324926 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.337032 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.366573 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.377693 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.388918 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.402005 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.402056 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.402065 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.402077 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.402086 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:33Z","lastTransitionTime":"2025-10-02T01:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.409392 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.427591 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.437446 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.449783 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.463051 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.475702 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.486909 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.498417 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.505121 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.505157 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.505168 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.505185 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.505197 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:33Z","lastTransitionTime":"2025-10-02T01:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.521985 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.536175 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.608477 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.608548 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.608567 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.608594 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.608613 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:33Z","lastTransitionTime":"2025-10-02T01:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.711135 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.711207 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.711229 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.711258 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.711279 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:33Z","lastTransitionTime":"2025-10-02T01:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.780983 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.801361 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.814112 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.814188 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.814212 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.814250 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.814275 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:33Z","lastTransitionTime":"2025-10-02T01:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.815542 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.835267 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.854393 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.873647 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.894492 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.916906 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.917008 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.917029 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.917050 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.917067 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:33Z","lastTransitionTime":"2025-10-02T01:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.918133 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.936723 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.953150 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.969340 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:33 crc kubenswrapper[4775]: I1002 01:41:33.998783 4775 generic.go:334] "Generic (PLEG): container finished" podID="9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a" containerID="b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099" exitCode=0 Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:33.998837 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" event={"ID":"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a","Type":"ContainerDied","Data":"b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099"} Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.005254 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.019406 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.019452 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.019473 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.019500 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.019518 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:34Z","lastTransitionTime":"2025-10-02T01:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.034288 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.047321 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.066304 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.081943 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.097466 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.122201 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.122247 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.122259 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.122275 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.122287 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:34Z","lastTransitionTime":"2025-10-02T01:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.123800 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.135285 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.148645 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.175214 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.195076 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.215071 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.226010 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.226088 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.226114 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.226146 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.226170 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:34Z","lastTransitionTime":"2025-10-02T01:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.232796 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.249921 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.268364 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.289879 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.304279 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.322385 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.329553 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.329615 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.329636 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.329661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.329679 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:34Z","lastTransitionTime":"2025-10-02T01:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.360000 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.433136 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.433213 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.433238 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.433272 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.433300 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:34Z","lastTransitionTime":"2025-10-02T01:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.538011 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.538065 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.538081 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.538105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.538121 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:34Z","lastTransitionTime":"2025-10-02T01:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.641548 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.641629 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.641655 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.641687 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.641709 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:34Z","lastTransitionTime":"2025-10-02T01:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.745094 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.745159 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.745177 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.745200 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.745218 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:34Z","lastTransitionTime":"2025-10-02T01:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.765148 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.765262 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:34 crc kubenswrapper[4775]: E1002 01:41:34.765362 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.765415 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:34 crc kubenswrapper[4775]: E1002 01:41:34.765583 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:34 crc kubenswrapper[4775]: E1002 01:41:34.765783 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.848653 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.848721 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.848739 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.848765 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.848785 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:34Z","lastTransitionTime":"2025-10-02T01:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.952658 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.952717 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.952738 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.952763 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:34 crc kubenswrapper[4775]: I1002 01:41:34.952781 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:34Z","lastTransitionTime":"2025-10-02T01:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.007448 4775 generic.go:334] "Generic (PLEG): container finished" podID="9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a" containerID="63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338" exitCode=0 Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.007531 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" event={"ID":"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a","Type":"ContainerDied","Data":"63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338"} Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.025204 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.063727 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.070076 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.070150 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.070237 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.070272 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.070295 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:35Z","lastTransitionTime":"2025-10-02T01:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.086747 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.117734 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.132866 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.153854 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.172983 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.173017 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.173029 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.173047 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.173060 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:35Z","lastTransitionTime":"2025-10-02T01:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.183446 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.208756 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.228654 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.245336 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.263767 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.276141 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.276205 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.276223 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.276247 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.276267 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:35Z","lastTransitionTime":"2025-10-02T01:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.283915 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.306942 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.320607 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.335349 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:35Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.378548 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.378807 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.379036 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.379206 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.379370 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:35Z","lastTransitionTime":"2025-10-02T01:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.482042 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.482106 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.482125 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.482151 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.482169 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:35Z","lastTransitionTime":"2025-10-02T01:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.585713 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.585752 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.585763 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.585780 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.585793 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:35Z","lastTransitionTime":"2025-10-02T01:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.689484 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.689550 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.689568 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.689594 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.689612 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:35Z","lastTransitionTime":"2025-10-02T01:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.791941 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.792031 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.792080 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.792108 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.792127 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:35Z","lastTransitionTime":"2025-10-02T01:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.895092 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.895165 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.895180 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.895214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.895231 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:35Z","lastTransitionTime":"2025-10-02T01:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.998449 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.998510 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.998526 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.998547 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:35 crc kubenswrapper[4775]: I1002 01:41:35.998559 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:35Z","lastTransitionTime":"2025-10-02T01:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.014276 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" event={"ID":"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a","Type":"ContainerStarted","Data":"fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5"} Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.028583 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.041772 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.096026 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.101556 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.101602 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.101621 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.101643 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.101658 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:36Z","lastTransitionTime":"2025-10-02T01:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.144773 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.156570 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.171992 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.186185 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.203350 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.203394 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.203409 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.203427 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.203441 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:36Z","lastTransitionTime":"2025-10-02T01:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.203616 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.215420 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.228568 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.243907 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.257568 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.267890 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.286111 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.301979 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:36Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.305874 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.305920 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.305931 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.305972 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.305984 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:36Z","lastTransitionTime":"2025-10-02T01:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.409555 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.409622 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.409645 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.409674 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.409697 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:36Z","lastTransitionTime":"2025-10-02T01:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.512832 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.512897 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.512917 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.512942 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.512986 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:36Z","lastTransitionTime":"2025-10-02T01:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.616442 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.616497 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.616514 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.616535 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.616552 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:36Z","lastTransitionTime":"2025-10-02T01:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.719801 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.719859 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.719909 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.719935 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.719998 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:36Z","lastTransitionTime":"2025-10-02T01:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.764832 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.764945 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:36 crc kubenswrapper[4775]: E1002 01:41:36.765051 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.765089 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:36 crc kubenswrapper[4775]: E1002 01:41:36.765273 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:36 crc kubenswrapper[4775]: E1002 01:41:36.765460 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.823021 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.823075 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.823092 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.823115 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.823133 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:36Z","lastTransitionTime":"2025-10-02T01:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.926562 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.926620 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.926638 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.926663 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:36 crc kubenswrapper[4775]: I1002 01:41:36.926680 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:36Z","lastTransitionTime":"2025-10-02T01:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.020515 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/0.log" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.024692 4775 generic.go:334] "Generic (PLEG): container finished" podID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerID="3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c" exitCode=1 Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.024787 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerDied","Data":"3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c"} Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.026340 4775 scope.go:117] "RemoveContainer" containerID="3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.028746 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.028801 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.028821 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.028847 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.028868 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:37Z","lastTransitionTime":"2025-10-02T01:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.060522 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.077476 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.097181 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.120843 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.131457 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.131635 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.131770 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.131905 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.132272 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:37Z","lastTransitionTime":"2025-10-02T01:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.149174 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.166833 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.189203 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.213121 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.232549 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.235131 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.235205 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.235372 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.235433 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.235536 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:37Z","lastTransitionTime":"2025-10-02T01:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.253249 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.272983 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.295654 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.309510 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.323523 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.339124 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.339170 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.339191 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.339215 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.339234 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:37Z","lastTransitionTime":"2025-10-02T01:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.361712 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"message\\\":\\\"7754 6051 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:35.987791 6051 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988116 6051 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.988472 6051 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:41:35.988683 6051 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988751 6051 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.988786 6051 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988913 6051 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.989516 6051 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.442044 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.442097 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.442113 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.442136 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.442154 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:37Z","lastTransitionTime":"2025-10-02T01:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.547759 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.547794 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.547804 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.547826 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.547840 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:37Z","lastTransitionTime":"2025-10-02T01:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.651182 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.651246 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.651259 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.651279 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.651291 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:37Z","lastTransitionTime":"2025-10-02T01:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.755162 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.755222 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.755239 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.755262 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.755280 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:37Z","lastTransitionTime":"2025-10-02T01:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.858508 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.858563 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.858580 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.858604 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.858620 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:37Z","lastTransitionTime":"2025-10-02T01:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.930724 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.951106 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.962415 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.962488 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.962516 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.962551 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.962574 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:37Z","lastTransitionTime":"2025-10-02T01:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:37 crc kubenswrapper[4775]: I1002 01:41:37.971221 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:37Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.011434 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"message\\\":\\\"7754 6051 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:35.987791 6051 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988116 6051 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.988472 6051 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:41:35.988683 6051 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988751 6051 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.988786 6051 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988913 6051 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.989516 6051 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.033593 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/0.log" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.040152 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerStarted","Data":"58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46"} Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.040831 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.048527 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.065749 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.065812 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.065921 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.065946 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.065993 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:38Z","lastTransitionTime":"2025-10-02T01:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.067620 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.088394 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.105403 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.127423 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.143355 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.161559 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.169664 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.169734 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.169753 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.169779 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.169797 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:38Z","lastTransitionTime":"2025-10-02T01:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.183053 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.204421 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.224099 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.243220 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.262258 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.273686 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.273744 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.273763 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.273787 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.273807 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:38Z","lastTransitionTime":"2025-10-02T01:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.279051 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.315379 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"message\\\":\\\"7754 6051 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:35.987791 6051 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988116 6051 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.988472 6051 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:41:35.988683 6051 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988751 6051 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.988786 6051 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988913 6051 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.989516 6051 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.321303 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6"] Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.322031 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.324935 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.325203 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.345695 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.376869 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.376931 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.377313 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.377389 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.377410 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:38Z","lastTransitionTime":"2025-10-02T01:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.381004 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.395858 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.422817 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.445059 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.459456 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7ee8fecb-23fc-4714-a079-c37b55da5640-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6wmv6\" (UID: \"7ee8fecb-23fc-4714-a079-c37b55da5640\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.459549 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7ee8fecb-23fc-4714-a079-c37b55da5640-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6wmv6\" (UID: \"7ee8fecb-23fc-4714-a079-c37b55da5640\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.459671 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhg82\" (UniqueName: \"kubernetes.io/projected/7ee8fecb-23fc-4714-a079-c37b55da5640-kube-api-access-rhg82\") pod \"ovnkube-control-plane-749d76644c-6wmv6\" (UID: \"7ee8fecb-23fc-4714-a079-c37b55da5640\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.459761 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7ee8fecb-23fc-4714-a079-c37b55da5640-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6wmv6\" (UID: \"7ee8fecb-23fc-4714-a079-c37b55da5640\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.465302 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.481170 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.481221 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.481238 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.481262 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.481283 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:38Z","lastTransitionTime":"2025-10-02T01:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.482388 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.501215 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.519221 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.538659 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.559100 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.560532 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7ee8fecb-23fc-4714-a079-c37b55da5640-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6wmv6\" (UID: \"7ee8fecb-23fc-4714-a079-c37b55da5640\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.560590 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7ee8fecb-23fc-4714-a079-c37b55da5640-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6wmv6\" (UID: \"7ee8fecb-23fc-4714-a079-c37b55da5640\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.560642 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhg82\" (UniqueName: \"kubernetes.io/projected/7ee8fecb-23fc-4714-a079-c37b55da5640-kube-api-access-rhg82\") pod \"ovnkube-control-plane-749d76644c-6wmv6\" (UID: \"7ee8fecb-23fc-4714-a079-c37b55da5640\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.560735 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7ee8fecb-23fc-4714-a079-c37b55da5640-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6wmv6\" (UID: \"7ee8fecb-23fc-4714-a079-c37b55da5640\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.561670 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7ee8fecb-23fc-4714-a079-c37b55da5640-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6wmv6\" (UID: \"7ee8fecb-23fc-4714-a079-c37b55da5640\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.562151 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7ee8fecb-23fc-4714-a079-c37b55da5640-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6wmv6\" (UID: \"7ee8fecb-23fc-4714-a079-c37b55da5640\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.570179 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7ee8fecb-23fc-4714-a079-c37b55da5640-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6wmv6\" (UID: \"7ee8fecb-23fc-4714-a079-c37b55da5640\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.581792 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.583831 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.584042 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.584128 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.584209 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.584266 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:38Z","lastTransitionTime":"2025-10-02T01:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.589760 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhg82\" (UniqueName: \"kubernetes.io/projected/7ee8fecb-23fc-4714-a079-c37b55da5640-kube-api-access-rhg82\") pod \"ovnkube-control-plane-749d76644c-6wmv6\" (UID: \"7ee8fecb-23fc-4714-a079-c37b55da5640\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.600416 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.626767 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.642595 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.642924 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.672064 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: W1002 01:41:38.672616 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7ee8fecb_23fc_4714_a079_c37b55da5640.slice/crio-56258144ffd12d112871a90e30299c702cdebaee69c3ddb84d4a09d6e164c0e8 WatchSource:0}: Error finding container 56258144ffd12d112871a90e30299c702cdebaee69c3ddb84d4a09d6e164c0e8: Status 404 returned error can't find the container with id 56258144ffd12d112871a90e30299c702cdebaee69c3ddb84d4a09d6e164c0e8 Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.687434 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.687483 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.687502 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.687527 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.687545 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:38Z","lastTransitionTime":"2025-10-02T01:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.699587 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.718915 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.740389 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.758336 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.764228 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.764245 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.764373 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:38 crc kubenswrapper[4775]: E1002 01:41:38.764520 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:38 crc kubenswrapper[4775]: E1002 01:41:38.764606 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:38 crc kubenswrapper[4775]: E1002 01:41:38.764728 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.781051 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.790698 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.790748 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.790765 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.790786 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.790805 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:38Z","lastTransitionTime":"2025-10-02T01:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.801541 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.817753 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.833914 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.865139 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"message\\\":\\\"7754 6051 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:35.987791 6051 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988116 6051 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.988472 6051 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:41:35.988683 6051 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988751 6051 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.988786 6051 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988913 6051 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.989516 6051 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.876843 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.893883 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.896495 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.896551 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.896573 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.896602 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.896624 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:38Z","lastTransitionTime":"2025-10-02T01:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.909431 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.933308 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:38Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.999231 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.999286 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.999348 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.999373 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:38 crc kubenswrapper[4775]: I1002 01:41:38.999391 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:38Z","lastTransitionTime":"2025-10-02T01:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.047153 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" event={"ID":"7ee8fecb-23fc-4714-a079-c37b55da5640","Type":"ContainerStarted","Data":"7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2"} Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.048330 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" event={"ID":"7ee8fecb-23fc-4714-a079-c37b55da5640","Type":"ContainerStarted","Data":"56258144ffd12d112871a90e30299c702cdebaee69c3ddb84d4a09d6e164c0e8"} Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.054373 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/1.log" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.054991 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/0.log" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.057821 4775 generic.go:334] "Generic (PLEG): container finished" podID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerID="58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46" exitCode=1 Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.057872 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerDied","Data":"58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46"} Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.057915 4775 scope.go:117] "RemoveContainer" containerID="3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.058430 4775 scope.go:117] "RemoveContainer" containerID="58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46" Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.058625 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.074495 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.089934 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.101510 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.101546 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.101557 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.101575 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.101588 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:39Z","lastTransitionTime":"2025-10-02T01:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.110079 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"message\\\":\\\"7754 6051 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:35.987791 6051 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988116 6051 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.988472 6051 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:41:35.988683 6051 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988751 6051 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.988786 6051 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988913 6051 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.989516 6051 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"al_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 01:41:37.989602 6248 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:37.989717 6248 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:37.989744 6248 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 01:41:37.989760 6248 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:37.989769 6248 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:37.989751 6248 factory.go:656] Stopping watch factory\\\\nI1002 01:41:37.989846 6248 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 01:41:37.989890 6248 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:37.989923 6248 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 01:41:37.990037 6248 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.130660 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.141630 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.154139 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.166713 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.178800 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.190600 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.199099 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.205251 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.205291 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.205300 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.205315 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.205325 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:39Z","lastTransitionTime":"2025-10-02T01:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.213273 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.224549 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.234623 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.246559 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.259606 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.275411 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.308294 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.308332 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.308343 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.308357 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.308367 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:39Z","lastTransitionTime":"2025-10-02T01:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.412053 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.412110 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.412135 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.412165 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.412188 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:39Z","lastTransitionTime":"2025-10-02T01:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.436994 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-x7pdc"] Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.437670 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.437751 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.456494 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.470310 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.470806 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.470944 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.471046 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.471091 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.471145 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.471200 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.471219 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.471218 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.471259 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.471287 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.471308 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.471144 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:41:55.471096687 +0000 UTC m=+52.637840777 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.471399 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.471461 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:55.471438276 +0000 UTC m=+52.638182366 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.471476 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.471496 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:55.471482597 +0000 UTC m=+52.638226677 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.471527 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:55.471506498 +0000 UTC m=+52.638250618 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.471549 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:55.471539869 +0000 UTC m=+52.638284029 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.484170 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.498159 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.514801 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.514846 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.514859 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.514879 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.514893 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:39Z","lastTransitionTime":"2025-10-02T01:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.515813 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.537509 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.552620 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.567906 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.572432 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwwg9\" (UniqueName: \"kubernetes.io/projected/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-kube-api-access-rwwg9\") pod \"network-metrics-daemon-x7pdc\" (UID: \"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\") " pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.572474 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs\") pod \"network-metrics-daemon-x7pdc\" (UID: \"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\") " pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.594256 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3e5941fcfd2a544a5219e2de3afa4ca30f7ae79c06ab0bc2ad0633a551cc030c\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"message\\\":\\\"7754 6051 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:35.987791 6051 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988116 6051 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.988472 6051 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1002 01:41:35.988683 6051 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988751 6051 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.988786 6051 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:35.988913 6051 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1002 01:41:35.989516 6051 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"al_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 01:41:37.989602 6248 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:37.989717 6248 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:37.989744 6248 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 01:41:37.989760 6248 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:37.989769 6248 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:37.989751 6248 factory.go:656] Stopping watch factory\\\\nI1002 01:41:37.989846 6248 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 01:41:37.989890 6248 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:37.989923 6248 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 01:41:37.990037 6248 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.613973 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.618167 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.618219 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.618237 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.618274 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.618293 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:39Z","lastTransitionTime":"2025-10-02T01:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.628300 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.638866 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.660006 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.673597 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwwg9\" (UniqueName: \"kubernetes.io/projected/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-kube-api-access-rwwg9\") pod \"network-metrics-daemon-x7pdc\" (UID: \"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\") " pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.673680 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs\") pod \"network-metrics-daemon-x7pdc\" (UID: \"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\") " pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.673844 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:41:39 crc kubenswrapper[4775]: E1002 01:41:39.673910 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs podName:ef954ceb-7b89-42cf-83dd-2e64dc9e44c1 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:40.173887933 +0000 UTC m=+37.340632003 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs") pod "network-metrics-daemon-x7pdc" (UID: "ef954ceb-7b89-42cf-83dd-2e64dc9e44c1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.674688 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.687925 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.690640 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwwg9\" (UniqueName: \"kubernetes.io/projected/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-kube-api-access-rwwg9\") pod \"network-metrics-daemon-x7pdc\" (UID: \"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\") " pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.708612 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.721128 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.721193 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.721206 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.721225 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.721237 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:39Z","lastTransitionTime":"2025-10-02T01:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.729377 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:39Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.824249 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.824319 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.824337 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.824366 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.824392 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:39Z","lastTransitionTime":"2025-10-02T01:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.927288 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.927384 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.927410 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.927438 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:39 crc kubenswrapper[4775]: I1002 01:41:39.927454 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:39Z","lastTransitionTime":"2025-10-02T01:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.030381 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.030439 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.030456 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.030479 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.030496 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.066554 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/1.log" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.071999 4775 scope.go:117] "RemoveContainer" containerID="58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46" Oct 02 01:41:40 crc kubenswrapper[4775]: E1002 01:41:40.072471 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.073387 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" event={"ID":"7ee8fecb-23fc-4714-a079-c37b55da5640","Type":"ContainerStarted","Data":"c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.097694 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.122909 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.133382 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.133430 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.133450 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.133474 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.133495 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.139662 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.163053 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.179353 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs\") pod \"network-metrics-daemon-x7pdc\" (UID: \"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\") " pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:40 crc kubenswrapper[4775]: E1002 01:41:40.180493 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:41:40 crc kubenswrapper[4775]: E1002 01:41:40.180684 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs podName:ef954ceb-7b89-42cf-83dd-2e64dc9e44c1 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:41.180637468 +0000 UTC m=+38.347381608 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs") pod "network-metrics-daemon-x7pdc" (UID: "ef954ceb-7b89-42cf-83dd-2e64dc9e44c1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.186350 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.207934 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.228612 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.236156 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.236198 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.236215 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.236238 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.236256 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.248285 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.268919 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.288410 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.306195 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.336837 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"al_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 01:41:37.989602 6248 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:37.989717 6248 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:37.989744 6248 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 01:41:37.989760 6248 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:37.989769 6248 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:37.989751 6248 factory.go:656] Stopping watch factory\\\\nI1002 01:41:37.989846 6248 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 01:41:37.989890 6248 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:37.989923 6248 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 01:41:37.990037 6248 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.338842 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.338888 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.338903 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.338927 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.338943 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.373164 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.390613 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.411236 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.428634 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.441850 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.441899 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.441916 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.441938 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.441982 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.445237 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.445903 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.446010 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.446030 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.446058 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.446076 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.462122 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: E1002 01:41:40.466524 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.476005 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.476081 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.476105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.476134 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.476157 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.490740 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: E1002 01:41:40.497216 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.503692 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.503762 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.503783 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.503812 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.503833 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.513865 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: E1002 01:41:40.524388 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.529408 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.529492 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.529513 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.529540 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.529558 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.534584 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: E1002 01:41:40.548876 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.554086 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.554566 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.554819 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.555039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.555234 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.555368 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: E1002 01:41:40.574760 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: E1002 01:41:40.575026 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.575139 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.577607 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.577872 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.578048 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.578214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.578339 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.596094 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.614208 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.631733 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.649452 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.665457 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.681423 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.681506 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.681567 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.681593 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.681615 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.696984 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"al_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 01:41:37.989602 6248 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:37.989717 6248 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:37.989744 6248 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 01:41:37.989760 6248 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:37.989769 6248 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:37.989751 6248 factory.go:656] Stopping watch factory\\\\nI1002 01:41:37.989846 6248 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 01:41:37.989890 6248 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:37.989923 6248 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 01:41:37.990037 6248 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.715939 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.727795 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.740827 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.764984 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.765064 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:40 crc kubenswrapper[4775]: E1002 01:41:40.765111 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.765228 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:40 crc kubenswrapper[4775]: E1002 01:41:40.765268 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.765303 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:40 crc kubenswrapper[4775]: E1002 01:41:40.765413 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:40 crc kubenswrapper[4775]: E1002 01:41:40.765504 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.772046 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.784074 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.784106 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.784114 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.784128 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.784137 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.784988 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:40Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.887395 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.887450 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.887467 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.887489 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.887506 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.991187 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.991251 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.991276 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.991300 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:40 crc kubenswrapper[4775]: I1002 01:41:40.991317 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:40Z","lastTransitionTime":"2025-10-02T01:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.094472 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.094558 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.094581 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.094608 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.094627 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:41Z","lastTransitionTime":"2025-10-02T01:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.190418 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs\") pod \"network-metrics-daemon-x7pdc\" (UID: \"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\") " pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:41 crc kubenswrapper[4775]: E1002 01:41:41.190687 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:41:41 crc kubenswrapper[4775]: E1002 01:41:41.190820 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs podName:ef954ceb-7b89-42cf-83dd-2e64dc9e44c1 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:43.190782067 +0000 UTC m=+40.357526187 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs") pod "network-metrics-daemon-x7pdc" (UID: "ef954ceb-7b89-42cf-83dd-2e64dc9e44c1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.198642 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.198698 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.198720 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.198747 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.198771 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:41Z","lastTransitionTime":"2025-10-02T01:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.302262 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.302306 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.302325 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.302346 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.302378 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:41Z","lastTransitionTime":"2025-10-02T01:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.405446 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.405507 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.405525 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.405547 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.405563 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:41Z","lastTransitionTime":"2025-10-02T01:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.509165 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.509217 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.509236 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.509259 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.509277 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:41Z","lastTransitionTime":"2025-10-02T01:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.612472 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.612534 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.612556 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.612583 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.612604 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:41Z","lastTransitionTime":"2025-10-02T01:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.716668 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.716761 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.716855 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.716898 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.716924 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:41Z","lastTransitionTime":"2025-10-02T01:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.820647 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.820715 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.820732 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.820757 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.820779 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:41Z","lastTransitionTime":"2025-10-02T01:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.923669 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.923731 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.923751 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.923782 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:41 crc kubenswrapper[4775]: I1002 01:41:41.923806 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:41Z","lastTransitionTime":"2025-10-02T01:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.026315 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.026353 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.026363 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.026379 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.026388 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:42Z","lastTransitionTime":"2025-10-02T01:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.129555 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.129655 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.129678 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.129711 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.129740 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:42Z","lastTransitionTime":"2025-10-02T01:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.232006 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.232077 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.232091 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.232413 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.232454 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:42Z","lastTransitionTime":"2025-10-02T01:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.335617 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.335671 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.335682 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.335700 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.335710 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:42Z","lastTransitionTime":"2025-10-02T01:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.439583 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.439654 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.439664 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.439692 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.439704 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:42Z","lastTransitionTime":"2025-10-02T01:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.542278 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.542337 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.542353 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.542378 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.542396 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:42Z","lastTransitionTime":"2025-10-02T01:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.645317 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.645370 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.645382 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.645401 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.645414 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:42Z","lastTransitionTime":"2025-10-02T01:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.748479 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.748546 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.748576 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.748607 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.748635 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:42Z","lastTransitionTime":"2025-10-02T01:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.764877 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.764918 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.764879 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.765028 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:42 crc kubenswrapper[4775]: E1002 01:41:42.765066 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:42 crc kubenswrapper[4775]: E1002 01:41:42.765258 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:42 crc kubenswrapper[4775]: E1002 01:41:42.765376 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:41:42 crc kubenswrapper[4775]: E1002 01:41:42.765550 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.852428 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.852518 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.852581 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.852618 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.852691 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:42Z","lastTransitionTime":"2025-10-02T01:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.956174 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.956231 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.956250 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.956273 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:42 crc kubenswrapper[4775]: I1002 01:41:42.956290 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:42Z","lastTransitionTime":"2025-10-02T01:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.059658 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.059721 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.059747 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.059776 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.059800 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:43Z","lastTransitionTime":"2025-10-02T01:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.162608 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.162691 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.162715 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.162745 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.162770 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:43Z","lastTransitionTime":"2025-10-02T01:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.216488 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs\") pod \"network-metrics-daemon-x7pdc\" (UID: \"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\") " pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:43 crc kubenswrapper[4775]: E1002 01:41:43.216892 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:41:43 crc kubenswrapper[4775]: E1002 01:41:43.217107 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs podName:ef954ceb-7b89-42cf-83dd-2e64dc9e44c1 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:47.217064705 +0000 UTC m=+44.383808775 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs") pod "network-metrics-daemon-x7pdc" (UID: "ef954ceb-7b89-42cf-83dd-2e64dc9e44c1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.268626 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.268700 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.268726 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.268774 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.268801 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:43Z","lastTransitionTime":"2025-10-02T01:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.373566 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.373638 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.373660 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.373691 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.373712 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:43Z","lastTransitionTime":"2025-10-02T01:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.476436 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.476547 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.476573 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.476604 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.476625 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:43Z","lastTransitionTime":"2025-10-02T01:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.580125 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.580179 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.580196 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.580220 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.580241 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:43Z","lastTransitionTime":"2025-10-02T01:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.683999 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.684064 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.684080 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.684104 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.684121 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:43Z","lastTransitionTime":"2025-10-02T01:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.784514 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.786680 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.786743 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.786759 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.786786 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.786806 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:43Z","lastTransitionTime":"2025-10-02T01:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.804141 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.832941 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"al_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 01:41:37.989602 6248 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:37.989717 6248 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:37.989744 6248 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 01:41:37.989760 6248 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:37.989769 6248 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:37.989751 6248 factory.go:656] Stopping watch factory\\\\nI1002 01:41:37.989846 6248 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 01:41:37.989890 6248 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:37.989923 6248 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 01:41:37.990037 6248 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.850300 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.874536 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.889835 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.889895 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.889914 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.889938 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.889986 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:43Z","lastTransitionTime":"2025-10-02T01:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.893362 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.910122 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.944081 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.965643 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.977803 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.990451 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:43Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.991820 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.991847 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.991856 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.991870 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:43 crc kubenswrapper[4775]: I1002 01:41:43.991882 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:43Z","lastTransitionTime":"2025-10-02T01:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.015978 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.040194 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.057143 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.074185 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.089157 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.093919 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.094029 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.094049 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.094078 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.094097 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:44Z","lastTransitionTime":"2025-10-02T01:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.106834 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:44Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.198846 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.198912 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.198931 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.198987 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.199008 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:44Z","lastTransitionTime":"2025-10-02T01:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.302539 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.302620 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.302640 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.302669 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.302732 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:44Z","lastTransitionTime":"2025-10-02T01:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.406469 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.406529 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.406549 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.406578 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.406596 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:44Z","lastTransitionTime":"2025-10-02T01:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.509717 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.509761 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.509780 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.509803 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.509820 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:44Z","lastTransitionTime":"2025-10-02T01:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.613226 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.613270 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.613287 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.613310 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.613328 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:44Z","lastTransitionTime":"2025-10-02T01:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.716486 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.716562 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.716586 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.716618 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.716641 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:44Z","lastTransitionTime":"2025-10-02T01:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.765288 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.765353 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.765390 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.765811 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:44 crc kubenswrapper[4775]: E1002 01:41:44.765806 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:44 crc kubenswrapper[4775]: E1002 01:41:44.766254 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:41:44 crc kubenswrapper[4775]: E1002 01:41:44.766401 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:44 crc kubenswrapper[4775]: E1002 01:41:44.766581 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.820193 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.820238 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.820256 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.820279 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.820296 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:44Z","lastTransitionTime":"2025-10-02T01:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.923663 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.923731 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.923754 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.923784 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:44 crc kubenswrapper[4775]: I1002 01:41:44.923807 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:44Z","lastTransitionTime":"2025-10-02T01:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.027195 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.027538 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.027716 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.027922 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.028165 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:45Z","lastTransitionTime":"2025-10-02T01:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.131473 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.131550 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.131570 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.131596 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.131616 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:45Z","lastTransitionTime":"2025-10-02T01:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.235504 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.235585 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.235609 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.235636 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.235659 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:45Z","lastTransitionTime":"2025-10-02T01:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.339096 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.339225 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.339296 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.339324 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.339389 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:45Z","lastTransitionTime":"2025-10-02T01:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.443607 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.443684 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.443702 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.443732 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.443751 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:45Z","lastTransitionTime":"2025-10-02T01:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.546791 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.547489 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.547676 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.547820 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.547991 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:45Z","lastTransitionTime":"2025-10-02T01:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.651266 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.651353 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.651378 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.651413 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.651436 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:45Z","lastTransitionTime":"2025-10-02T01:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.754743 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.755149 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.755292 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.755422 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.755618 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:45Z","lastTransitionTime":"2025-10-02T01:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.859678 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.859757 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.859813 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.859841 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.859859 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:45Z","lastTransitionTime":"2025-10-02T01:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.962468 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.962520 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.962539 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.962565 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:45 crc kubenswrapper[4775]: I1002 01:41:45.962582 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:45Z","lastTransitionTime":"2025-10-02T01:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.065694 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.065789 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.065808 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.065839 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.065861 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:46Z","lastTransitionTime":"2025-10-02T01:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.168571 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.168636 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.168653 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.168677 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.168696 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:46Z","lastTransitionTime":"2025-10-02T01:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.271991 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.272056 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.272074 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.272098 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.272115 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:46Z","lastTransitionTime":"2025-10-02T01:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.375742 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.375806 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.375824 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.375852 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.375906 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:46Z","lastTransitionTime":"2025-10-02T01:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.479254 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.479330 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.479348 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.479370 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.479387 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:46Z","lastTransitionTime":"2025-10-02T01:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.582108 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.582169 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.582190 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.582219 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.582241 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:46Z","lastTransitionTime":"2025-10-02T01:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.686173 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.686234 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.686257 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.686287 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.686309 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:46Z","lastTransitionTime":"2025-10-02T01:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.764324 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.764340 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.764469 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.764528 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:46 crc kubenswrapper[4775]: E1002 01:41:46.764675 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:46 crc kubenswrapper[4775]: E1002 01:41:46.764798 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:46 crc kubenswrapper[4775]: E1002 01:41:46.764922 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:41:46 crc kubenswrapper[4775]: E1002 01:41:46.765080 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.790160 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.790214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.790232 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.790256 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.790273 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:46Z","lastTransitionTime":"2025-10-02T01:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.894345 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.894417 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.894442 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.894475 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.894496 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:46Z","lastTransitionTime":"2025-10-02T01:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.997362 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.997430 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.997452 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.997473 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:46 crc kubenswrapper[4775]: I1002 01:41:46.997491 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:46Z","lastTransitionTime":"2025-10-02T01:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.100691 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.100753 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.100775 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.100805 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.100828 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:47Z","lastTransitionTime":"2025-10-02T01:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.203562 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.203623 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.203664 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.203688 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.203705 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:47Z","lastTransitionTime":"2025-10-02T01:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.268457 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs\") pod \"network-metrics-daemon-x7pdc\" (UID: \"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\") " pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:47 crc kubenswrapper[4775]: E1002 01:41:47.268654 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:41:47 crc kubenswrapper[4775]: E1002 01:41:47.268768 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs podName:ef954ceb-7b89-42cf-83dd-2e64dc9e44c1 nodeName:}" failed. No retries permitted until 2025-10-02 01:41:55.268731208 +0000 UTC m=+52.435475288 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs") pod "network-metrics-daemon-x7pdc" (UID: "ef954ceb-7b89-42cf-83dd-2e64dc9e44c1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.307065 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.307133 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.307157 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.307184 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.307205 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:47Z","lastTransitionTime":"2025-10-02T01:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.410852 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.410923 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.410941 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.410993 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.411012 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:47Z","lastTransitionTime":"2025-10-02T01:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.513932 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.514010 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.514028 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.514051 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.514068 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:47Z","lastTransitionTime":"2025-10-02T01:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.616381 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.616437 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.616454 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.616479 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.616512 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:47Z","lastTransitionTime":"2025-10-02T01:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.719519 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.719624 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.719649 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.719683 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.719706 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:47Z","lastTransitionTime":"2025-10-02T01:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.822902 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.823000 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.823026 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.823049 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.823066 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:47Z","lastTransitionTime":"2025-10-02T01:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.926840 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.926913 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.926937 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.926993 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:47 crc kubenswrapper[4775]: I1002 01:41:47.927011 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:47Z","lastTransitionTime":"2025-10-02T01:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.029211 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.029272 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.029290 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.029313 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.029332 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:48Z","lastTransitionTime":"2025-10-02T01:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.131885 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.131947 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.131999 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.132025 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.132044 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:48Z","lastTransitionTime":"2025-10-02T01:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.235301 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.235364 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.235382 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.235406 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.235423 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:48Z","lastTransitionTime":"2025-10-02T01:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.338387 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.338463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.338486 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.338513 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.338537 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:48Z","lastTransitionTime":"2025-10-02T01:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.442642 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.442694 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.442706 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.442723 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.442737 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:48Z","lastTransitionTime":"2025-10-02T01:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.544702 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.544744 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.544754 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.544769 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.544782 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:48Z","lastTransitionTime":"2025-10-02T01:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.647331 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.647436 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.647452 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.647474 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.647490 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:48Z","lastTransitionTime":"2025-10-02T01:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.750017 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.750161 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.750192 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.750222 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.750247 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:48Z","lastTransitionTime":"2025-10-02T01:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.764430 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.764501 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:48 crc kubenswrapper[4775]: E1002 01:41:48.764625 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.764672 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.764767 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:48 crc kubenswrapper[4775]: E1002 01:41:48.764902 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:41:48 crc kubenswrapper[4775]: E1002 01:41:48.765024 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:48 crc kubenswrapper[4775]: E1002 01:41:48.765194 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.852435 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.852527 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.852546 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.852571 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.852588 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:48Z","lastTransitionTime":"2025-10-02T01:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.955527 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.955588 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.955605 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.955630 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:48 crc kubenswrapper[4775]: I1002 01:41:48.955647 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:48Z","lastTransitionTime":"2025-10-02T01:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.058749 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.058822 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.058839 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.058863 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.058880 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:49Z","lastTransitionTime":"2025-10-02T01:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.161777 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.161813 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.161825 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.161839 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.161850 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:49Z","lastTransitionTime":"2025-10-02T01:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.265560 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.265638 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.265658 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.265681 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.265700 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:49Z","lastTransitionTime":"2025-10-02T01:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.369782 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.369864 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.369889 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.369924 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.369948 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:49Z","lastTransitionTime":"2025-10-02T01:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.473905 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.473996 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.474014 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.474038 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.474060 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:49Z","lastTransitionTime":"2025-10-02T01:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.577713 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.577793 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.577811 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.577840 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.577864 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:49Z","lastTransitionTime":"2025-10-02T01:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.680574 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.680624 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.680642 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.680664 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.680679 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:49Z","lastTransitionTime":"2025-10-02T01:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.783186 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.783238 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.783254 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.783322 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.783340 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:49Z","lastTransitionTime":"2025-10-02T01:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.886702 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.887232 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.887277 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.887306 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.887326 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:49Z","lastTransitionTime":"2025-10-02T01:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.990165 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.990227 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.990243 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.990271 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:49 crc kubenswrapper[4775]: I1002 01:41:49.990290 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:49Z","lastTransitionTime":"2025-10-02T01:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.093002 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.093055 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.093064 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.093083 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.093095 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:50Z","lastTransitionTime":"2025-10-02T01:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.195780 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.195858 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.195877 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.195902 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.195924 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:50Z","lastTransitionTime":"2025-10-02T01:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.299078 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.299153 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.299178 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.299207 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.299230 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:50Z","lastTransitionTime":"2025-10-02T01:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.402180 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.402249 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.402263 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.402283 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.402300 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:50Z","lastTransitionTime":"2025-10-02T01:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.505145 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.505208 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.505226 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.505250 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.505267 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:50Z","lastTransitionTime":"2025-10-02T01:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.607644 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.607697 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.607714 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.607732 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.607744 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:50Z","lastTransitionTime":"2025-10-02T01:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.710297 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.710342 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.710354 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.710368 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.710379 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:50Z","lastTransitionTime":"2025-10-02T01:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.765080 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.765160 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.765193 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:50 crc kubenswrapper[4775]: E1002 01:41:50.765328 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.765426 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:50 crc kubenswrapper[4775]: E1002 01:41:50.765568 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:41:50 crc kubenswrapper[4775]: E1002 01:41:50.765664 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:50 crc kubenswrapper[4775]: E1002 01:41:50.765734 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.813307 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.813368 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.813389 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.813417 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.813442 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:50Z","lastTransitionTime":"2025-10-02T01:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.890436 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.890484 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.890500 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.893448 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.893504 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:50Z","lastTransitionTime":"2025-10-02T01:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:50 crc kubenswrapper[4775]: E1002 01:41:50.915384 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:50Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.922124 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.922173 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.922190 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.922214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.922232 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:50Z","lastTransitionTime":"2025-10-02T01:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:50 crc kubenswrapper[4775]: E1002 01:41:50.937215 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:50Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.941662 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.941746 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.941775 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.941806 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.941828 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:50Z","lastTransitionTime":"2025-10-02T01:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:50 crc kubenswrapper[4775]: E1002 01:41:50.966783 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:50Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.971012 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.971070 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.971091 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.971117 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.971136 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:50Z","lastTransitionTime":"2025-10-02T01:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:50 crc kubenswrapper[4775]: E1002 01:41:50.990559 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:50Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.995217 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.995255 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.995267 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.995285 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:50 crc kubenswrapper[4775]: I1002 01:41:50.995299 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:50Z","lastTransitionTime":"2025-10-02T01:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:51 crc kubenswrapper[4775]: E1002 01:41:51.011995 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:51Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:51 crc kubenswrapper[4775]: E1002 01:41:51.012159 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.013824 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.013874 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.013890 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.013913 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.014000 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:51Z","lastTransitionTime":"2025-10-02T01:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.116080 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.116137 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.116146 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.116168 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.116181 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:51Z","lastTransitionTime":"2025-10-02T01:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.219475 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.219545 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.219563 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.219586 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.219603 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:51Z","lastTransitionTime":"2025-10-02T01:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.321917 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.322021 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.322039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.322063 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.322080 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:51Z","lastTransitionTime":"2025-10-02T01:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.424661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.424701 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.424712 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.424729 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.424740 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:51Z","lastTransitionTime":"2025-10-02T01:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.528056 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.528125 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.528141 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.528163 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.528182 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:51Z","lastTransitionTime":"2025-10-02T01:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.630491 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.630539 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.630555 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.630579 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.630595 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:51Z","lastTransitionTime":"2025-10-02T01:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.733128 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.733187 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.733206 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.733234 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.733253 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:51Z","lastTransitionTime":"2025-10-02T01:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.838413 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.838489 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.838507 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.838535 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.838558 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:51Z","lastTransitionTime":"2025-10-02T01:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.941505 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.941588 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.941610 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.941637 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:51 crc kubenswrapper[4775]: I1002 01:41:51.941660 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:51Z","lastTransitionTime":"2025-10-02T01:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.045181 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.045253 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.045278 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.045309 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.045331 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:52Z","lastTransitionTime":"2025-10-02T01:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.151144 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.151195 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.151217 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.151245 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.151267 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:52Z","lastTransitionTime":"2025-10-02T01:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.255393 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.255492 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.255514 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.255539 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.255556 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:52Z","lastTransitionTime":"2025-10-02T01:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.358413 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.358443 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.358453 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.358467 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.358477 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:52Z","lastTransitionTime":"2025-10-02T01:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.466813 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.466916 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.466936 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.466997 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.467015 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:52Z","lastTransitionTime":"2025-10-02T01:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.569832 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.569882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.569894 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.569911 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.569922 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:52Z","lastTransitionTime":"2025-10-02T01:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.672915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.673010 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.673035 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.673060 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.673079 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:52Z","lastTransitionTime":"2025-10-02T01:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.764236 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.764389 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:52 crc kubenswrapper[4775]: E1002 01:41:52.764486 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.764570 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.764583 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:52 crc kubenswrapper[4775]: E1002 01:41:52.764904 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:41:52 crc kubenswrapper[4775]: E1002 01:41:52.765034 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:52 crc kubenswrapper[4775]: E1002 01:41:52.765410 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.766533 4775 scope.go:117] "RemoveContainer" containerID="58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.781374 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.781448 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.781469 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.781502 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.781526 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:52Z","lastTransitionTime":"2025-10-02T01:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.884924 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.884986 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.884996 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.885013 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.885024 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:52Z","lastTransitionTime":"2025-10-02T01:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.987633 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.987700 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.987718 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.987743 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:52 crc kubenswrapper[4775]: I1002 01:41:52.987766 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:52Z","lastTransitionTime":"2025-10-02T01:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.090474 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.090514 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.090526 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.090541 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.090552 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:53Z","lastTransitionTime":"2025-10-02T01:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.125134 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/1.log" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.128740 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerStarted","Data":"af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995"} Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.129289 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.144230 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.161095 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.193272 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.193305 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.193313 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.193327 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.193335 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:53Z","lastTransitionTime":"2025-10-02T01:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.198626 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.218602 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.241086 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.262455 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.291233 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.296039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.296076 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.296090 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.296111 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.296127 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:53Z","lastTransitionTime":"2025-10-02T01:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.326517 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.342310 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.358253 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.371448 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.385309 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.397868 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.397894 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.397902 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.397915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.397924 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:53Z","lastTransitionTime":"2025-10-02T01:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.402945 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.417668 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.430794 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.452390 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"al_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 01:41:37.989602 6248 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:37.989717 6248 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:37.989744 6248 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 01:41:37.989760 6248 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:37.989769 6248 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:37.989751 6248 factory.go:656] Stopping watch factory\\\\nI1002 01:41:37.989846 6248 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 01:41:37.989890 6248 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:37.989923 6248 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 01:41:37.990037 6248 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.465728 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.500663 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.500699 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.500711 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.500727 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.500736 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:53Z","lastTransitionTime":"2025-10-02T01:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.603337 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.603434 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.603458 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.603487 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.603508 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:53Z","lastTransitionTime":"2025-10-02T01:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.707291 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.707366 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.707390 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.707418 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.707438 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:53Z","lastTransitionTime":"2025-10-02T01:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.783249 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.797824 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.810758 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.810803 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.810815 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.810833 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.810845 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:53Z","lastTransitionTime":"2025-10-02T01:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.818430 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.867432 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.885311 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.901257 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.913473 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.913539 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.913560 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.913584 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.913603 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:53Z","lastTransitionTime":"2025-10-02T01:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.918347 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.933885 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.951327 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.964768 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.980117 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:53 crc kubenswrapper[4775]: I1002 01:41:53.998362 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"al_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 01:41:37.989602 6248 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:37.989717 6248 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:37.989744 6248 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 01:41:37.989760 6248 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:37.989769 6248 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:37.989751 6248 factory.go:656] Stopping watch factory\\\\nI1002 01:41:37.989846 6248 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 01:41:37.989890 6248 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:37.989923 6248 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 01:41:37.990037 6248 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:53Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.009126 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.015676 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.015697 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.015705 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.015721 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.015730 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:54Z","lastTransitionTime":"2025-10-02T01:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.027106 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.043204 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.058013 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.076775 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.118450 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.118497 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.118511 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.118532 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.118546 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:54Z","lastTransitionTime":"2025-10-02T01:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.132900 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/2.log" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.133882 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/1.log" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.137287 4775 generic.go:334] "Generic (PLEG): container finished" podID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerID="af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995" exitCode=1 Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.137327 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerDied","Data":"af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995"} Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.137364 4775 scope.go:117] "RemoveContainer" containerID="58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.138074 4775 scope.go:117] "RemoveContainer" containerID="af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995" Oct 02 01:41:54 crc kubenswrapper[4775]: E1002 01:41:54.138282 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.156619 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.177602 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.192509 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.206999 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.225350 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.225737 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.225768 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.225786 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.225803 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.225814 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:54Z","lastTransitionTime":"2025-10-02T01:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.241867 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.260569 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.279447 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.295508 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.316096 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.328718 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.329005 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.329239 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.329430 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.329591 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:54Z","lastTransitionTime":"2025-10-02T01:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.335337 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.362804 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"al_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 01:41:37.989602 6248 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:37.989717 6248 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:37.989744 6248 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 01:41:37.989760 6248 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:37.989769 6248 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:37.989751 6248 factory.go:656] Stopping watch factory\\\\nI1002 01:41:37.989846 6248 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 01:41:37.989890 6248 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:37.989923 6248 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 01:41:37.990037 6248 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:53Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:53.805456 6444 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 01:41:53.805551 6444 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:53.805561 6444 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:53.805591 6444 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 01:41:53.805602 6444 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 01:41:53.805613 6444 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 01:41:53.805630 6444 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:53.805641 6444 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:53.805647 6444 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:53.805652 6444 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:41:53.805671 6444 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 01:41:53.805689 6444 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 01:41:53.805711 6444 factory.go:656] Stopping watch factory\\\\nI1002 01:41:53.805719 6444 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 01:41:53.805727 6444 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.394168 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.412541 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.433379 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.433435 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.433447 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.433467 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.433488 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:54Z","lastTransitionTime":"2025-10-02T01:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.434673 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.452084 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.470320 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:54Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.536996 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.537071 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.537093 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.537122 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.537140 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:54Z","lastTransitionTime":"2025-10-02T01:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.640640 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.641025 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.641180 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.641322 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.641474 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:54Z","lastTransitionTime":"2025-10-02T01:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.745076 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.745138 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.745156 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.745179 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.745197 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:54Z","lastTransitionTime":"2025-10-02T01:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.765293 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.765332 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.765332 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:54 crc kubenswrapper[4775]: E1002 01:41:54.765455 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:54 crc kubenswrapper[4775]: E1002 01:41:54.765649 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:41:54 crc kubenswrapper[4775]: E1002 01:41:54.765773 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.766332 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:54 crc kubenswrapper[4775]: E1002 01:41:54.766702 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.849351 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.849799 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.850161 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.850414 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.850906 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:54Z","lastTransitionTime":"2025-10-02T01:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.954385 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.954455 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.954473 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.954497 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:54 crc kubenswrapper[4775]: I1002 01:41:54.954516 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:54Z","lastTransitionTime":"2025-10-02T01:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.042833 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.058476 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.059321 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.059375 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.059393 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.059416 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.059434 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:55Z","lastTransitionTime":"2025-10-02T01:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.071695 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.085651 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.101884 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.124227 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.143454 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.144088 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/2.log" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.149409 4775 scope.go:117] "RemoveContainer" containerID="af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995" Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.149621 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.163652 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.163702 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.163719 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.163739 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.163756 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:55Z","lastTransitionTime":"2025-10-02T01:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.164373 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.187799 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.206214 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.227841 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.248097 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.267305 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.267373 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.267395 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.267425 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.267345 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.267448 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:55Z","lastTransitionTime":"2025-10-02T01:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.306903 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58a8ae5442bff2a1e9d77aa33429f9d35cfaa8aa45ff66cff7aa598f4bc19b46\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"message\\\":\\\"al_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-kube-apiserver/apiserver]} name:Service_openshift-kube-apiserver/apiserver_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.93:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {d71b38eb-32af-4c0f-9490-7c317c111e3a}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1002 01:41:37.989602 6248 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:37.989717 6248 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:37.989744 6248 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1002 01:41:37.989760 6248 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:37.989769 6248 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:37.989751 6248 factory.go:656] Stopping watch factory\\\\nI1002 01:41:37.989846 6248 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1002 01:41:37.989890 6248 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:37.989923 6248 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1002 01:41:37.990037 6248 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:53Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:53.805456 6444 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 01:41:53.805551 6444 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:53.805561 6444 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:53.805591 6444 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 01:41:53.805602 6444 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 01:41:53.805613 6444 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 01:41:53.805630 6444 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:53.805641 6444 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:53.805647 6444 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:53.805652 6444 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:41:53.805671 6444 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 01:41:53.805689 6444 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 01:41:53.805711 6444 factory.go:656] Stopping watch factory\\\\nI1002 01:41:53.805719 6444 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 01:41:53.805727 6444 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.323610 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.343793 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.363128 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.368685 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs\") pod \"network-metrics-daemon-x7pdc\" (UID: \"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\") " pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.368923 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.369257 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs podName:ef954ceb-7b89-42cf-83dd-2e64dc9e44c1 nodeName:}" failed. No retries permitted until 2025-10-02 01:42:11.369213747 +0000 UTC m=+68.535957827 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs") pod "network-metrics-daemon-x7pdc" (UID: "ef954ceb-7b89-42cf-83dd-2e64dc9e44c1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.370725 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.370812 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.370840 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.370877 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.370903 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:55Z","lastTransitionTime":"2025-10-02T01:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.379313 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.415167 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.433427 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b00713a-180e-4e32-bdb6-cb34fd64e1ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ef520bf5435b6a70fb920599ee3b72d11062641a24a364340ced83ff966eeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a349484f3015205e6074e79651f9e3a242b3b2597b70e60231115430ee41cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b0204478719ee799c2b7cff276ce9514d66b27c9eafa538ceec0e6478c0bf47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.450297 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.466924 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.474245 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.474285 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.474301 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.474324 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.474342 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:55Z","lastTransitionTime":"2025-10-02T01:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.498601 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:53Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:53.805456 6444 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 01:41:53.805551 6444 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:53.805561 6444 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:53.805591 6444 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 01:41:53.805602 6444 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 01:41:53.805613 6444 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 01:41:53.805630 6444 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:53.805641 6444 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:53.805647 6444 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:53.805652 6444 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:41:53.805671 6444 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 01:41:53.805689 6444 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 01:41:53.805711 6444 factory.go:656] Stopping watch factory\\\\nI1002 01:41:53.805719 6444 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 01:41:53.805727 6444 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.530865 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.547788 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.569664 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.571618 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.571864 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.571998 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:42:27.571934921 +0000 UTC m=+84.738679001 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.572103 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.572170 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.572109 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.572221 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.572246 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.572257 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.572332 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:42:27.572301631 +0000 UTC m=+84.739045711 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.572378 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.572404 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.572176 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.572440 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.572501 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:42:27.572486256 +0000 UTC m=+84.739230336 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.572565 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:42:27.572522527 +0000 UTC m=+84.739266597 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.572418 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:55 crc kubenswrapper[4775]: E1002 01:41:55.572661 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:42:27.57264083 +0000 UTC m=+84.739385140 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.577055 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.577097 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.577113 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.577136 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.577153 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:55Z","lastTransitionTime":"2025-10-02T01:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.588209 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.606853 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.630303 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.654759 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.670912 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.680555 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.680627 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.680651 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.680679 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.680697 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:55Z","lastTransitionTime":"2025-10-02T01:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.692423 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.715988 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.736172 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.755313 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.776708 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.785068 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.785129 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.785148 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.785176 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.785197 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:55Z","lastTransitionTime":"2025-10-02T01:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.800303 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:41:55Z is after 2025-08-24T17:21:41Z" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.888621 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.888677 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.888696 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.888719 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.888735 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:55Z","lastTransitionTime":"2025-10-02T01:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.991778 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.991829 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.991848 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.991873 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:55 crc kubenswrapper[4775]: I1002 01:41:55.991895 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:55Z","lastTransitionTime":"2025-10-02T01:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.095477 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.095545 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.095568 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.095597 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.095617 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:56Z","lastTransitionTime":"2025-10-02T01:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.199036 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.199089 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.199107 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.199134 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.199156 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:56Z","lastTransitionTime":"2025-10-02T01:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.307103 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.307177 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.307195 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.307223 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.307241 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:56Z","lastTransitionTime":"2025-10-02T01:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.410615 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.410725 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.410747 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.410771 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.410790 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:56Z","lastTransitionTime":"2025-10-02T01:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.514237 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.514290 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.514307 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.514326 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.514340 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:56Z","lastTransitionTime":"2025-10-02T01:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.617777 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.617850 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.617862 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.617884 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.617896 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:56Z","lastTransitionTime":"2025-10-02T01:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.720931 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.721030 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.721048 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.721075 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.721095 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:56Z","lastTransitionTime":"2025-10-02T01:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.764611 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.764674 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.764736 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:56 crc kubenswrapper[4775]: E1002 01:41:56.764844 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:56 crc kubenswrapper[4775]: E1002 01:41:56.764890 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:56 crc kubenswrapper[4775]: E1002 01:41:56.765050 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.765265 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:56 crc kubenswrapper[4775]: E1002 01:41:56.765508 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.824764 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.824828 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.824847 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.824903 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.824923 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:56Z","lastTransitionTime":"2025-10-02T01:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.929113 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.929168 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.929179 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.929202 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:56 crc kubenswrapper[4775]: I1002 01:41:56.929217 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:56Z","lastTransitionTime":"2025-10-02T01:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.032291 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.032361 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.032382 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.032408 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.032426 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:57Z","lastTransitionTime":"2025-10-02T01:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.135471 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.135547 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.135566 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.135593 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.135612 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:57Z","lastTransitionTime":"2025-10-02T01:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.238286 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.238353 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.238372 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.238395 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.238414 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:57Z","lastTransitionTime":"2025-10-02T01:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.341188 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.341250 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.341268 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.341293 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.341310 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:57Z","lastTransitionTime":"2025-10-02T01:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.444378 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.444424 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.444434 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.444451 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.444461 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:57Z","lastTransitionTime":"2025-10-02T01:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.547796 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.547875 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.547897 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.547929 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.547982 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:57Z","lastTransitionTime":"2025-10-02T01:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.650876 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.651152 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.651190 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.651221 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.651241 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:57Z","lastTransitionTime":"2025-10-02T01:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.754504 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.754554 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.754571 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.754594 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.754613 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:57Z","lastTransitionTime":"2025-10-02T01:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.857520 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.857568 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.857586 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.857607 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.857627 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:57Z","lastTransitionTime":"2025-10-02T01:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.960475 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.960522 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.960531 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.960547 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:57 crc kubenswrapper[4775]: I1002 01:41:57.960572 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:57Z","lastTransitionTime":"2025-10-02T01:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.063471 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.063529 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.063551 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.063580 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.063606 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:58Z","lastTransitionTime":"2025-10-02T01:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.166379 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.166423 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.166437 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.166456 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.166472 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:58Z","lastTransitionTime":"2025-10-02T01:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.269687 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.269769 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.269798 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.269831 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.269855 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:58Z","lastTransitionTime":"2025-10-02T01:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.372721 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.372794 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.372813 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.372841 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.372861 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:58Z","lastTransitionTime":"2025-10-02T01:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.476131 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.476225 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.476243 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.476297 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.476315 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:58Z","lastTransitionTime":"2025-10-02T01:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.579374 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.579434 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.579470 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.579499 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.579520 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:58Z","lastTransitionTime":"2025-10-02T01:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.683066 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.683132 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.683152 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.683177 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.683197 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:58Z","lastTransitionTime":"2025-10-02T01:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.764868 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.764918 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.765025 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:41:58 crc kubenswrapper[4775]: E1002 01:41:58.765110 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.765229 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:41:58 crc kubenswrapper[4775]: E1002 01:41:58.765236 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:41:58 crc kubenswrapper[4775]: E1002 01:41:58.765311 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:41:58 crc kubenswrapper[4775]: E1002 01:41:58.765428 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.785545 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.785591 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.785608 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.785629 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.785648 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:58Z","lastTransitionTime":"2025-10-02T01:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.888811 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.888868 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.888886 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.888911 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.888928 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:58Z","lastTransitionTime":"2025-10-02T01:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.992351 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.992418 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.992439 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.992464 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:58 crc kubenswrapper[4775]: I1002 01:41:58.992483 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:58Z","lastTransitionTime":"2025-10-02T01:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.095683 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.095731 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.095772 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.095795 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.095810 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:59Z","lastTransitionTime":"2025-10-02T01:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.198156 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.198221 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.198240 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.198263 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.198282 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:59Z","lastTransitionTime":"2025-10-02T01:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.300938 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.301013 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.301061 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.301088 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.301106 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:59Z","lastTransitionTime":"2025-10-02T01:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.404738 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.404801 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.404818 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.404844 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.404866 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:59Z","lastTransitionTime":"2025-10-02T01:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.507200 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.507256 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.507272 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.507293 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.507311 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:59Z","lastTransitionTime":"2025-10-02T01:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.610041 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.610105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.610130 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.610157 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.610177 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:59Z","lastTransitionTime":"2025-10-02T01:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.713418 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.713494 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.713517 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.713548 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.713570 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:59Z","lastTransitionTime":"2025-10-02T01:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.816419 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.816481 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.816498 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.816522 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.816540 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:59Z","lastTransitionTime":"2025-10-02T01:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.919917 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.920044 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.920064 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.920094 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:41:59 crc kubenswrapper[4775]: I1002 01:41:59.920111 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:41:59Z","lastTransitionTime":"2025-10-02T01:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.023559 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.023640 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.023666 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.023700 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.023724 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:00Z","lastTransitionTime":"2025-10-02T01:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.126695 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.126928 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.126981 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.127009 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.127028 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:00Z","lastTransitionTime":"2025-10-02T01:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.229535 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.229608 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.229627 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.229651 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.229669 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:00Z","lastTransitionTime":"2025-10-02T01:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.333375 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.333444 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.333463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.333486 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.333503 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:00Z","lastTransitionTime":"2025-10-02T01:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.436725 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.436812 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.436830 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.436855 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.436872 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:00Z","lastTransitionTime":"2025-10-02T01:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.540040 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.540079 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.540088 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.540105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.540115 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:00Z","lastTransitionTime":"2025-10-02T01:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.642422 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.642521 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.642546 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.642575 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.642598 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:00Z","lastTransitionTime":"2025-10-02T01:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.745835 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.745897 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.745914 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.745941 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.745986 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:00Z","lastTransitionTime":"2025-10-02T01:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.765186 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.765291 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.765227 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.765198 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:00 crc kubenswrapper[4775]: E1002 01:42:00.765424 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:00 crc kubenswrapper[4775]: E1002 01:42:00.765519 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:00 crc kubenswrapper[4775]: E1002 01:42:00.765730 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:00 crc kubenswrapper[4775]: E1002 01:42:00.766035 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.848328 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.848375 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.848386 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.848401 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.848439 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:00Z","lastTransitionTime":"2025-10-02T01:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.951167 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.951224 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.951242 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.951265 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:00 crc kubenswrapper[4775]: I1002 01:42:00.951282 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:00Z","lastTransitionTime":"2025-10-02T01:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.053837 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.053888 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.053902 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.053920 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.053934 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:01Z","lastTransitionTime":"2025-10-02T01:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.156772 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.156835 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.156852 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.156875 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.156891 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:01Z","lastTransitionTime":"2025-10-02T01:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.178386 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.178428 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.178443 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.178466 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.178482 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:01Z","lastTransitionTime":"2025-10-02T01:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:01 crc kubenswrapper[4775]: E1002 01:42:01.198379 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.203148 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.203220 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.203242 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.203270 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.203294 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:01Z","lastTransitionTime":"2025-10-02T01:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:01 crc kubenswrapper[4775]: E1002 01:42:01.221897 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.226494 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.226540 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.226557 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.226579 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.226595 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:01Z","lastTransitionTime":"2025-10-02T01:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:01 crc kubenswrapper[4775]: E1002 01:42:01.243711 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.248661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.248717 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.248733 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.248762 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.248784 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:01Z","lastTransitionTime":"2025-10-02T01:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:01 crc kubenswrapper[4775]: E1002 01:42:01.269852 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.274697 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.274770 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.274795 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.274823 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.274847 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:01Z","lastTransitionTime":"2025-10-02T01:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:01 crc kubenswrapper[4775]: E1002 01:42:01.296673 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:01Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:01 crc kubenswrapper[4775]: E1002 01:42:01.297043 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.299381 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.299455 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.299477 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.299500 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.299517 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:01Z","lastTransitionTime":"2025-10-02T01:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.403515 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.403578 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.403596 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.403619 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.403636 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:01Z","lastTransitionTime":"2025-10-02T01:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.506760 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.506815 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.506829 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.506847 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.506861 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:01Z","lastTransitionTime":"2025-10-02T01:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.609038 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.609145 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.609161 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.609180 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.609194 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:01Z","lastTransitionTime":"2025-10-02T01:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.711785 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.711833 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.711851 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.711873 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.711888 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:01Z","lastTransitionTime":"2025-10-02T01:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.814771 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.814836 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.814857 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.814891 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.814911 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:01Z","lastTransitionTime":"2025-10-02T01:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.918443 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.918491 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.918510 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.918534 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:01 crc kubenswrapper[4775]: I1002 01:42:01.918550 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:01Z","lastTransitionTime":"2025-10-02T01:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.021483 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.021543 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.021560 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.021586 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.021604 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:02Z","lastTransitionTime":"2025-10-02T01:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.124877 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.124929 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.124948 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.125004 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.125021 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:02Z","lastTransitionTime":"2025-10-02T01:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.227848 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.227929 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.227990 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.228027 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.228049 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:02Z","lastTransitionTime":"2025-10-02T01:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.330661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.330715 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.330731 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.330757 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.330775 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:02Z","lastTransitionTime":"2025-10-02T01:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.433543 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.433602 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.433621 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.433644 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.433660 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:02Z","lastTransitionTime":"2025-10-02T01:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.536202 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.536279 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.536302 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.536333 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.536357 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:02Z","lastTransitionTime":"2025-10-02T01:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.639362 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.639489 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.639512 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.639532 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.639547 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:02Z","lastTransitionTime":"2025-10-02T01:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.742878 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.742942 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.743010 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.743039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.743062 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:02Z","lastTransitionTime":"2025-10-02T01:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.764532 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.764597 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.764546 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.764541 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:02 crc kubenswrapper[4775]: E1002 01:42:02.764740 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:02 crc kubenswrapper[4775]: E1002 01:42:02.764889 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:02 crc kubenswrapper[4775]: E1002 01:42:02.765038 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:02 crc kubenswrapper[4775]: E1002 01:42:02.765155 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.846117 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.846162 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.846177 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.846196 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.846213 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:02Z","lastTransitionTime":"2025-10-02T01:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.949063 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.949129 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.949146 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.949166 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:02 crc kubenswrapper[4775]: I1002 01:42:02.949182 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:02Z","lastTransitionTime":"2025-10-02T01:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.051393 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.051462 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.051479 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.051504 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.051521 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:03Z","lastTransitionTime":"2025-10-02T01:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.154706 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.154794 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.154811 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.154833 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.154852 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:03Z","lastTransitionTime":"2025-10-02T01:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.257797 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.257839 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.257853 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.257870 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.257882 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:03Z","lastTransitionTime":"2025-10-02T01:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.360899 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.360983 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.361001 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.361028 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.361045 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:03Z","lastTransitionTime":"2025-10-02T01:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.464779 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.464842 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.464859 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.464885 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.464903 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:03Z","lastTransitionTime":"2025-10-02T01:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.568092 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.568162 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.568180 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.568202 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.568219 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:03Z","lastTransitionTime":"2025-10-02T01:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.671452 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.671495 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.671510 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.671533 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.671550 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:03Z","lastTransitionTime":"2025-10-02T01:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.774139 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.774193 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.774219 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.774241 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.774258 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:03Z","lastTransitionTime":"2025-10-02T01:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.789172 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.812853 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.829663 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.848734 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.871742 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.876804 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.876858 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.876876 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.876901 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.876918 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:03Z","lastTransitionTime":"2025-10-02T01:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.888220 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.902168 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.916180 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.928660 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.940196 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b00713a-180e-4e32-bdb6-cb34fd64e1ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ef520bf5435b6a70fb920599ee3b72d11062641a24a364340ced83ff966eeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a349484f3015205e6074e79651f9e3a242b3b2597b70e60231115430ee41cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b0204478719ee799c2b7cff276ce9514d66b27c9eafa538ceec0e6478c0bf47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.951004 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.962402 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.979559 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.979608 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.979619 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.979636 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.979648 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:03Z","lastTransitionTime":"2025-10-02T01:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:03 crc kubenswrapper[4775]: I1002 01:42:03.983169 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:53Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:53.805456 6444 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 01:41:53.805551 6444 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:53.805561 6444 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:53.805591 6444 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 01:41:53.805602 6444 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 01:41:53.805613 6444 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 01:41:53.805630 6444 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:53.805641 6444 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:53.805647 6444 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:53.805652 6444 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:41:53.805671 6444 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 01:41:53.805689 6444 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 01:41:53.805711 6444 factory.go:656] Stopping watch factory\\\\nI1002 01:41:53.805719 6444 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 01:41:53.805727 6444 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.001021 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:03Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.014039 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:04Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.028138 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:04Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.041915 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:04Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.059198 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:04Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.083305 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.083371 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.083394 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.083493 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.083519 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:04Z","lastTransitionTime":"2025-10-02T01:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.186726 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.186784 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.186795 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.186813 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.186824 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:04Z","lastTransitionTime":"2025-10-02T01:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.292444 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.292506 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.292524 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.292553 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.292570 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:04Z","lastTransitionTime":"2025-10-02T01:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.395791 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.395882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.395915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.395933 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.395948 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:04Z","lastTransitionTime":"2025-10-02T01:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.506769 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.506838 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.506857 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.507035 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.507073 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:04Z","lastTransitionTime":"2025-10-02T01:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.610726 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.610793 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.610812 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.610837 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.610857 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:04Z","lastTransitionTime":"2025-10-02T01:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.714213 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.714602 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.714620 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.714643 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.714661 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:04Z","lastTransitionTime":"2025-10-02T01:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.764224 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.764278 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.764253 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.764380 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:04 crc kubenswrapper[4775]: E1002 01:42:04.764647 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:04 crc kubenswrapper[4775]: E1002 01:42:04.764766 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:04 crc kubenswrapper[4775]: E1002 01:42:04.764992 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:04 crc kubenswrapper[4775]: E1002 01:42:04.765146 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.818643 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.818710 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.818731 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.818759 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.818785 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:04Z","lastTransitionTime":"2025-10-02T01:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.922003 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.922073 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.922083 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.922096 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:04 crc kubenswrapper[4775]: I1002 01:42:04.922107 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:04Z","lastTransitionTime":"2025-10-02T01:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.025823 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.025890 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.025908 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.025932 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.025975 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:05Z","lastTransitionTime":"2025-10-02T01:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.129015 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.129117 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.129141 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.129170 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.129192 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:05Z","lastTransitionTime":"2025-10-02T01:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.232452 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.232554 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.232574 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.232599 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.232616 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:05Z","lastTransitionTime":"2025-10-02T01:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.335616 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.335720 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.335739 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.335805 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.335835 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:05Z","lastTransitionTime":"2025-10-02T01:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.439466 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.439533 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.439552 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.439575 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.439592 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:05Z","lastTransitionTime":"2025-10-02T01:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.543148 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.543214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.543231 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.543257 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.543274 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:05Z","lastTransitionTime":"2025-10-02T01:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.646769 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.646822 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.646839 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.646864 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.646880 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:05Z","lastTransitionTime":"2025-10-02T01:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.750312 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.750410 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.750427 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.750450 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.750468 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:05Z","lastTransitionTime":"2025-10-02T01:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.853730 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.853796 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.853813 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.853834 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.853850 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:05Z","lastTransitionTime":"2025-10-02T01:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.956901 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.956997 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.957016 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.957068 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:05 crc kubenswrapper[4775]: I1002 01:42:05.957091 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:05Z","lastTransitionTime":"2025-10-02T01:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.060419 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.060480 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.060510 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.060546 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.060563 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:06Z","lastTransitionTime":"2025-10-02T01:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.163468 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.163537 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.163554 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.163577 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.163596 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:06Z","lastTransitionTime":"2025-10-02T01:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.266611 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.266673 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.266689 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.266712 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.266728 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:06Z","lastTransitionTime":"2025-10-02T01:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.369944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.370027 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.370044 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.370067 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.370085 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:06Z","lastTransitionTime":"2025-10-02T01:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.473792 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.473866 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.473890 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.473917 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.473942 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:06Z","lastTransitionTime":"2025-10-02T01:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.576293 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.576349 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.576367 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.576390 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.576407 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:06Z","lastTransitionTime":"2025-10-02T01:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.678627 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.678661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.678672 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.678687 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.678699 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:06Z","lastTransitionTime":"2025-10-02T01:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.764885 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:06 crc kubenswrapper[4775]: E1002 01:42:06.765165 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.765461 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:06 crc kubenswrapper[4775]: E1002 01:42:06.765622 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.766187 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:06 crc kubenswrapper[4775]: E1002 01:42:06.766296 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.766351 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:06 crc kubenswrapper[4775]: E1002 01:42:06.766436 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.781932 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.782044 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.782068 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.782104 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.782123 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:06Z","lastTransitionTime":"2025-10-02T01:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.885267 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.885328 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.885345 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.885372 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.885392 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:06Z","lastTransitionTime":"2025-10-02T01:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.989126 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.989171 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.989180 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.989196 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:06 crc kubenswrapper[4775]: I1002 01:42:06.989209 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:06Z","lastTransitionTime":"2025-10-02T01:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.092146 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.092197 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.092214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.092236 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.092252 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:07Z","lastTransitionTime":"2025-10-02T01:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.194833 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.194866 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.194875 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.194889 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.194899 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:07Z","lastTransitionTime":"2025-10-02T01:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.297404 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.297473 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.297493 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.297514 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.297527 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:07Z","lastTransitionTime":"2025-10-02T01:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.400550 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.400680 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.400706 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.400730 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.400747 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:07Z","lastTransitionTime":"2025-10-02T01:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.503790 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.503858 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.503880 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.503908 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.503928 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:07Z","lastTransitionTime":"2025-10-02T01:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.607261 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.607318 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.607336 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.607361 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.607379 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:07Z","lastTransitionTime":"2025-10-02T01:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.710465 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.710518 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.710535 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.710557 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.710574 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:07Z","lastTransitionTime":"2025-10-02T01:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.813316 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.813357 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.813368 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.813389 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.813404 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:07Z","lastTransitionTime":"2025-10-02T01:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.925152 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.925220 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.925237 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.925263 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:07 crc kubenswrapper[4775]: I1002 01:42:07.925279 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:07Z","lastTransitionTime":"2025-10-02T01:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.028433 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.028499 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.028519 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.028542 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.028559 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:08Z","lastTransitionTime":"2025-10-02T01:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.131578 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.131638 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.131656 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.131679 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.131696 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:08Z","lastTransitionTime":"2025-10-02T01:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.234408 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.234513 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.234553 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.234586 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.234611 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:08Z","lastTransitionTime":"2025-10-02T01:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.338041 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.338069 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.338077 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.338091 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.338099 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:08Z","lastTransitionTime":"2025-10-02T01:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.441129 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.441194 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.441209 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.441232 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.441249 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:08Z","lastTransitionTime":"2025-10-02T01:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.543887 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.544005 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.544027 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.544046 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.544093 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:08Z","lastTransitionTime":"2025-10-02T01:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.646710 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.646743 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.646753 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.646767 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.646777 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:08Z","lastTransitionTime":"2025-10-02T01:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.749716 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.749775 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.749795 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.749818 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.749857 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:08Z","lastTransitionTime":"2025-10-02T01:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.764508 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.764556 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.764558 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.764641 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:08 crc kubenswrapper[4775]: E1002 01:42:08.764759 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:08 crc kubenswrapper[4775]: E1002 01:42:08.765030 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:08 crc kubenswrapper[4775]: E1002 01:42:08.765107 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:08 crc kubenswrapper[4775]: E1002 01:42:08.765167 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.852549 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.852605 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.852622 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.852644 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.852660 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:08Z","lastTransitionTime":"2025-10-02T01:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.955923 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.956051 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.956074 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.956104 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:08 crc kubenswrapper[4775]: I1002 01:42:08.956133 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:08Z","lastTransitionTime":"2025-10-02T01:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.058756 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.058801 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.058809 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.058823 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.058831 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:09Z","lastTransitionTime":"2025-10-02T01:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.161399 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.161449 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.161467 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.161488 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.161505 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:09Z","lastTransitionTime":"2025-10-02T01:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.264413 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.264471 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.264489 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.264511 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.264528 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:09Z","lastTransitionTime":"2025-10-02T01:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.366512 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.366803 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.366923 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.367030 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.367133 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:09Z","lastTransitionTime":"2025-10-02T01:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.469426 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.469666 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.469730 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.469794 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.469853 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:09Z","lastTransitionTime":"2025-10-02T01:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.572626 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.572686 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.572698 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.572715 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.572727 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:09Z","lastTransitionTime":"2025-10-02T01:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.675444 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.675479 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.675488 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.675504 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.675515 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:09Z","lastTransitionTime":"2025-10-02T01:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.777660 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.777712 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.777728 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.777749 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.777768 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:09Z","lastTransitionTime":"2025-10-02T01:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.880911 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.880994 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.881063 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.881086 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.881105 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:09Z","lastTransitionTime":"2025-10-02T01:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.984514 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.984559 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.984572 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.984589 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:09 crc kubenswrapper[4775]: I1002 01:42:09.984603 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:09Z","lastTransitionTime":"2025-10-02T01:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.087620 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.087667 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.087688 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.087719 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.087742 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:10Z","lastTransitionTime":"2025-10-02T01:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.191014 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.191073 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.191084 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.191105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.191128 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:10Z","lastTransitionTime":"2025-10-02T01:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.292993 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.293037 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.293048 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.293091 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.293101 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:10Z","lastTransitionTime":"2025-10-02T01:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.395417 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.395460 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.395474 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.395491 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.395507 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:10Z","lastTransitionTime":"2025-10-02T01:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.497788 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.497827 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.497836 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.497851 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.497861 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:10Z","lastTransitionTime":"2025-10-02T01:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.600457 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.600509 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.600527 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.600552 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.600569 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:10Z","lastTransitionTime":"2025-10-02T01:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.703162 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.703205 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.703217 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.703233 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.703245 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:10Z","lastTransitionTime":"2025-10-02T01:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.765168 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.765219 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.765173 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:10 crc kubenswrapper[4775]: E1002 01:42:10.765290 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.765369 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:10 crc kubenswrapper[4775]: E1002 01:42:10.765705 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:10 crc kubenswrapper[4775]: E1002 01:42:10.765765 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:10 crc kubenswrapper[4775]: E1002 01:42:10.765858 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.766313 4775 scope.go:117] "RemoveContainer" containerID="af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995" Oct 02 01:42:10 crc kubenswrapper[4775]: E1002 01:42:10.766750 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.805315 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.805363 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.805374 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.805390 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.805400 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:10Z","lastTransitionTime":"2025-10-02T01:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.908223 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.908278 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.908295 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.908319 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:10 crc kubenswrapper[4775]: I1002 01:42:10.908337 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:10Z","lastTransitionTime":"2025-10-02T01:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.011170 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.011235 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.011259 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.011285 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.011307 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:11Z","lastTransitionTime":"2025-10-02T01:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.113680 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.113739 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.113756 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.113779 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.113796 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:11Z","lastTransitionTime":"2025-10-02T01:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.216202 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.216289 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.216313 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.216350 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.216374 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:11Z","lastTransitionTime":"2025-10-02T01:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.318922 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.318992 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.319005 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.319022 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.319036 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:11Z","lastTransitionTime":"2025-10-02T01:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.390104 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.390148 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.390157 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.390170 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.390179 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:11Z","lastTransitionTime":"2025-10-02T01:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:11 crc kubenswrapper[4775]: E1002 01:42:11.405007 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.410500 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.410535 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.410550 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.410565 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.410575 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:11Z","lastTransitionTime":"2025-10-02T01:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:11 crc kubenswrapper[4775]: E1002 01:42:11.431076 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.436375 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.436415 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.436425 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.436441 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.436451 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:11Z","lastTransitionTime":"2025-10-02T01:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:11 crc kubenswrapper[4775]: E1002 01:42:11.456569 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.461138 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.461206 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.461226 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.461255 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.461274 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:11Z","lastTransitionTime":"2025-10-02T01:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.461333 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs\") pod \"network-metrics-daemon-x7pdc\" (UID: \"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\") " pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:11 crc kubenswrapper[4775]: E1002 01:42:11.461452 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:42:11 crc kubenswrapper[4775]: E1002 01:42:11.461505 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs podName:ef954ceb-7b89-42cf-83dd-2e64dc9e44c1 nodeName:}" failed. No retries permitted until 2025-10-02 01:42:43.461489712 +0000 UTC m=+100.628233752 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs") pod "network-metrics-daemon-x7pdc" (UID: "ef954ceb-7b89-42cf-83dd-2e64dc9e44c1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:42:11 crc kubenswrapper[4775]: E1002 01:42:11.480593 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.484378 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.484405 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.484413 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.484425 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.484435 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:11Z","lastTransitionTime":"2025-10-02T01:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:11 crc kubenswrapper[4775]: E1002 01:42:11.499636 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:11Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:11 crc kubenswrapper[4775]: E1002 01:42:11.499738 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.501632 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.501682 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.501693 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.501711 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.501722 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:11Z","lastTransitionTime":"2025-10-02T01:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.604186 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.604249 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.604260 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.604274 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.604285 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:11Z","lastTransitionTime":"2025-10-02T01:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.706642 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.706702 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.706719 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.706745 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.706762 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:11Z","lastTransitionTime":"2025-10-02T01:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.808386 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.808443 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.808463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.808489 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.808509 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:11Z","lastTransitionTime":"2025-10-02T01:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.910945 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.911022 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.911039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.911061 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:11 crc kubenswrapper[4775]: I1002 01:42:11.911078 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:11Z","lastTransitionTime":"2025-10-02T01:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.013983 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.014065 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.014081 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.014104 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.014118 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:12Z","lastTransitionTime":"2025-10-02T01:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.117348 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.117413 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.117440 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.117471 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.117520 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:12Z","lastTransitionTime":"2025-10-02T01:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.220723 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.220803 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.220824 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.220848 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.220867 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:12Z","lastTransitionTime":"2025-10-02T01:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.323228 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.323293 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.323310 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.323336 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.323355 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:12Z","lastTransitionTime":"2025-10-02T01:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.425745 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.425782 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.425791 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.425805 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.425814 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:12Z","lastTransitionTime":"2025-10-02T01:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.528740 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.528787 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.528808 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.528837 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.528858 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:12Z","lastTransitionTime":"2025-10-02T01:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.630440 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.630476 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.630484 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.630498 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.630507 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:12Z","lastTransitionTime":"2025-10-02T01:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.732355 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.732390 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.732400 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.732413 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.732423 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:12Z","lastTransitionTime":"2025-10-02T01:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.765028 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.765075 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.765130 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:12 crc kubenswrapper[4775]: E1002 01:42:12.765140 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.765082 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:12 crc kubenswrapper[4775]: E1002 01:42:12.765317 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:12 crc kubenswrapper[4775]: E1002 01:42:12.765418 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:12 crc kubenswrapper[4775]: E1002 01:42:12.765537 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.834642 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.834705 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.834727 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.834757 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.834782 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:12Z","lastTransitionTime":"2025-10-02T01:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.938195 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.938239 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.938249 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.938265 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:12 crc kubenswrapper[4775]: I1002 01:42:12.938278 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:12Z","lastTransitionTime":"2025-10-02T01:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.040378 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.040421 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.040432 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.040449 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.040461 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:13Z","lastTransitionTime":"2025-10-02T01:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.142555 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.142597 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.142609 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.142624 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.142638 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:13Z","lastTransitionTime":"2025-10-02T01:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.244758 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.244820 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.244839 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.244862 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.244879 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:13Z","lastTransitionTime":"2025-10-02T01:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.347354 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.347413 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.347430 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.347452 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.347468 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:13Z","lastTransitionTime":"2025-10-02T01:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.449560 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.449609 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.449626 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.449647 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.449664 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:13Z","lastTransitionTime":"2025-10-02T01:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.553424 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.553458 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.553471 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.553486 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.553495 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:13Z","lastTransitionTime":"2025-10-02T01:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.656227 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.656268 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.656278 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.656292 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.656301 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:13Z","lastTransitionTime":"2025-10-02T01:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.758587 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.758689 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.758715 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.758744 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.758767 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:13Z","lastTransitionTime":"2025-10-02T01:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.803405 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.816081 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.829832 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.841551 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.852926 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.861008 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.861038 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.861046 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.861060 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.861070 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:13Z","lastTransitionTime":"2025-10-02T01:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.865859 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.881216 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.891827 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.903755 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.917197 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.934863 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.947923 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.959667 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.963863 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.963911 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.963929 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.963981 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.964001 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:13Z","lastTransitionTime":"2025-10-02T01:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.975584 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:13 crc kubenswrapper[4775]: I1002 01:42:13.989762 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b00713a-180e-4e32-bdb6-cb34fd64e1ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ef520bf5435b6a70fb920599ee3b72d11062641a24a364340ced83ff966eeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a349484f3015205e6074e79651f9e3a242b3b2597b70e60231115430ee41cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b0204478719ee799c2b7cff276ce9514d66b27c9eafa538ceec0e6478c0bf47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:13Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.001920 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.012688 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.032375 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:53Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:53.805456 6444 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 01:41:53.805551 6444 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:53.805561 6444 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:53.805591 6444 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 01:41:53.805602 6444 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 01:41:53.805613 6444 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 01:41:53.805630 6444 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:53.805641 6444 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:53.805647 6444 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:53.805652 6444 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:41:53.805671 6444 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 01:41:53.805689 6444 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 01:41:53.805711 6444 factory.go:656] Stopping watch factory\\\\nI1002 01:41:53.805719 6444 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 01:41:53.805727 6444 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:14Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.066669 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.066708 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.066720 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.066737 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.066749 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:14Z","lastTransitionTime":"2025-10-02T01:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.169299 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.169344 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.169353 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.169367 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.169378 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:14Z","lastTransitionTime":"2025-10-02T01:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.275521 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.275564 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.275572 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.275586 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.275596 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:14Z","lastTransitionTime":"2025-10-02T01:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.378388 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.378446 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.378481 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.378508 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.378529 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:14Z","lastTransitionTime":"2025-10-02T01:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.481775 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.481818 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.481828 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.481840 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.481849 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:14Z","lastTransitionTime":"2025-10-02T01:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.584394 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.584457 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.584468 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.584484 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.584496 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:14Z","lastTransitionTime":"2025-10-02T01:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.688080 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.688128 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.688167 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.688188 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.688206 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:14Z","lastTransitionTime":"2025-10-02T01:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.765114 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.765149 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.765217 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:14 crc kubenswrapper[4775]: E1002 01:42:14.765257 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.765287 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:14 crc kubenswrapper[4775]: E1002 01:42:14.765385 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:14 crc kubenswrapper[4775]: E1002 01:42:14.765430 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:14 crc kubenswrapper[4775]: E1002 01:42:14.765626 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.791733 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.791791 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.791809 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.791834 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.791854 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:14Z","lastTransitionTime":"2025-10-02T01:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.894417 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.894467 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.894482 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.894504 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.894522 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:14Z","lastTransitionTime":"2025-10-02T01:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.997887 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.997925 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.997934 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.997947 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:14 crc kubenswrapper[4775]: I1002 01:42:14.997976 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:14Z","lastTransitionTime":"2025-10-02T01:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.100773 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.100812 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.100825 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.100842 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.100853 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:15Z","lastTransitionTime":"2025-10-02T01:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.203512 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.203569 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.203584 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.203604 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.203620 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:15Z","lastTransitionTime":"2025-10-02T01:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.229444 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2xv98_ab7ffbf4-19df-4fff-aacc-344eab1d1089/kube-multus/0.log" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.229519 4775 generic.go:334] "Generic (PLEG): container finished" podID="ab7ffbf4-19df-4fff-aacc-344eab1d1089" containerID="30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a" exitCode=1 Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.229557 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2xv98" event={"ID":"ab7ffbf4-19df-4fff-aacc-344eab1d1089","Type":"ContainerDied","Data":"30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a"} Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.230105 4775 scope.go:117] "RemoveContainer" containerID="30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.245253 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.261247 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.279027 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.296791 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.305746 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.305771 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.305781 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.305793 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.305802 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:15Z","lastTransitionTime":"2025-10-02T01:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.315767 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.330395 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.341257 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.357171 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:53Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:53.805456 6444 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 01:41:53.805551 6444 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:53.805561 6444 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:53.805591 6444 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 01:41:53.805602 6444 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 01:41:53.805613 6444 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 01:41:53.805630 6444 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:53.805641 6444 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:53.805647 6444 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:53.805652 6444 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:41:53.805671 6444 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 01:41:53.805689 6444 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 01:41:53.805711 6444 factory.go:656] Stopping watch factory\\\\nI1002 01:41:53.805719 6444 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 01:41:53.805727 6444 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.368449 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b00713a-180e-4e32-bdb6-cb34fd64e1ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ef520bf5435b6a70fb920599ee3b72d11062641a24a364340ced83ff966eeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a349484f3015205e6074e79651f9e3a242b3b2597b70e60231115430ee41cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b0204478719ee799c2b7cff276ce9514d66b27c9eafa538ceec0e6478c0bf47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.378313 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.387402 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.395870 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.407977 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.408255 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.408354 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.408439 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.408523 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:15Z","lastTransitionTime":"2025-10-02T01:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.412106 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.421276 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.436869 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:15Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:42:14Z\\\",\\\"message\\\":\\\"2025-10-02T01:41:29+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6207de43-1968-48d7-82ca-6591efe1a6ee\\\\n2025-10-02T01:41:29+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6207de43-1968-48d7-82ca-6591efe1a6ee to /host/opt/cni/bin/\\\\n2025-10-02T01:41:29Z [verbose] multus-daemon started\\\\n2025-10-02T01:41:29Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:42:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.447324 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.459775 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.467888 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:15Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.510200 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.510231 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.510241 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.510261 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.510270 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:15Z","lastTransitionTime":"2025-10-02T01:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.612828 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.613239 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.613384 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.613556 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.613705 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:15Z","lastTransitionTime":"2025-10-02T01:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.716203 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.716304 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.716326 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.716352 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.716369 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:15Z","lastTransitionTime":"2025-10-02T01:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.818625 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.818652 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.818660 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.818671 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.818679 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:15Z","lastTransitionTime":"2025-10-02T01:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.920819 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.920872 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.920884 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.920900 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:15 crc kubenswrapper[4775]: I1002 01:42:15.920911 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:15Z","lastTransitionTime":"2025-10-02T01:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.022900 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.022976 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.022993 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.023034 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.023050 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:16Z","lastTransitionTime":"2025-10-02T01:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.124715 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.124800 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.124824 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.124854 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.124877 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:16Z","lastTransitionTime":"2025-10-02T01:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.227658 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.227732 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.227755 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.227784 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.227805 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:16Z","lastTransitionTime":"2025-10-02T01:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.235785 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2xv98_ab7ffbf4-19df-4fff-aacc-344eab1d1089/kube-multus/0.log" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.235864 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2xv98" event={"ID":"ab7ffbf4-19df-4fff-aacc-344eab1d1089","Type":"ContainerStarted","Data":"816181b47894f3934c2d223e7a376a1adf06cc23e5d3529fc4fbe0c5f029503e"} Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.255793 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.275686 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.294651 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.314990 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.330344 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.330394 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.330412 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.330436 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.330455 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:16Z","lastTransitionTime":"2025-10-02T01:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.334853 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.355545 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.375661 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b00713a-180e-4e32-bdb6-cb34fd64e1ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ef520bf5435b6a70fb920599ee3b72d11062641a24a364340ced83ff966eeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a349484f3015205e6074e79651f9e3a242b3b2597b70e60231115430ee41cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b0204478719ee799c2b7cff276ce9514d66b27c9eafa538ceec0e6478c0bf47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.396410 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.412854 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.434863 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.434926 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.434944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.434999 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.435017 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:16Z","lastTransitionTime":"2025-10-02T01:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.443433 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:53Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:53.805456 6444 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 01:41:53.805551 6444 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:53.805561 6444 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:53.805591 6444 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 01:41:53.805602 6444 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 01:41:53.805613 6444 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 01:41:53.805630 6444 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:53.805641 6444 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:53.805647 6444 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:53.805652 6444 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:41:53.805671 6444 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 01:41:53.805689 6444 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 01:41:53.805711 6444 factory.go:656] Stopping watch factory\\\\nI1002 01:41:53.805719 6444 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 01:41:53.805727 6444 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.476524 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.490217 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.508570 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://816181b47894f3934c2d223e7a376a1adf06cc23e5d3529fc4fbe0c5f029503e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:42:14Z\\\",\\\"message\\\":\\\"2025-10-02T01:41:29+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6207de43-1968-48d7-82ca-6591efe1a6ee\\\\n2025-10-02T01:41:29+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6207de43-1968-48d7-82ca-6591efe1a6ee to /host/opt/cni/bin/\\\\n2025-10-02T01:41:29Z [verbose] multus-daemon started\\\\n2025-10-02T01:41:29Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:42:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.524523 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.536093 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.536914 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.536940 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.536949 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.536975 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.536985 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:16Z","lastTransitionTime":"2025-10-02T01:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.550320 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.571876 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.587777 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:16Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.640102 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.640140 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.640176 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.640195 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.640206 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:16Z","lastTransitionTime":"2025-10-02T01:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.743055 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.743097 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.743109 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.743128 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.743139 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:16Z","lastTransitionTime":"2025-10-02T01:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.764837 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.764891 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.764897 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.764895 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:16 crc kubenswrapper[4775]: E1002 01:42:16.765021 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:16 crc kubenswrapper[4775]: E1002 01:42:16.765142 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:16 crc kubenswrapper[4775]: E1002 01:42:16.765240 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:16 crc kubenswrapper[4775]: E1002 01:42:16.765379 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.846525 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.846593 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.846611 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.846636 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.846653 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:16Z","lastTransitionTime":"2025-10-02T01:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.950157 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.950220 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.950241 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.950267 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:16 crc kubenswrapper[4775]: I1002 01:42:16.950285 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:16Z","lastTransitionTime":"2025-10-02T01:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.053444 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.053503 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.053520 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.053550 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.053573 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:17Z","lastTransitionTime":"2025-10-02T01:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.156931 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.157012 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.157030 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.157053 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.157070 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:17Z","lastTransitionTime":"2025-10-02T01:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.260257 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.260342 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.260379 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.260412 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.260436 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:17Z","lastTransitionTime":"2025-10-02T01:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.362680 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.362718 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.362728 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.362744 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.362758 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:17Z","lastTransitionTime":"2025-10-02T01:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.465025 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.465061 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.465073 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.465088 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.465099 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:17Z","lastTransitionTime":"2025-10-02T01:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.567427 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.567495 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.567512 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.567536 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.567553 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:17Z","lastTransitionTime":"2025-10-02T01:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.670361 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.670415 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.670432 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.670453 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.670470 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:17Z","lastTransitionTime":"2025-10-02T01:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.772463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.772513 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.772528 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.772547 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.772566 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:17Z","lastTransitionTime":"2025-10-02T01:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.875809 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.875882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.875914 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.875944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.876035 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:17Z","lastTransitionTime":"2025-10-02T01:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.978303 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.978363 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.978383 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.978411 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:17 crc kubenswrapper[4775]: I1002 01:42:17.978433 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:17Z","lastTransitionTime":"2025-10-02T01:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.081463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.081616 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.081637 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.081661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.081710 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:18Z","lastTransitionTime":"2025-10-02T01:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.184233 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.184301 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.184322 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.184355 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.184376 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:18Z","lastTransitionTime":"2025-10-02T01:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.287794 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.287897 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.287915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.288030 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.288058 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:18Z","lastTransitionTime":"2025-10-02T01:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.391060 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.391123 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.391141 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.391166 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.391185 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:18Z","lastTransitionTime":"2025-10-02T01:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.494128 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.494193 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.494217 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.494245 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.494266 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:18Z","lastTransitionTime":"2025-10-02T01:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.596527 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.596596 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.596619 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.596649 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.596671 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:18Z","lastTransitionTime":"2025-10-02T01:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.699930 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.700040 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.700064 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.700093 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.700118 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:18Z","lastTransitionTime":"2025-10-02T01:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.764389 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.764440 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.764459 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:18 crc kubenswrapper[4775]: E1002 01:42:18.764541 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.764571 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:18 crc kubenswrapper[4775]: E1002 01:42:18.764744 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:18 crc kubenswrapper[4775]: E1002 01:42:18.764768 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:18 crc kubenswrapper[4775]: E1002 01:42:18.764930 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.802701 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.802764 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.802790 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.802819 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.802840 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:18Z","lastTransitionTime":"2025-10-02T01:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.905914 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.905998 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.906016 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.906036 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:18 crc kubenswrapper[4775]: I1002 01:42:18.906052 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:18Z","lastTransitionTime":"2025-10-02T01:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.009003 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.009078 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.009098 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.009122 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.009142 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:19Z","lastTransitionTime":"2025-10-02T01:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.112013 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.112087 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.112109 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.112138 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.112166 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:19Z","lastTransitionTime":"2025-10-02T01:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.214897 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.215007 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.215034 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.215062 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.215085 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:19Z","lastTransitionTime":"2025-10-02T01:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.317933 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.318005 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.318021 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.318042 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.318061 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:19Z","lastTransitionTime":"2025-10-02T01:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.420874 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.420929 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.420945 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.421002 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.421026 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:19Z","lastTransitionTime":"2025-10-02T01:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.524731 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.525170 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.525188 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.525210 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.525228 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:19Z","lastTransitionTime":"2025-10-02T01:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.628336 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.628405 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.628428 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.628460 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.628484 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:19Z","lastTransitionTime":"2025-10-02T01:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.731215 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.731287 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.731308 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.731344 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.731365 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:19Z","lastTransitionTime":"2025-10-02T01:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.834578 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.834640 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.834665 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.834693 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.834718 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:19Z","lastTransitionTime":"2025-10-02T01:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.941490 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.941576 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.941602 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.941635 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:19 crc kubenswrapper[4775]: I1002 01:42:19.941669 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:19Z","lastTransitionTime":"2025-10-02T01:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.045272 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.045333 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.045350 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.045375 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.045393 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:20Z","lastTransitionTime":"2025-10-02T01:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.148424 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.148468 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.148484 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.148506 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.148523 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:20Z","lastTransitionTime":"2025-10-02T01:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.251040 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.251087 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.251104 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.251124 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.251138 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:20Z","lastTransitionTime":"2025-10-02T01:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.354199 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.354434 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.354453 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.354477 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.354496 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:20Z","lastTransitionTime":"2025-10-02T01:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.457311 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.457369 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.457385 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.457408 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.457424 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:20Z","lastTransitionTime":"2025-10-02T01:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.561043 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.561116 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.561142 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.561171 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.561194 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:20Z","lastTransitionTime":"2025-10-02T01:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.665167 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.665226 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.665245 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.665269 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.665294 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:20Z","lastTransitionTime":"2025-10-02T01:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.764323 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.764377 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.764380 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.764518 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:20 crc kubenswrapper[4775]: E1002 01:42:20.764702 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:20 crc kubenswrapper[4775]: E1002 01:42:20.764871 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:20 crc kubenswrapper[4775]: E1002 01:42:20.765029 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:20 crc kubenswrapper[4775]: E1002 01:42:20.765190 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.767491 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.767554 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.767613 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.767642 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.767664 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:20Z","lastTransitionTime":"2025-10-02T01:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.871090 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.871158 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.871175 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.871201 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.871222 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:20Z","lastTransitionTime":"2025-10-02T01:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.974573 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.974636 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.974652 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.974676 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:20 crc kubenswrapper[4775]: I1002 01:42:20.974694 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:20Z","lastTransitionTime":"2025-10-02T01:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.077512 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.077594 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.077612 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.077637 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.077654 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:21Z","lastTransitionTime":"2025-10-02T01:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.181335 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.181395 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.181412 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.181436 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.181452 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:21Z","lastTransitionTime":"2025-10-02T01:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.284650 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.284709 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.284727 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.284750 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.284767 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:21Z","lastTransitionTime":"2025-10-02T01:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.388146 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.388223 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.388247 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.388277 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.388299 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:21Z","lastTransitionTime":"2025-10-02T01:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.491199 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.491264 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.491280 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.491304 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.491324 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:21Z","lastTransitionTime":"2025-10-02T01:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.520826 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.520901 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.520914 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.520933 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.520970 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:21Z","lastTransitionTime":"2025-10-02T01:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:21 crc kubenswrapper[4775]: E1002 01:42:21.537030 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.541919 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.542061 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.542082 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.542108 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.542127 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:21Z","lastTransitionTime":"2025-10-02T01:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:21 crc kubenswrapper[4775]: E1002 01:42:21.562774 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.567184 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.567243 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.567261 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.567285 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.567302 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:21Z","lastTransitionTime":"2025-10-02T01:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:21 crc kubenswrapper[4775]: E1002 01:42:21.587269 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.592140 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.592197 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.592213 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.592238 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.592256 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:21Z","lastTransitionTime":"2025-10-02T01:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:21 crc kubenswrapper[4775]: E1002 01:42:21.613091 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.617671 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.617731 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.617748 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.617773 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.617790 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:21Z","lastTransitionTime":"2025-10-02T01:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:21 crc kubenswrapper[4775]: E1002 01:42:21.637929 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:21Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:21 crc kubenswrapper[4775]: E1002 01:42:21.638289 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.643666 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.644090 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.644120 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.644153 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.644188 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:21Z","lastTransitionTime":"2025-10-02T01:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.747613 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.747683 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.747700 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.747722 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.747760 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:21Z","lastTransitionTime":"2025-10-02T01:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.783235 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.850383 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.850447 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.850465 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.850489 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.850509 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:21Z","lastTransitionTime":"2025-10-02T01:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.954118 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.954186 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.954204 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.954228 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:21 crc kubenswrapper[4775]: I1002 01:42:21.954246 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:21Z","lastTransitionTime":"2025-10-02T01:42:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.057912 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.058010 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.058034 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.058062 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.058082 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:22Z","lastTransitionTime":"2025-10-02T01:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.161783 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.161878 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.161906 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.161935 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.161994 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:22Z","lastTransitionTime":"2025-10-02T01:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.264140 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.264205 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.264224 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.264451 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.264492 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:22Z","lastTransitionTime":"2025-10-02T01:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.367480 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.367555 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.367578 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.367606 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.367627 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:22Z","lastTransitionTime":"2025-10-02T01:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.471123 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.471185 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.471203 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.471229 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.471244 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:22Z","lastTransitionTime":"2025-10-02T01:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.574242 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.574380 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.574403 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.574427 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.574483 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:22Z","lastTransitionTime":"2025-10-02T01:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.678189 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.678251 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.678267 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.678289 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.678309 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:22Z","lastTransitionTime":"2025-10-02T01:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.765252 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.765292 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.765302 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.765378 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:22 crc kubenswrapper[4775]: E1002 01:42:22.765495 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:22 crc kubenswrapper[4775]: E1002 01:42:22.765617 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:22 crc kubenswrapper[4775]: E1002 01:42:22.765731 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:22 crc kubenswrapper[4775]: E1002 01:42:22.765838 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.781494 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.781557 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.781574 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.781597 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.781615 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:22Z","lastTransitionTime":"2025-10-02T01:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.885002 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.885051 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.885069 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.885094 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.885111 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:22Z","lastTransitionTime":"2025-10-02T01:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.987590 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.987640 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.987656 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.987680 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:22 crc kubenswrapper[4775]: I1002 01:42:22.987697 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:22Z","lastTransitionTime":"2025-10-02T01:42:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.093010 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.093187 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.093264 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.093301 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.093384 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:23Z","lastTransitionTime":"2025-10-02T01:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.196347 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.196433 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.196452 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.196474 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.196490 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:23Z","lastTransitionTime":"2025-10-02T01:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.298475 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.298521 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.298533 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.298548 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.298561 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:23Z","lastTransitionTime":"2025-10-02T01:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.448126 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.448193 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.448216 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.448246 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.448266 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:23Z","lastTransitionTime":"2025-10-02T01:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.552199 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.552280 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.552298 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.552324 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.552343 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:23Z","lastTransitionTime":"2025-10-02T01:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.654762 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.654828 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.654846 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.654871 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.654888 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:23Z","lastTransitionTime":"2025-10-02T01:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.757525 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.757607 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.757631 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.757662 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.757684 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:23Z","lastTransitionTime":"2025-10-02T01:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.810661 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:53Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:53.805456 6444 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 01:41:53.805551 6444 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:53.805561 6444 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:53.805591 6444 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 01:41:53.805602 6444 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 01:41:53.805613 6444 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 01:41:53.805630 6444 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:53.805641 6444 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:53.805647 6444 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:53.805652 6444 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:41:53.805671 6444 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 01:41:53.805689 6444 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 01:41:53.805711 6444 factory.go:656] Stopping watch factory\\\\nI1002 01:41:53.805719 6444 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 01:41:53.805727 6444 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:23Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.833139 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b00713a-180e-4e32-bdb6-cb34fd64e1ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ef520bf5435b6a70fb920599ee3b72d11062641a24a364340ced83ff966eeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a349484f3015205e6074e79651f9e3a242b3b2597b70e60231115430ee41cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b0204478719ee799c2b7cff276ce9514d66b27c9eafa538ceec0e6478c0bf47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:23Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.849068 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a78ddf-0a02-405b-9617-6c35e0935250\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfca4ef06649242b5de270bcec809ecc2fcee3f340129e6581efe764f2fecf52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6578795fea6ce646a7bfcbd7599591e0e189c7fd8ba2088f6b959b96be7964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6578795fea6ce646a7bfcbd7599591e0e189c7fd8ba2088f6b959b96be7964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:23Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.860940 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.861001 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.861013 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.861029 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.861040 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:23Z","lastTransitionTime":"2025-10-02T01:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.867514 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:23Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.887157 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:23Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.904353 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:23Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.938466 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:23Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.955337 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:23Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.964225 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.964280 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.964297 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.964321 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.964337 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:23Z","lastTransitionTime":"2025-10-02T01:42:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.978484 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://816181b47894f3934c2d223e7a376a1adf06cc23e5d3529fc4fbe0c5f029503e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:42:14Z\\\",\\\"message\\\":\\\"2025-10-02T01:41:29+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6207de43-1968-48d7-82ca-6591efe1a6ee\\\\n2025-10-02T01:41:29+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6207de43-1968-48d7-82ca-6591efe1a6ee to /host/opt/cni/bin/\\\\n2025-10-02T01:41:29Z [verbose] multus-daemon started\\\\n2025-10-02T01:41:29Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:42:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:23Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:23 crc kubenswrapper[4775]: I1002 01:42:23.993754 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:23Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.009457 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.028733 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.044765 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.063657 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.067217 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.067282 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.067302 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.067340 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.067359 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:24Z","lastTransitionTime":"2025-10-02T01:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.082205 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.100436 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.123545 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.144156 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.162174 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:24Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.170944 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.171035 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.171047 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.171066 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.171078 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:24Z","lastTransitionTime":"2025-10-02T01:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.273825 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.273871 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.273882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.273899 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.273911 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:24Z","lastTransitionTime":"2025-10-02T01:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.376818 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.376921 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.376940 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.376990 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.377011 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:24Z","lastTransitionTime":"2025-10-02T01:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.480456 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.480540 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.480563 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.480585 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.480602 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:24Z","lastTransitionTime":"2025-10-02T01:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.584014 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.584083 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.584105 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.584132 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.584154 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:24Z","lastTransitionTime":"2025-10-02T01:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.687481 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.687523 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.687540 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.687560 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.687576 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:24Z","lastTransitionTime":"2025-10-02T01:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.765113 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:24 crc kubenswrapper[4775]: E1002 01:42:24.765258 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.765482 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:24 crc kubenswrapper[4775]: E1002 01:42:24.765567 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.765753 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:24 crc kubenswrapper[4775]: E1002 01:42:24.765848 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.766093 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:24 crc kubenswrapper[4775]: E1002 01:42:24.766202 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.790306 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.790356 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.790372 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.790392 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.790407 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:24Z","lastTransitionTime":"2025-10-02T01:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.893527 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.893579 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.893592 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.893622 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.893633 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:24Z","lastTransitionTime":"2025-10-02T01:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.997142 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.997207 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.997223 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.997246 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:24 crc kubenswrapper[4775]: I1002 01:42:24.997262 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:24Z","lastTransitionTime":"2025-10-02T01:42:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.100352 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.100427 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.100452 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.100480 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.100501 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:25Z","lastTransitionTime":"2025-10-02T01:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.203504 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.203560 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.203576 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.203601 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.203618 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:25Z","lastTransitionTime":"2025-10-02T01:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.306750 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.306861 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.306884 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.306912 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.306936 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:25Z","lastTransitionTime":"2025-10-02T01:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.410877 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.410941 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.410993 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.411020 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.411037 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:25Z","lastTransitionTime":"2025-10-02T01:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.514033 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.514100 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.514122 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.514153 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.514172 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:25Z","lastTransitionTime":"2025-10-02T01:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.617310 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.617546 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.617558 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.617576 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.617588 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:25Z","lastTransitionTime":"2025-10-02T01:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.720905 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.721028 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.721049 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.721082 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.721110 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:25Z","lastTransitionTime":"2025-10-02T01:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.766258 4775 scope.go:117] "RemoveContainer" containerID="af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.824052 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.824111 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.824126 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.824152 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.824172 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:25Z","lastTransitionTime":"2025-10-02T01:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.927093 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.927186 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.927206 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.927229 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:25 crc kubenswrapper[4775]: I1002 01:42:25.927245 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:25Z","lastTransitionTime":"2025-10-02T01:42:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.030382 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.030435 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.030452 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.030474 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.030490 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:26Z","lastTransitionTime":"2025-10-02T01:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.133354 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.133433 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.133451 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.133532 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.133554 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:26Z","lastTransitionTime":"2025-10-02T01:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.237083 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.237138 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.237155 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.237177 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.237193 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:26Z","lastTransitionTime":"2025-10-02T01:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.272008 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/2.log" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.275788 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerStarted","Data":"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3"} Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.276548 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.306002 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.324837 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.339454 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.339524 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.339539 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.339588 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.339602 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:26Z","lastTransitionTime":"2025-10-02T01:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.343054 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.357020 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.374052 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.389418 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.405614 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.417942 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.429746 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.441941 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.442003 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.442017 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.442035 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.442049 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:26Z","lastTransitionTime":"2025-10-02T01:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.442106 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b00713a-180e-4e32-bdb6-cb34fd64e1ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ef520bf5435b6a70fb920599ee3b72d11062641a24a364340ced83ff966eeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a349484f3015205e6074e79651f9e3a242b3b2597b70e60231115430ee41cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b0204478719ee799c2b7cff276ce9514d66b27c9eafa538ceec0e6478c0bf47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.460862 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a78ddf-0a02-405b-9617-6c35e0935250\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfca4ef06649242b5de270bcec809ecc2fcee3f340129e6581efe764f2fecf52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6578795fea6ce646a7bfcbd7599591e0e189c7fd8ba2088f6b959b96be7964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6578795fea6ce646a7bfcbd7599591e0e189c7fd8ba2088f6b959b96be7964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.474148 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.489771 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.511026 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:53Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:53.805456 6444 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 01:41:53.805551 6444 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:53.805561 6444 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:53.805591 6444 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 01:41:53.805602 6444 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 01:41:53.805613 6444 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 01:41:53.805630 6444 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:53.805641 6444 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:53.805647 6444 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:53.805652 6444 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:41:53.805671 6444 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 01:41:53.805689 6444 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 01:41:53.805711 6444 factory.go:656] Stopping watch factory\\\\nI1002 01:41:53.805719 6444 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 01:41:53.805727 6444 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:42:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.534404 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.544408 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.544487 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.544510 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.544563 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.544586 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:26Z","lastTransitionTime":"2025-10-02T01:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.546392 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.560344 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://816181b47894f3934c2d223e7a376a1adf06cc23e5d3529fc4fbe0c5f029503e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:42:14Z\\\",\\\"message\\\":\\\"2025-10-02T01:41:29+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6207de43-1968-48d7-82ca-6591efe1a6ee\\\\n2025-10-02T01:41:29+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6207de43-1968-48d7-82ca-6591efe1a6ee to /host/opt/cni/bin/\\\\n2025-10-02T01:41:29Z [verbose] multus-daemon started\\\\n2025-10-02T01:41:29Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:42:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.572976 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.584105 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:26Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.647000 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.647040 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.647051 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.647068 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.647079 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:26Z","lastTransitionTime":"2025-10-02T01:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.751375 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.751659 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.751800 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.751925 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.752083 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:26Z","lastTransitionTime":"2025-10-02T01:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.764996 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.765108 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.765005 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:26 crc kubenswrapper[4775]: E1002 01:42:26.765421 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:26 crc kubenswrapper[4775]: E1002 01:42:26.765683 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.765784 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:26 crc kubenswrapper[4775]: E1002 01:42:26.765927 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:26 crc kubenswrapper[4775]: E1002 01:42:26.766109 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.855500 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.855561 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.855579 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.855602 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.855621 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:26Z","lastTransitionTime":"2025-10-02T01:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.959689 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.959772 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.959792 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.959819 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:26 crc kubenswrapper[4775]: I1002 01:42:26.959839 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:26Z","lastTransitionTime":"2025-10-02T01:42:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.063607 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.063669 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.063686 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.063713 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.063729 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:27Z","lastTransitionTime":"2025-10-02T01:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.166724 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.166786 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.166807 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.166835 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.166856 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:27Z","lastTransitionTime":"2025-10-02T01:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.270092 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.270160 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.270186 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.270214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.270237 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:27Z","lastTransitionTime":"2025-10-02T01:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.282746 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/3.log" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.283798 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/2.log" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.287676 4775 generic.go:334] "Generic (PLEG): container finished" podID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerID="12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3" exitCode=1 Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.287740 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerDied","Data":"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3"} Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.287796 4775 scope.go:117] "RemoveContainer" containerID="af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.290111 4775 scope.go:117] "RemoveContainer" containerID="12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3" Oct 02 01:42:27 crc kubenswrapper[4775]: E1002 01:42:27.290675 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.310276 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.327090 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.359338 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.372554 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.372595 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.372622 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.372667 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.372693 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:27Z","lastTransitionTime":"2025-10-02T01:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.376668 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.396073 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://816181b47894f3934c2d223e7a376a1adf06cc23e5d3529fc4fbe0c5f029503e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:42:14Z\\\",\\\"message\\\":\\\"2025-10-02T01:41:29+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6207de43-1968-48d7-82ca-6591efe1a6ee\\\\n2025-10-02T01:41:29+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6207de43-1968-48d7-82ca-6591efe1a6ee to /host/opt/cni/bin/\\\\n2025-10-02T01:41:29Z [verbose] multus-daemon started\\\\n2025-10-02T01:41:29Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:42:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.415265 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.436459 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.455218 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.476409 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.476509 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.476576 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.476596 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.476620 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.476639 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:27Z","lastTransitionTime":"2025-10-02T01:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.496492 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.515618 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.535309 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.561027 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.580470 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.580564 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.580619 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.580644 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.580662 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:27Z","lastTransitionTime":"2025-10-02T01:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.586288 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.610634 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.641292 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af5b958ccd779029a93649d909dbce9983ed027d4c61afb3da8c4c687d37a995\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:41:53Z\\\",\\\"message\\\":\\\"y (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1002 01:41:53.805456 6444 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1002 01:41:53.805551 6444 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1002 01:41:53.805561 6444 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1002 01:41:53.805591 6444 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1002 01:41:53.805602 6444 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1002 01:41:53.805613 6444 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1002 01:41:53.805630 6444 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1002 01:41:53.805641 6444 handler.go:208] Removed *v1.Node event handler 2\\\\nI1002 01:41:53.805647 6444 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1002 01:41:53.805652 6444 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1002 01:41:53.805671 6444 handler.go:208] Removed *v1.Node event handler 7\\\\nI1002 01:41:53.805689 6444 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1002 01:41:53.805711 6444 factory.go:656] Stopping watch factory\\\\nI1002 01:41:53.805719 6444 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1002 01:41:53.805727 6444 ovnkube.go:599] Stopped ovnkube\\\\nI1002 01:41:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:42:27Z\\\",\\\"message\\\":\\\"vices.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"4607c9b7-15f9-4ba0-86e5-0021ba7e4488\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 01:42:26.759226 6840 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:42:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.650301 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.650576 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:27 crc kubenswrapper[4775]: E1002 01:42:27.650602 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.650563178 +0000 UTC m=+148.817307268 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.650649 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.650741 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:27 crc kubenswrapper[4775]: E1002 01:42:27.650758 4775 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.650804 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:27 crc kubenswrapper[4775]: E1002 01:42:27.650855 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.650829715 +0000 UTC m=+148.817573795 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 02 01:42:27 crc kubenswrapper[4775]: E1002 01:42:27.651032 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:42:27 crc kubenswrapper[4775]: E1002 01:42:27.651070 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:42:27 crc kubenswrapper[4775]: E1002 01:42:27.651098 4775 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:42:27 crc kubenswrapper[4775]: E1002 01:42:27.651136 4775 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:42:27 crc kubenswrapper[4775]: E1002 01:42:27.651175 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.651148164 +0000 UTC m=+148.817892234 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:42:27 crc kubenswrapper[4775]: E1002 01:42:27.651032 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 02 01:42:27 crc kubenswrapper[4775]: E1002 01:42:27.651216 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.651189945 +0000 UTC m=+148.817934025 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 02 01:42:27 crc kubenswrapper[4775]: E1002 01:42:27.651232 4775 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 02 01:42:27 crc kubenswrapper[4775]: E1002 01:42:27.651252 4775 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:42:27 crc kubenswrapper[4775]: E1002 01:42:27.651302 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.651287788 +0000 UTC m=+148.818031858 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.658883 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b00713a-180e-4e32-bdb6-cb34fd64e1ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ef520bf5435b6a70fb920599ee3b72d11062641a24a364340ced83ff966eeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a349484f3015205e6074e79651f9e3a242b3b2597b70e60231115430ee41cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b0204478719ee799c2b7cff276ce9514d66b27c9eafa538ceec0e6478c0bf47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.674671 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a78ddf-0a02-405b-9617-6c35e0935250\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfca4ef06649242b5de270bcec809ecc2fcee3f340129e6581efe764f2fecf52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6578795fea6ce646a7bfcbd7599591e0e189c7fd8ba2088f6b959b96be7964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6578795fea6ce646a7bfcbd7599591e0e189c7fd8ba2088f6b959b96be7964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.683495 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.683576 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.683595 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.683623 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.683642 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:27Z","lastTransitionTime":"2025-10-02T01:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.694775 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:27Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.787072 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.787157 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.787175 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.787203 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.787221 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:27Z","lastTransitionTime":"2025-10-02T01:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.890358 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.890426 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.890448 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.890478 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.890500 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:27Z","lastTransitionTime":"2025-10-02T01:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.992857 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.992915 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.992931 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.992984 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:27 crc kubenswrapper[4775]: I1002 01:42:27.993009 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:27Z","lastTransitionTime":"2025-10-02T01:42:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.096690 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.096750 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.096774 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.096804 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.096828 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:28Z","lastTransitionTime":"2025-10-02T01:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.202658 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.202726 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.202742 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.202769 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.202786 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:28Z","lastTransitionTime":"2025-10-02T01:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.294371 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/3.log" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.300454 4775 scope.go:117] "RemoveContainer" containerID="12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3" Oct 02 01:42:28 crc kubenswrapper[4775]: E1002 01:42:28.301046 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.304637 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.304689 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.304706 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.304728 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.304746 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:28Z","lastTransitionTime":"2025-10-02T01:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.319161 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.334150 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.348693 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.368196 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.384138 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.400280 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.407975 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.408031 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.408050 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.408073 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.408091 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:28Z","lastTransitionTime":"2025-10-02T01:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.419908 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b00713a-180e-4e32-bdb6-cb34fd64e1ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ef520bf5435b6a70fb920599ee3b72d11062641a24a364340ced83ff966eeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a349484f3015205e6074e79651f9e3a242b3b2597b70e60231115430ee41cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b0204478719ee799c2b7cff276ce9514d66b27c9eafa538ceec0e6478c0bf47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.432058 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a78ddf-0a02-405b-9617-6c35e0935250\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfca4ef06649242b5de270bcec809ecc2fcee3f340129e6581efe764f2fecf52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6578795fea6ce646a7bfcbd7599591e0e189c7fd8ba2088f6b959b96be7964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6578795fea6ce646a7bfcbd7599591e0e189c7fd8ba2088f6b959b96be7964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.447982 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.461811 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.481300 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:42:27Z\\\",\\\"message\\\":\\\"vices.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"4607c9b7-15f9-4ba0-86e5-0021ba7e4488\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 01:42:26.759226 6840 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:42:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.510608 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.510661 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.510681 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.510706 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.510728 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:28Z","lastTransitionTime":"2025-10-02T01:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.510836 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.525413 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.544257 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://816181b47894f3934c2d223e7a376a1adf06cc23e5d3529fc4fbe0c5f029503e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:42:14Z\\\",\\\"message\\\":\\\"2025-10-02T01:41:29+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6207de43-1968-48d7-82ca-6591efe1a6ee\\\\n2025-10-02T01:41:29+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6207de43-1968-48d7-82ca-6591efe1a6ee to /host/opt/cni/bin/\\\\n2025-10-02T01:41:29Z [verbose] multus-daemon started\\\\n2025-10-02T01:41:29Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:42:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.561586 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.578115 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.597797 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.619933 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.620042 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.620307 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.620339 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.620367 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:28Z","lastTransitionTime":"2025-10-02T01:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.626234 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.641728 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:28Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.724067 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.724135 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.724147 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.724185 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.724200 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:28Z","lastTransitionTime":"2025-10-02T01:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.764474 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.764504 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.764554 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:28 crc kubenswrapper[4775]: E1002 01:42:28.764661 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.764763 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:28 crc kubenswrapper[4775]: E1002 01:42:28.764931 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:28 crc kubenswrapper[4775]: E1002 01:42:28.765089 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:28 crc kubenswrapper[4775]: E1002 01:42:28.765239 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.827495 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.827560 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.827579 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.827604 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.827621 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:28Z","lastTransitionTime":"2025-10-02T01:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.931259 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.931317 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.931335 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.931358 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:28 crc kubenswrapper[4775]: I1002 01:42:28.931375 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:28Z","lastTransitionTime":"2025-10-02T01:42:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.034601 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.034672 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.034699 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.034725 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.034745 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:29Z","lastTransitionTime":"2025-10-02T01:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.137657 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.137717 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.137733 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.137758 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.137774 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:29Z","lastTransitionTime":"2025-10-02T01:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.242034 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.242143 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.242161 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.242184 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.242201 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:29Z","lastTransitionTime":"2025-10-02T01:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.345442 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.345484 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.345496 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.345512 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.345521 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:29Z","lastTransitionTime":"2025-10-02T01:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.448236 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.448305 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.448322 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.448348 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.448366 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:29Z","lastTransitionTime":"2025-10-02T01:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.551862 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.551935 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.551979 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.552004 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.552021 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:29Z","lastTransitionTime":"2025-10-02T01:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.655743 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.655830 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.655855 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.655884 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.655903 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:29Z","lastTransitionTime":"2025-10-02T01:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.759316 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.759396 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.759414 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.759445 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.759463 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:29Z","lastTransitionTime":"2025-10-02T01:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.862992 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.863060 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.863076 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.863100 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.863122 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:29Z","lastTransitionTime":"2025-10-02T01:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.967321 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.967386 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.967404 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.967427 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:29 crc kubenswrapper[4775]: I1002 01:42:29.967444 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:29Z","lastTransitionTime":"2025-10-02T01:42:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.070768 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.070839 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.070851 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.070869 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.070880 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:30Z","lastTransitionTime":"2025-10-02T01:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.175854 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.175902 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.175913 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.175932 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.175973 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:30Z","lastTransitionTime":"2025-10-02T01:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.279599 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.279667 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.279683 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.279704 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.279720 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:30Z","lastTransitionTime":"2025-10-02T01:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.383116 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.383214 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.383233 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.383321 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.383392 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:30Z","lastTransitionTime":"2025-10-02T01:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.486831 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.486886 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.486903 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.486931 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.486948 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:30Z","lastTransitionTime":"2025-10-02T01:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.590322 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.590395 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.590406 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.590425 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.590436 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:30Z","lastTransitionTime":"2025-10-02T01:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.693924 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.694029 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.694052 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.694083 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.694108 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:30Z","lastTransitionTime":"2025-10-02T01:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.764549 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.764717 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:30 crc kubenswrapper[4775]: E1002 01:42:30.764775 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.764840 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:30 crc kubenswrapper[4775]: E1002 01:42:30.765053 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:30 crc kubenswrapper[4775]: E1002 01:42:30.765168 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.765256 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:30 crc kubenswrapper[4775]: E1002 01:42:30.765371 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.797039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.797095 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.797118 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.797147 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.797167 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:30Z","lastTransitionTime":"2025-10-02T01:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.899262 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.899317 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.899326 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.899346 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:30 crc kubenswrapper[4775]: I1002 01:42:30.899372 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:30Z","lastTransitionTime":"2025-10-02T01:42:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.002102 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.002163 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.002180 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.002204 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.002222 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:31Z","lastTransitionTime":"2025-10-02T01:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.105334 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.105397 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.105416 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.105440 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.105457 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:31Z","lastTransitionTime":"2025-10-02T01:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.208589 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.208674 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.208697 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.208727 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.208747 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:31Z","lastTransitionTime":"2025-10-02T01:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.311912 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.312000 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.312016 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.312039 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.312055 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:31Z","lastTransitionTime":"2025-10-02T01:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.415429 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.415508 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.416378 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.416450 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.416470 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:31Z","lastTransitionTime":"2025-10-02T01:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.519597 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.519650 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.519667 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.519727 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.519747 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:31Z","lastTransitionTime":"2025-10-02T01:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.622397 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.622437 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.622448 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.622463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.622474 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:31Z","lastTransitionTime":"2025-10-02T01:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.725657 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.725718 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.725737 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.725761 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.725779 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:31Z","lastTransitionTime":"2025-10-02T01:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.831766 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.831885 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.831901 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.831924 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.831943 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:31Z","lastTransitionTime":"2025-10-02T01:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.935439 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.935503 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.935521 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.935545 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:31 crc kubenswrapper[4775]: I1002 01:42:31.935563 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:31Z","lastTransitionTime":"2025-10-02T01:42:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.000482 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.000549 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.000564 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.000605 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.000616 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:32Z","lastTransitionTime":"2025-10-02T01:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:32 crc kubenswrapper[4775]: E1002 01:42:32.021348 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.026877 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.026939 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.026986 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.027008 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.027026 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:32Z","lastTransitionTime":"2025-10-02T01:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:32 crc kubenswrapper[4775]: E1002 01:42:32.045882 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.051228 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.051280 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.051292 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.051310 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.051322 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:32Z","lastTransitionTime":"2025-10-02T01:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:32 crc kubenswrapper[4775]: E1002 01:42:32.071464 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.077660 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.077735 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.077756 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.077783 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.077805 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:32Z","lastTransitionTime":"2025-10-02T01:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:32 crc kubenswrapper[4775]: E1002 01:42:32.096820 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.101936 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.102069 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.102089 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.102111 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.102130 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:32Z","lastTransitionTime":"2025-10-02T01:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:32 crc kubenswrapper[4775]: E1002 01:42:32.121223 4775 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"5a28e63a-b617-4331-805d-ce489133ccf9\\\",\\\"systemUUID\\\":\\\"c12787e0-b782-423d-acab-1fed869cc978\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:32Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:32 crc kubenswrapper[4775]: E1002 01:42:32.121500 4775 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.123622 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.123674 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.123689 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.123710 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.123722 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:32Z","lastTransitionTime":"2025-10-02T01:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.226321 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.226385 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.226404 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.226430 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.226450 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:32Z","lastTransitionTime":"2025-10-02T01:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.329098 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.329160 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.329181 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.329208 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.329229 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:32Z","lastTransitionTime":"2025-10-02T01:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.433195 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.433248 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.433266 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.433299 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.433317 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:32Z","lastTransitionTime":"2025-10-02T01:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.536463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.536513 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.536532 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.536555 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.536572 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:32Z","lastTransitionTime":"2025-10-02T01:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.640268 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.640338 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.640356 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.640381 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.640398 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:32Z","lastTransitionTime":"2025-10-02T01:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.743711 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.743770 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.743788 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.743811 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.743828 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:32Z","lastTransitionTime":"2025-10-02T01:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.764941 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.765413 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.765421 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:32 crc kubenswrapper[4775]: E1002 01:42:32.765590 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.765627 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:32 crc kubenswrapper[4775]: E1002 01:42:32.765838 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:32 crc kubenswrapper[4775]: E1002 01:42:32.766059 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:32 crc kubenswrapper[4775]: E1002 01:42:32.766175 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.847439 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.847507 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.847523 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.847548 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.847565 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:32Z","lastTransitionTime":"2025-10-02T01:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.950875 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.950937 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.951008 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.951041 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:32 crc kubenswrapper[4775]: I1002 01:42:32.951062 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:32Z","lastTransitionTime":"2025-10-02T01:42:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.054046 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.054115 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.054133 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.054157 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.054174 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:33Z","lastTransitionTime":"2025-10-02T01:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.157500 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.157552 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.157567 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.157589 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.157606 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:33Z","lastTransitionTime":"2025-10-02T01:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.260361 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.260441 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.260463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.260489 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.260507 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:33Z","lastTransitionTime":"2025-10-02T01:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.363271 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.363324 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.363342 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.363364 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.363383 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:33Z","lastTransitionTime":"2025-10-02T01:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.466567 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.466629 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.466651 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.466679 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.466704 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:33Z","lastTransitionTime":"2025-10-02T01:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.570605 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.570680 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.570698 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.570725 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.570743 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:33Z","lastTransitionTime":"2025-10-02T01:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.674053 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.674131 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.674157 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.674186 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.674208 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:33Z","lastTransitionTime":"2025-10-02T01:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.776917 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.777015 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.777041 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.777071 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.777128 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:33Z","lastTransitionTime":"2025-10-02T01:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.789241 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.806928 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.827866 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ae693d6d-7a34-46db-b76e-c7fb43b5b0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://46dfd4b159c65a228c3d7d91e0a5135e3c6cb00dfcdf96707bc3905fc6016796\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7e88fbc0ea73154eb115cd8530509403c56209d13da482225ccf30f416ebb14e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5dedccdd9ac5dfb9866cf4abcb28c3b58a44dca850986398e24c2101d92899\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d28d5cdf3577d30dc114868a37211f7c2db7f30cadfe55b594885b8d56b8b9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.845647 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"720ef65a-69b0-4129-bc82-74f10f03f5a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a688930b50acf802252391fc68570f75fe78864098454cf704fe264cd5608a29\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://235524bb94a72bf9f99664a2b3a3460e7fdcaaea0a56f3fe66cc3885d16970e4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7483c8accedafb222e17ddc13d648fc51a56679a011a0b7d9898ddf4d5958508\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2657475e7559d053838a40659192bd1bbf94d3e49f202be44d0b0f606e24c09c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f012b0e1aef6566a2c2198b82c2e1bb16789413ea3cebd8db6a9211d1cfd7778\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-02T01:41:23Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1002 01:41:17.467787 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1002 01:41:17.470233 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-288718946/tls.crt::/tmp/serving-cert-288718946/tls.key\\\\\\\"\\\\nI1002 01:41:23.582513 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1002 01:41:23.588520 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1002 01:41:23.588556 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1002 01:41:23.588590 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1002 01:41:23.588600 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1002 01:41:23.601254 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1002 01:41:23.601299 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601310 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1002 01:41:23.601320 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1002 01:41:23.601311 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1002 01:41:23.601328 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1002 01:41:23.601369 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1002 01:41:23.601379 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1002 01:41:23.604698 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dead2efd865dc51d464317a2224534609bcc17ceb0a1b95ab7b6c9ba647e8653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b3afee439c1370c38ffffd3c97ad61948fafe29e332408067e4780343d9c9afa\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.864511 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4d5ea55bacb8a8eea3a6bb5b51f9d998047c2ace1956f07cc68f224829243fd6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea6bdafe32212d0e4e466fc8e70134316a64086fdb16f4c2aad807b065faf330\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.880463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.880528 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.880552 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.880584 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.880610 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:33Z","lastTransitionTime":"2025-10-02T01:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.885837 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.914789 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7178a2dd-9182-400a-959d-c0c4181b6f18\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:42:27Z\\\",\\\"message\\\":\\\"vices.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"4607c9b7-15f9-4ba0-86e5-0021ba7e4488\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-controller-manager-operator/metrics_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-controller-manager-operator/metrics\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.58\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nF1002 01:42:26.759226 6840 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:42:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-27cwc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-zbgms\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.936019 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0b00713a-180e-4e32-bdb6-cb34fd64e1ae\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4ef520bf5435b6a70fb920599ee3b72d11062641a24a364340ced83ff966eeab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a349484f3015205e6074e79651f9e3a242b3b2597b70e60231115430ee41cd5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7b0204478719ee799c2b7cff276ce9514d66b27c9eafa538ceec0e6478c0bf47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1e028466bd8a0714f7cfd622e717ab2e8688f2210d661c9ccfe839c6fdbfb345\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.951761 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d6a78ddf-0a02-405b-9617-6c35e0935250\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dfca4ef06649242b5de270bcec809ecc2fcee3f340129e6581efe764f2fecf52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ce6578795fea6ce646a7bfcbd7599591e0e189c7fd8ba2088f6b959b96be7964\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ce6578795fea6ce646a7bfcbd7599591e0e189c7fd8ba2088f6b959b96be7964\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.970845 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9ffe58bea8a09bd80c8244e158e52d579ed5ce326492c5a0c8af6f8fa12281e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.984333 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.984411 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.984434 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.984463 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.984488 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:33Z","lastTransitionTime":"2025-10-02T01:42:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:33 crc kubenswrapper[4775]: I1002 01:42:33.991092 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c390c5b3-f65b-4f2e-9d03-bb09cf613134\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://407467401e03e4748fb8f791819805c851fa353ff0ab7605d6a6e6093cec8c78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vxqq5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-945lh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:33Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.009726 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:39Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rwwg9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:39Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-x7pdc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.046505 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c2fefdc-754c-4b8b-a0b9-a7b33a8d49b1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://12855a4534f4b87a9032bf2dd358b114dc1fa91eacaf9eb3cdb00de999ad8b6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f8f15bc8ddbed5007bc33cf7ef68967e21bf2b789f711dec2f283f7e51e3f90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2029a82e9787d12c85b2939b8b3b4182cd2dfa48e3c2c0127c3a4168c2729cc0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2a2a48b03b4a663526909e9cdf3489a20d5aedfccb5666e73fb13b5135baa0f7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://329e439fedb753ac97f1062dfaedf8492e5adedd726acc101b178d19fb6dc153\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21eb63477b907531914bfe95b081c909ef89436d0a871b9698e4f5cc7bd4a423\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9abdf66957641a3611d453b973503f124c647db9560ec20c12f4e3fd7c519666\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:06Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6396346f142fa507ec29d5e799de1b1bb14e0f30563c03a4b9cfc45f59c1fbe8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.064302 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9fxjb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88740fe4-c496-4ab9-a518-cfdaeb7fee36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3906f2fd942b8fa1941c3a971f414d5eaa3f70652cb5985539b3de2a0bb95c3e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26cqh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:24Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9fxjb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.086213 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-2xv98" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab7ffbf4-19df-4fff-aacc-344eab1d1089\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:42:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://816181b47894f3934c2d223e7a376a1adf06cc23e5d3529fc4fbe0c5f029503e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-02T01:42:14Z\\\",\\\"message\\\":\\\"2025-10-02T01:41:29+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_6207de43-1968-48d7-82ca-6591efe1a6ee\\\\n2025-10-02T01:41:29+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_6207de43-1968-48d7-82ca-6591efe1a6ee to /host/opt/cni/bin/\\\\n2025-10-02T01:41:29Z [verbose] multus-daemon started\\\\n2025-10-02T01:41:29Z [verbose] Readiness Indicator file check\\\\n2025-10-02T01:42:14Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:42:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4xw9j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-2xv98\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.087814 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.087869 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.087887 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.087912 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.087932 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:34Z","lastTransitionTime":"2025-10-02T01:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.110129 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7ee8fecb-23fc-4714-a079-c37b55da5640\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7cf41599a2f56233bb1796d23bf984a79cce70595b0cf24c34a684a8f5515eb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c48a00477bb58a5611a9ce40216962a5ded8cf3e2494bd0e60f1fe384c526824\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rhg82\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:38Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6wmv6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.132211 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:24Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27539aefe79b448891ce8e4a5833bdcfa95271e586ba99eda6075a61149b3105\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.156651 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d24b3c6-c6ef-4a3c-aa18-635b09af6d7a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcd05d5dbcfcedd5e41f03c217abb7d3c359a74cd112458316d88562d221c2b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ebee7d4b7b045b786ee4e38f7ed6704ab7d72eb2a1c8eb836010cf3ec3d637b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://89aec05fed93248b75c2e303ceb920c77e1a5878edc61e5bce9ae96bc24ffba9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://921d9f695914a88e5d214fa64c4de437ed7ef3d3ac250f29de354074f6f17b9c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2747881e566fe5abde2339c5b5127d2fbb3851cd5f506bac98d1956c86d744cc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b51fb0247a3be3d95d25fb4e66baf7ce44168d7b00c41c6147d0b38938578099\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://63af023a0d16ce04855e307114de766eadbe376bf03dbbf364f4e90511f5c338\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-02T01:41:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-02T01:41:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m4ndm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:25Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ftrn8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.173653 4775 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zrblh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"520ea1b5-f5b6-4de5-93dc-dadd513c4a33\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T01:41:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d00e50e8f2852af4b31b72fce5e99c8d1831ff82a537fe7ed9e7ea2278b5ad0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-02T01:41:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-4wntq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T01:41:27Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zrblh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-02T01:42:34Z is after 2025-08-24T17:21:41Z" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.191348 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.191392 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.191408 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.191433 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.191451 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:34Z","lastTransitionTime":"2025-10-02T01:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.294657 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.294723 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.294739 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.294764 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.294782 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:34Z","lastTransitionTime":"2025-10-02T01:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.397916 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.398024 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.398044 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.398079 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.398102 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:34Z","lastTransitionTime":"2025-10-02T01:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.501245 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.501345 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.501410 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.501437 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.501499 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:34Z","lastTransitionTime":"2025-10-02T01:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.605045 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.605103 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.605122 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.605147 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.605164 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:34Z","lastTransitionTime":"2025-10-02T01:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.708500 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.708596 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.708614 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.708638 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.708654 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:34Z","lastTransitionTime":"2025-10-02T01:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.765223 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.765283 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.765326 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:34 crc kubenswrapper[4775]: E1002 01:42:34.765412 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.765467 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:34 crc kubenswrapper[4775]: E1002 01:42:34.765715 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:34 crc kubenswrapper[4775]: E1002 01:42:34.766014 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:34 crc kubenswrapper[4775]: E1002 01:42:34.766417 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.812212 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.812301 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.812322 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.812343 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.812361 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:34Z","lastTransitionTime":"2025-10-02T01:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.915556 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.915946 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.916001 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.916026 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:34 crc kubenswrapper[4775]: I1002 01:42:34.916048 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:34Z","lastTransitionTime":"2025-10-02T01:42:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.022153 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.022218 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.022235 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.022260 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.022277 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:35Z","lastTransitionTime":"2025-10-02T01:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.125878 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.125976 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.126003 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.126027 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.126045 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:35Z","lastTransitionTime":"2025-10-02T01:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.228986 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.229053 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.229071 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.229094 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.229112 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:35Z","lastTransitionTime":"2025-10-02T01:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.331371 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.331433 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.331450 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.331476 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.331494 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:35Z","lastTransitionTime":"2025-10-02T01:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.434579 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.434677 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.434704 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.434734 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.434758 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:35Z","lastTransitionTime":"2025-10-02T01:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.539124 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.539187 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.539209 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.539240 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.539261 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:35Z","lastTransitionTime":"2025-10-02T01:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.642863 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.643092 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.643127 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.643156 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.643180 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:35Z","lastTransitionTime":"2025-10-02T01:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.746851 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.746912 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.746930 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.746985 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.747006 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:35Z","lastTransitionTime":"2025-10-02T01:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.849628 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.849703 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.849731 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.849762 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.849783 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:35Z","lastTransitionTime":"2025-10-02T01:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.952521 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.952632 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.952655 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.952683 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:35 crc kubenswrapper[4775]: I1002 01:42:35.952705 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:35Z","lastTransitionTime":"2025-10-02T01:42:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.056417 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.056476 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.056491 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.056511 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.056527 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:36Z","lastTransitionTime":"2025-10-02T01:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.159341 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.159397 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.159414 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.159437 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.159453 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:36Z","lastTransitionTime":"2025-10-02T01:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.262002 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.262067 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.262092 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.262124 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.262150 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:36Z","lastTransitionTime":"2025-10-02T01:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.365479 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.365565 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.365586 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.365611 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.365628 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:36Z","lastTransitionTime":"2025-10-02T01:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.468372 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.468426 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.468447 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.468517 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.468543 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:36Z","lastTransitionTime":"2025-10-02T01:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.571667 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.571726 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.571739 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.571758 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.571771 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:36Z","lastTransitionTime":"2025-10-02T01:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.674544 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.674640 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.674667 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.674695 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.674715 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:36Z","lastTransitionTime":"2025-10-02T01:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.764683 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.764750 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.764762 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.764711 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:36 crc kubenswrapper[4775]: E1002 01:42:36.764856 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:36 crc kubenswrapper[4775]: E1002 01:42:36.764994 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:36 crc kubenswrapper[4775]: E1002 01:42:36.765099 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:36 crc kubenswrapper[4775]: E1002 01:42:36.765234 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.777811 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.777882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.777906 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.777934 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.777997 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:36Z","lastTransitionTime":"2025-10-02T01:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.881118 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.881202 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.881222 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.881242 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.881257 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:36Z","lastTransitionTime":"2025-10-02T01:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.984255 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.984316 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.984332 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.984391 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:36 crc kubenswrapper[4775]: I1002 01:42:36.984414 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:36Z","lastTransitionTime":"2025-10-02T01:42:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.088010 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.088072 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.088092 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.088115 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.088132 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:37Z","lastTransitionTime":"2025-10-02T01:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.190865 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.190933 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.190999 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.191031 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.191053 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:37Z","lastTransitionTime":"2025-10-02T01:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.293447 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.293502 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.293518 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.293541 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.293560 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:37Z","lastTransitionTime":"2025-10-02T01:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.397284 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.397349 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.397370 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.397393 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.397409 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:37Z","lastTransitionTime":"2025-10-02T01:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.499580 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.499636 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.499653 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.499676 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.499698 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:37Z","lastTransitionTime":"2025-10-02T01:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.602900 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.603152 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.603219 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.603309 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.603372 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:37Z","lastTransitionTime":"2025-10-02T01:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.706764 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.706822 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.706840 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.706867 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.706883 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:37Z","lastTransitionTime":"2025-10-02T01:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.810142 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.810225 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.810247 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.810271 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.810293 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:37Z","lastTransitionTime":"2025-10-02T01:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.913688 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.913760 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.913779 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.913802 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:37 crc kubenswrapper[4775]: I1002 01:42:37.913818 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:37Z","lastTransitionTime":"2025-10-02T01:42:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.016762 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.016824 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.016846 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.016869 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.016886 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:38Z","lastTransitionTime":"2025-10-02T01:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.121133 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.121213 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.121237 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.121264 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.121287 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:38Z","lastTransitionTime":"2025-10-02T01:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.224197 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.224262 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.224279 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.224309 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.224338 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:38Z","lastTransitionTime":"2025-10-02T01:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.326982 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.327047 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.327065 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.327088 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.327105 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:38Z","lastTransitionTime":"2025-10-02T01:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.430110 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.430184 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.430203 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.430232 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.430252 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:38Z","lastTransitionTime":"2025-10-02T01:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.532927 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.533014 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.533031 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.533057 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.533074 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:38Z","lastTransitionTime":"2025-10-02T01:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.635606 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.635676 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.635684 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.635699 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.635709 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:38Z","lastTransitionTime":"2025-10-02T01:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.738714 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.738806 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.738830 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.738858 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.738879 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:38Z","lastTransitionTime":"2025-10-02T01:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.764918 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.765008 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:38 crc kubenswrapper[4775]: E1002 01:42:38.765092 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.765164 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:38 crc kubenswrapper[4775]: E1002 01:42:38.765268 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:38 crc kubenswrapper[4775]: E1002 01:42:38.765350 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.765563 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:38 crc kubenswrapper[4775]: E1002 01:42:38.765691 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.842773 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.842833 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.842850 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.842872 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.842890 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:38Z","lastTransitionTime":"2025-10-02T01:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.946579 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.946632 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.946650 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.946673 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:38 crc kubenswrapper[4775]: I1002 01:42:38.946689 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:38Z","lastTransitionTime":"2025-10-02T01:42:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.050930 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.051020 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.051040 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.051067 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.051097 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:39Z","lastTransitionTime":"2025-10-02T01:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.154493 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.154552 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.154574 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.154597 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.154614 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:39Z","lastTransitionTime":"2025-10-02T01:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.258181 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.258259 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.258282 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.258313 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.258336 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:39Z","lastTransitionTime":"2025-10-02T01:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.360844 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.360925 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.360980 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.361013 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.361035 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:39Z","lastTransitionTime":"2025-10-02T01:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.463508 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.463571 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.463587 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.463611 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.463632 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:39Z","lastTransitionTime":"2025-10-02T01:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.567104 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.567166 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.567183 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.567206 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.567224 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:39Z","lastTransitionTime":"2025-10-02T01:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.670346 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.670403 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.670419 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.670442 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.670460 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:39Z","lastTransitionTime":"2025-10-02T01:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.772705 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.772839 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.772861 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.772881 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.772899 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:39Z","lastTransitionTime":"2025-10-02T01:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.876425 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.876516 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.876545 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.876574 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.876593 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:39Z","lastTransitionTime":"2025-10-02T01:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.979611 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.979710 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.979729 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.979752 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:39 crc kubenswrapper[4775]: I1002 01:42:39.979771 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:39Z","lastTransitionTime":"2025-10-02T01:42:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.082235 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.082291 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.082310 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.082334 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.082353 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:40Z","lastTransitionTime":"2025-10-02T01:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.185945 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.186222 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.186243 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.186272 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.186306 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:40Z","lastTransitionTime":"2025-10-02T01:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.289782 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.289857 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.289878 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.289903 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.289920 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:40Z","lastTransitionTime":"2025-10-02T01:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.392709 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.392774 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.392797 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.392826 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.392846 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:40Z","lastTransitionTime":"2025-10-02T01:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.495093 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.495169 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.495191 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.495219 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.495239 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:40Z","lastTransitionTime":"2025-10-02T01:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.598522 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.598586 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.598607 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.598631 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.598648 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:40Z","lastTransitionTime":"2025-10-02T01:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.701390 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.701453 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.701472 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.701495 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.701513 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:40Z","lastTransitionTime":"2025-10-02T01:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.764436 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.764510 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:40 crc kubenswrapper[4775]: E1002 01:42:40.764568 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.764628 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:40 crc kubenswrapper[4775]: E1002 01:42:40.764744 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.764788 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:40 crc kubenswrapper[4775]: E1002 01:42:40.764889 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:40 crc kubenswrapper[4775]: E1002 01:42:40.765061 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.804472 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.804532 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.804549 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.804571 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.804590 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:40Z","lastTransitionTime":"2025-10-02T01:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.906988 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.907046 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.907065 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.907090 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:40 crc kubenswrapper[4775]: I1002 01:42:40.907106 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:40Z","lastTransitionTime":"2025-10-02T01:42:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.039305 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.039384 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.039403 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.039427 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.039444 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:41Z","lastTransitionTime":"2025-10-02T01:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.143042 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.143108 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.143128 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.143158 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.143181 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:41Z","lastTransitionTime":"2025-10-02T01:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.245843 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.245882 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.245892 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.245904 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.245912 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:41Z","lastTransitionTime":"2025-10-02T01:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.347649 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.347705 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.347717 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.347732 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.347746 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:41Z","lastTransitionTime":"2025-10-02T01:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.450866 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.450930 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.450949 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.451002 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.451019 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:41Z","lastTransitionTime":"2025-10-02T01:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.554733 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.554797 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.554817 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.554842 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.554858 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:41Z","lastTransitionTime":"2025-10-02T01:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.658417 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.658518 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.658538 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.658560 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.658575 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:41Z","lastTransitionTime":"2025-10-02T01:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.761720 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.761834 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.761858 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.761887 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.761907 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:41Z","lastTransitionTime":"2025-10-02T01:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.865285 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.865358 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.865382 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.865410 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.865432 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:41Z","lastTransitionTime":"2025-10-02T01:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.968184 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.968241 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.968257 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.968281 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:41 crc kubenswrapper[4775]: I1002 01:42:41.968297 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:41Z","lastTransitionTime":"2025-10-02T01:42:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.072052 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.072126 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.072150 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.072175 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.072191 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:42Z","lastTransitionTime":"2025-10-02T01:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.175205 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.175282 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.175307 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.175335 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.175356 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:42Z","lastTransitionTime":"2025-10-02T01:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.187682 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.187750 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.187774 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.187805 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.187827 4775 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-02T01:42:42Z","lastTransitionTime":"2025-10-02T01:42:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.296019 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld"] Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.296550 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.300398 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.300404 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.300619 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.300600 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.378247 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podStartSLOduration=78.378170867 podStartE2EDuration="1m18.378170867s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:42:42.338872379 +0000 UTC m=+99.505616469" watchObservedRunningTime="2025-10-02 01:42:42.378170867 +0000 UTC m=+99.544914937" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.411422 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/14a09e78-7039-48b2-8e0a-c17067f31cc2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.411538 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14a09e78-7039-48b2-8e0a-c17067f31cc2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.411591 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/14a09e78-7039-48b2-8e0a-c17067f31cc2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.411646 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/14a09e78-7039-48b2-8e0a-c17067f31cc2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.411701 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14a09e78-7039-48b2-8e0a-c17067f31cc2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.413778 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=21.413764351 podStartE2EDuration="21.413764351s" podCreationTimestamp="2025-10-02 01:42:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:42:42.413205235 +0000 UTC m=+99.579949285" watchObservedRunningTime="2025-10-02 01:42:42.413764351 +0000 UTC m=+99.580508401" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.413878 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=47.413873284 podStartE2EDuration="47.413873284s" podCreationTimestamp="2025-10-02 01:41:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:42:42.399498048 +0000 UTC m=+99.566242128" watchObservedRunningTime="2025-10-02 01:42:42.413873284 +0000 UTC m=+99.580617334" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.458660 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6wmv6" podStartSLOduration=77.458641346 podStartE2EDuration="1m17.458641346s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:42:42.45806558 +0000 UTC m=+99.624809660" watchObservedRunningTime="2025-10-02 01:42:42.458641346 +0000 UTC m=+99.625385396" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.458829 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-2xv98" podStartSLOduration=78.458824091 podStartE2EDuration="1m18.458824091s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:42:42.432316344 +0000 UTC m=+99.599060414" watchObservedRunningTime="2025-10-02 01:42:42.458824091 +0000 UTC m=+99.625568141" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.512785 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14a09e78-7039-48b2-8e0a-c17067f31cc2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.512872 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/14a09e78-7039-48b2-8e0a-c17067f31cc2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.512911 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14a09e78-7039-48b2-8e0a-c17067f31cc2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.512933 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/14a09e78-7039-48b2-8e0a-c17067f31cc2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.512984 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/14a09e78-7039-48b2-8e0a-c17067f31cc2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.513054 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/14a09e78-7039-48b2-8e0a-c17067f31cc2-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.513283 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/14a09e78-7039-48b2-8e0a-c17067f31cc2-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.514200 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/14a09e78-7039-48b2-8e0a-c17067f31cc2-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.520340 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=75.520316936 podStartE2EDuration="1m15.520316936s" podCreationTimestamp="2025-10-02 01:41:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:42:42.506729283 +0000 UTC m=+99.673473343" watchObservedRunningTime="2025-10-02 01:42:42.520316936 +0000 UTC m=+99.687060986" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.521028 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/14a09e78-7039-48b2-8e0a-c17067f31cc2-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.533731 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-9fxjb" podStartSLOduration=79.533711294 podStartE2EDuration="1m19.533711294s" podCreationTimestamp="2025-10-02 01:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:42:42.518846014 +0000 UTC m=+99.685590064" watchObservedRunningTime="2025-10-02 01:42:42.533711294 +0000 UTC m=+99.700455344" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.534255 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14a09e78-7039-48b2-8e0a-c17067f31cc2-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2r5ld\" (UID: \"14a09e78-7039-48b2-8e0a-c17067f31cc2\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.546494 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-zrblh" podStartSLOduration=78.546476674 podStartE2EDuration="1m18.546476674s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:42:42.534389023 +0000 UTC m=+99.701133073" watchObservedRunningTime="2025-10-02 01:42:42.546476674 +0000 UTC m=+99.713220714" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.585351 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-ftrn8" podStartSLOduration=78.585332309 podStartE2EDuration="1m18.585332309s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:42:42.584913098 +0000 UTC m=+99.751657188" watchObservedRunningTime="2025-10-02 01:42:42.585332309 +0000 UTC m=+99.752076339" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.619733 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.664309 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=75.664284666 podStartE2EDuration="1m15.664284666s" podCreationTimestamp="2025-10-02 01:41:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:42:42.66334169 +0000 UTC m=+99.830085780" watchObservedRunningTime="2025-10-02 01:42:42.664284666 +0000 UTC m=+99.831028716" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.683095 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=79.683075806 podStartE2EDuration="1m19.683075806s" podCreationTimestamp="2025-10-02 01:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:42:42.682526351 +0000 UTC m=+99.849270411" watchObservedRunningTime="2025-10-02 01:42:42.683075806 +0000 UTC m=+99.849819886" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.764494 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.764494 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.764570 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:42 crc kubenswrapper[4775]: I1002 01:42:42.764630 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:42 crc kubenswrapper[4775]: E1002 01:42:42.764747 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:42 crc kubenswrapper[4775]: E1002 01:42:42.764925 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:42 crc kubenswrapper[4775]: E1002 01:42:42.765040 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:42 crc kubenswrapper[4775]: E1002 01:42:42.765093 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:43 crc kubenswrapper[4775]: I1002 01:42:43.354680 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" event={"ID":"14a09e78-7039-48b2-8e0a-c17067f31cc2","Type":"ContainerStarted","Data":"0dfd89ea348757e27d2ca447b11e4647715cd2ab84627c0aaceb024ed409ad5d"} Oct 02 01:42:43 crc kubenswrapper[4775]: I1002 01:42:43.354804 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" event={"ID":"14a09e78-7039-48b2-8e0a-c17067f31cc2","Type":"ContainerStarted","Data":"e5d666067f932f1f77330c3ff5ac24460eaa545c11ac7c63f11d17e4f8847708"} Oct 02 01:42:43 crc kubenswrapper[4775]: I1002 01:42:43.379880 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2r5ld" podStartSLOduration=79.379862498 podStartE2EDuration="1m19.379862498s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:42:43.373131038 +0000 UTC m=+100.539875108" watchObservedRunningTime="2025-10-02 01:42:43.379862498 +0000 UTC m=+100.546606538" Oct 02 01:42:43 crc kubenswrapper[4775]: I1002 01:42:43.523719 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs\") pod \"network-metrics-daemon-x7pdc\" (UID: \"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\") " pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:43 crc kubenswrapper[4775]: E1002 01:42:43.523910 4775 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:42:43 crc kubenswrapper[4775]: E1002 01:42:43.524070 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs podName:ef954ceb-7b89-42cf-83dd-2e64dc9e44c1 nodeName:}" failed. No retries permitted until 2025-10-02 01:43:47.524029264 +0000 UTC m=+164.690773344 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs") pod "network-metrics-daemon-x7pdc" (UID: "ef954ceb-7b89-42cf-83dd-2e64dc9e44c1") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 02 01:42:43 crc kubenswrapper[4775]: I1002 01:42:43.767786 4775 scope.go:117] "RemoveContainer" containerID="12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3" Oct 02 01:42:43 crc kubenswrapper[4775]: E1002 01:42:43.768114 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" Oct 02 01:42:44 crc kubenswrapper[4775]: I1002 01:42:44.764845 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:44 crc kubenswrapper[4775]: I1002 01:42:44.764924 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:44 crc kubenswrapper[4775]: I1002 01:42:44.765011 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:44 crc kubenswrapper[4775]: I1002 01:42:44.765054 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:44 crc kubenswrapper[4775]: E1002 01:42:44.766287 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:44 crc kubenswrapper[4775]: E1002 01:42:44.766539 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:44 crc kubenswrapper[4775]: E1002 01:42:44.766708 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:44 crc kubenswrapper[4775]: E1002 01:42:44.767218 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:46 crc kubenswrapper[4775]: I1002 01:42:46.764340 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:46 crc kubenswrapper[4775]: I1002 01:42:46.764393 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:46 crc kubenswrapper[4775]: I1002 01:42:46.764464 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:46 crc kubenswrapper[4775]: I1002 01:42:46.765035 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:46 crc kubenswrapper[4775]: E1002 01:42:46.764922 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:46 crc kubenswrapper[4775]: E1002 01:42:46.765281 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:46 crc kubenswrapper[4775]: E1002 01:42:46.765422 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:46 crc kubenswrapper[4775]: E1002 01:42:46.765518 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:48 crc kubenswrapper[4775]: I1002 01:42:48.764795 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:48 crc kubenswrapper[4775]: I1002 01:42:48.764857 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:48 crc kubenswrapper[4775]: I1002 01:42:48.766044 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:48 crc kubenswrapper[4775]: I1002 01:42:48.766300 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:48 crc kubenswrapper[4775]: E1002 01:42:48.766498 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:48 crc kubenswrapper[4775]: E1002 01:42:48.766281 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:48 crc kubenswrapper[4775]: E1002 01:42:48.766796 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:48 crc kubenswrapper[4775]: E1002 01:42:48.766876 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:50 crc kubenswrapper[4775]: I1002 01:42:50.765212 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:50 crc kubenswrapper[4775]: I1002 01:42:50.765310 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:50 crc kubenswrapper[4775]: E1002 01:42:50.765370 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:50 crc kubenswrapper[4775]: E1002 01:42:50.765479 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:50 crc kubenswrapper[4775]: I1002 01:42:50.765552 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:50 crc kubenswrapper[4775]: E1002 01:42:50.765636 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:50 crc kubenswrapper[4775]: I1002 01:42:50.765720 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:50 crc kubenswrapper[4775]: E1002 01:42:50.765790 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:52 crc kubenswrapper[4775]: I1002 01:42:52.765024 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:52 crc kubenswrapper[4775]: I1002 01:42:52.765148 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:52 crc kubenswrapper[4775]: E1002 01:42:52.765277 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:52 crc kubenswrapper[4775]: I1002 01:42:52.765347 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:52 crc kubenswrapper[4775]: E1002 01:42:52.765351 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:52 crc kubenswrapper[4775]: E1002 01:42:52.765521 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:52 crc kubenswrapper[4775]: I1002 01:42:52.766137 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:52 crc kubenswrapper[4775]: E1002 01:42:52.766282 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:54 crc kubenswrapper[4775]: I1002 01:42:54.764791 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:54 crc kubenswrapper[4775]: E1002 01:42:54.765095 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:54 crc kubenswrapper[4775]: I1002 01:42:54.765179 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:54 crc kubenswrapper[4775]: I1002 01:42:54.765189 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:54 crc kubenswrapper[4775]: E1002 01:42:54.765439 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:54 crc kubenswrapper[4775]: E1002 01:42:54.765554 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:54 crc kubenswrapper[4775]: I1002 01:42:54.766166 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:54 crc kubenswrapper[4775]: E1002 01:42:54.766445 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:54 crc kubenswrapper[4775]: I1002 01:42:54.766568 4775 scope.go:117] "RemoveContainer" containerID="12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3" Oct 02 01:42:54 crc kubenswrapper[4775]: E1002 01:42:54.767035 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-zbgms_openshift-ovn-kubernetes(7178a2dd-9182-400a-959d-c0c4181b6f18)\"" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" Oct 02 01:42:56 crc kubenswrapper[4775]: I1002 01:42:56.765046 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:56 crc kubenswrapper[4775]: E1002 01:42:56.765248 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:56 crc kubenswrapper[4775]: I1002 01:42:56.765265 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:56 crc kubenswrapper[4775]: E1002 01:42:56.765373 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:42:56 crc kubenswrapper[4775]: I1002 01:42:56.765470 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:56 crc kubenswrapper[4775]: E1002 01:42:56.765645 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:56 crc kubenswrapper[4775]: I1002 01:42:56.766107 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:56 crc kubenswrapper[4775]: E1002 01:42:56.766289 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:58 crc kubenswrapper[4775]: I1002 01:42:58.765198 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:42:58 crc kubenswrapper[4775]: I1002 01:42:58.765244 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:42:58 crc kubenswrapper[4775]: E1002 01:42:58.765367 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:42:58 crc kubenswrapper[4775]: I1002 01:42:58.765433 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:42:58 crc kubenswrapper[4775]: I1002 01:42:58.765592 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:42:58 crc kubenswrapper[4775]: E1002 01:42:58.765628 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:42:58 crc kubenswrapper[4775]: E1002 01:42:58.765770 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:42:58 crc kubenswrapper[4775]: E1002 01:42:58.766287 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:43:00 crc kubenswrapper[4775]: I1002 01:43:00.764924 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:43:00 crc kubenswrapper[4775]: I1002 01:43:00.765000 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:00 crc kubenswrapper[4775]: I1002 01:43:00.765031 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:00 crc kubenswrapper[4775]: I1002 01:43:00.765147 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:43:00 crc kubenswrapper[4775]: E1002 01:43:00.765270 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:43:00 crc kubenswrapper[4775]: E1002 01:43:00.765543 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:43:00 crc kubenswrapper[4775]: E1002 01:43:00.765666 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:43:00 crc kubenswrapper[4775]: E1002 01:43:00.765766 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:43:01 crc kubenswrapper[4775]: I1002 01:43:01.421140 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2xv98_ab7ffbf4-19df-4fff-aacc-344eab1d1089/kube-multus/1.log" Oct 02 01:43:01 crc kubenswrapper[4775]: I1002 01:43:01.422031 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2xv98_ab7ffbf4-19df-4fff-aacc-344eab1d1089/kube-multus/0.log" Oct 02 01:43:01 crc kubenswrapper[4775]: I1002 01:43:01.422110 4775 generic.go:334] "Generic (PLEG): container finished" podID="ab7ffbf4-19df-4fff-aacc-344eab1d1089" containerID="816181b47894f3934c2d223e7a376a1adf06cc23e5d3529fc4fbe0c5f029503e" exitCode=1 Oct 02 01:43:01 crc kubenswrapper[4775]: I1002 01:43:01.422156 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2xv98" event={"ID":"ab7ffbf4-19df-4fff-aacc-344eab1d1089","Type":"ContainerDied","Data":"816181b47894f3934c2d223e7a376a1adf06cc23e5d3529fc4fbe0c5f029503e"} Oct 02 01:43:01 crc kubenswrapper[4775]: I1002 01:43:01.422208 4775 scope.go:117] "RemoveContainer" containerID="30f8cd595e17e43f4bf1dc33d260b0124a768c36f5cb518cd59d2c8a9f577f5a" Oct 02 01:43:01 crc kubenswrapper[4775]: I1002 01:43:01.422730 4775 scope.go:117] "RemoveContainer" containerID="816181b47894f3934c2d223e7a376a1adf06cc23e5d3529fc4fbe0c5f029503e" Oct 02 01:43:01 crc kubenswrapper[4775]: E1002 01:43:01.423021 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-2xv98_openshift-multus(ab7ffbf4-19df-4fff-aacc-344eab1d1089)\"" pod="openshift-multus/multus-2xv98" podUID="ab7ffbf4-19df-4fff-aacc-344eab1d1089" Oct 02 01:43:02 crc kubenswrapper[4775]: I1002 01:43:02.427928 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2xv98_ab7ffbf4-19df-4fff-aacc-344eab1d1089/kube-multus/1.log" Oct 02 01:43:02 crc kubenswrapper[4775]: I1002 01:43:02.764934 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:02 crc kubenswrapper[4775]: I1002 01:43:02.765032 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:43:02 crc kubenswrapper[4775]: E1002 01:43:02.765154 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:43:02 crc kubenswrapper[4775]: I1002 01:43:02.765230 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:43:02 crc kubenswrapper[4775]: I1002 01:43:02.765229 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:02 crc kubenswrapper[4775]: E1002 01:43:02.765366 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:43:02 crc kubenswrapper[4775]: E1002 01:43:02.765482 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:43:02 crc kubenswrapper[4775]: E1002 01:43:02.765566 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:43:03 crc kubenswrapper[4775]: E1002 01:43:03.711281 4775 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 02 01:43:03 crc kubenswrapper[4775]: E1002 01:43:03.876019 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 01:43:04 crc kubenswrapper[4775]: I1002 01:43:04.764540 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:43:04 crc kubenswrapper[4775]: I1002 01:43:04.764550 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:04 crc kubenswrapper[4775]: I1002 01:43:04.764608 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:43:04 crc kubenswrapper[4775]: I1002 01:43:04.764716 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:04 crc kubenswrapper[4775]: E1002 01:43:04.765343 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:43:04 crc kubenswrapper[4775]: E1002 01:43:04.765167 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:43:04 crc kubenswrapper[4775]: E1002 01:43:04.765385 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:43:04 crc kubenswrapper[4775]: E1002 01:43:04.764997 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:43:06 crc kubenswrapper[4775]: I1002 01:43:06.765129 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:43:06 crc kubenswrapper[4775]: I1002 01:43:06.765193 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:06 crc kubenswrapper[4775]: E1002 01:43:06.765368 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:43:06 crc kubenswrapper[4775]: I1002 01:43:06.765484 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:43:06 crc kubenswrapper[4775]: I1002 01:43:06.765478 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:06 crc kubenswrapper[4775]: E1002 01:43:06.765642 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:43:06 crc kubenswrapper[4775]: E1002 01:43:06.765810 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:43:06 crc kubenswrapper[4775]: E1002 01:43:06.765933 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:43:07 crc kubenswrapper[4775]: I1002 01:43:07.766380 4775 scope.go:117] "RemoveContainer" containerID="12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3" Oct 02 01:43:08 crc kubenswrapper[4775]: I1002 01:43:08.454717 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/3.log" Oct 02 01:43:08 crc kubenswrapper[4775]: I1002 01:43:08.459411 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerStarted","Data":"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca"} Oct 02 01:43:08 crc kubenswrapper[4775]: I1002 01:43:08.459912 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:43:08 crc kubenswrapper[4775]: I1002 01:43:08.504017 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podStartSLOduration=104.503999227 podStartE2EDuration="1m44.503999227s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:08.503413221 +0000 UTC m=+125.670157301" watchObservedRunningTime="2025-10-02 01:43:08.503999227 +0000 UTC m=+125.670743277" Oct 02 01:43:08 crc kubenswrapper[4775]: I1002 01:43:08.742878 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-x7pdc"] Oct 02 01:43:08 crc kubenswrapper[4775]: I1002 01:43:08.743046 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:43:08 crc kubenswrapper[4775]: E1002 01:43:08.743186 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:43:08 crc kubenswrapper[4775]: I1002 01:43:08.765243 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:08 crc kubenswrapper[4775]: I1002 01:43:08.765326 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:43:08 crc kubenswrapper[4775]: I1002 01:43:08.765376 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:08 crc kubenswrapper[4775]: E1002 01:43:08.765440 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:43:08 crc kubenswrapper[4775]: E1002 01:43:08.765597 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:43:08 crc kubenswrapper[4775]: E1002 01:43:08.765748 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:43:08 crc kubenswrapper[4775]: E1002 01:43:08.877092 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 01:43:10 crc kubenswrapper[4775]: I1002 01:43:10.765052 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:43:10 crc kubenswrapper[4775]: I1002 01:43:10.765159 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:10 crc kubenswrapper[4775]: I1002 01:43:10.765395 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:43:10 crc kubenswrapper[4775]: E1002 01:43:10.765531 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:43:10 crc kubenswrapper[4775]: I1002 01:43:10.765569 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:10 crc kubenswrapper[4775]: E1002 01:43:10.765706 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:43:10 crc kubenswrapper[4775]: E1002 01:43:10.765810 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:43:10 crc kubenswrapper[4775]: E1002 01:43:10.765897 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:43:12 crc kubenswrapper[4775]: I1002 01:43:12.764706 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:43:12 crc kubenswrapper[4775]: I1002 01:43:12.764713 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:12 crc kubenswrapper[4775]: E1002 01:43:12.764932 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:43:12 crc kubenswrapper[4775]: I1002 01:43:12.764752 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:12 crc kubenswrapper[4775]: I1002 01:43:12.764754 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:43:12 crc kubenswrapper[4775]: E1002 01:43:12.765102 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:43:12 crc kubenswrapper[4775]: E1002 01:43:12.765216 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:43:12 crc kubenswrapper[4775]: E1002 01:43:12.765360 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:43:13 crc kubenswrapper[4775]: E1002 01:43:13.877868 4775 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 01:43:14 crc kubenswrapper[4775]: I1002 01:43:14.764606 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:14 crc kubenswrapper[4775]: I1002 01:43:14.764685 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:43:14 crc kubenswrapper[4775]: I1002 01:43:14.764756 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:43:14 crc kubenswrapper[4775]: I1002 01:43:14.764861 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:14 crc kubenswrapper[4775]: E1002 01:43:14.764829 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:43:14 crc kubenswrapper[4775]: E1002 01:43:14.765047 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:43:14 crc kubenswrapper[4775]: E1002 01:43:14.765263 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:43:14 crc kubenswrapper[4775]: E1002 01:43:14.765427 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:43:15 crc kubenswrapper[4775]: I1002 01:43:15.766270 4775 scope.go:117] "RemoveContainer" containerID="816181b47894f3934c2d223e7a376a1adf06cc23e5d3529fc4fbe0c5f029503e" Oct 02 01:43:16 crc kubenswrapper[4775]: I1002 01:43:16.495180 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2xv98_ab7ffbf4-19df-4fff-aacc-344eab1d1089/kube-multus/1.log" Oct 02 01:43:16 crc kubenswrapper[4775]: I1002 01:43:16.495512 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2xv98" event={"ID":"ab7ffbf4-19df-4fff-aacc-344eab1d1089","Type":"ContainerStarted","Data":"e6e3e3d42f733f6c72de0970398567ac16e6e561b792060ddeaa55cb5db1de67"} Oct 02 01:43:16 crc kubenswrapper[4775]: I1002 01:43:16.764768 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:16 crc kubenswrapper[4775]: I1002 01:43:16.764837 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:43:16 crc kubenswrapper[4775]: I1002 01:43:16.764871 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:43:16 crc kubenswrapper[4775]: E1002 01:43:16.765025 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:43:16 crc kubenswrapper[4775]: E1002 01:43:16.765208 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:43:16 crc kubenswrapper[4775]: I1002 01:43:16.765338 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:16 crc kubenswrapper[4775]: E1002 01:43:16.765332 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:43:16 crc kubenswrapper[4775]: E1002 01:43:16.765727 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:43:18 crc kubenswrapper[4775]: I1002 01:43:18.764279 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:43:18 crc kubenswrapper[4775]: I1002 01:43:18.764380 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:18 crc kubenswrapper[4775]: I1002 01:43:18.764388 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:43:18 crc kubenswrapper[4775]: E1002 01:43:18.764474 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-x7pdc" podUID="ef954ceb-7b89-42cf-83dd-2e64dc9e44c1" Oct 02 01:43:18 crc kubenswrapper[4775]: I1002 01:43:18.764274 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:18 crc kubenswrapper[4775]: E1002 01:43:18.764625 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 02 01:43:18 crc kubenswrapper[4775]: E1002 01:43:18.764717 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 02 01:43:18 crc kubenswrapper[4775]: E1002 01:43:18.764794 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 02 01:43:20 crc kubenswrapper[4775]: I1002 01:43:20.764582 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:20 crc kubenswrapper[4775]: I1002 01:43:20.764654 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:43:20 crc kubenswrapper[4775]: I1002 01:43:20.764659 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:20 crc kubenswrapper[4775]: I1002 01:43:20.764603 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:43:20 crc kubenswrapper[4775]: I1002 01:43:20.767691 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 02 01:43:20 crc kubenswrapper[4775]: I1002 01:43:20.768574 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 02 01:43:20 crc kubenswrapper[4775]: I1002 01:43:20.768602 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 02 01:43:20 crc kubenswrapper[4775]: I1002 01:43:20.768604 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 02 01:43:20 crc kubenswrapper[4775]: I1002 01:43:20.768993 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 02 01:43:20 crc kubenswrapper[4775]: I1002 01:43:20.772098 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.249602 4775 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.307063 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rkpgm"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.307701 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.309372 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.326911 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.328153 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.328149 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.330099 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.330319 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.332346 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.332646 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.332795 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.333893 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.334418 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.334604 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5v4h8"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.335421 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.336610 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.339703 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.343419 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.343968 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.344020 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.344546 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.345979 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bxp9r"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.346358 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.349799 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-47dws"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.350775 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.352289 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.352849 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gfcf"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.353531 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.353709 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.354124 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.354298 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.361216 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.361591 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.363432 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.363814 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.365755 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nv4tc"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.366187 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.366646 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnzm6\" (UniqueName: \"kubernetes.io/projected/958d5b03-aa2b-4068-b6fc-6efb125179a0-kube-api-access-qnzm6\") pod \"machine-api-operator-5694c8668f-5v4h8\" (UID: \"958d5b03-aa2b-4068-b6fc-6efb125179a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.366728 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e44853a-a5c9-4e24-9d55-4a8952561c98-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bd8s9\" (UID: \"6e44853a-a5c9-4e24-9d55-4a8952561c98\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.366758 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/958d5b03-aa2b-4068-b6fc-6efb125179a0-config\") pod \"machine-api-operator-5694c8668f-5v4h8\" (UID: \"958d5b03-aa2b-4068-b6fc-6efb125179a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.366807 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5887047-3348-478d-a1d5-694762ac0dec-serving-cert\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.366830 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh5np\" (UniqueName: \"kubernetes.io/projected/f5887047-3348-478d-a1d5-694762ac0dec-kube-api-access-xh5np\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.366872 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/958d5b03-aa2b-4068-b6fc-6efb125179a0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5v4h8\" (UID: \"958d5b03-aa2b-4068-b6fc-6efb125179a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.366892 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.366910 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgl29\" (UniqueName: \"kubernetes.io/projected/a0d23779-3169-4e6a-9d8e-3760b45a145e-kube-api-access-sgl29\") pod \"cluster-samples-operator-665b6dd947-57899\" (UID: \"a0d23779-3169-4e6a-9d8e-3760b45a145e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.366977 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/59534b10-69de-4d5d-ba59-27c4f24df1a7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9jq6t\" (UID: \"59534b10-69de-4d5d-ba59-27c4f24df1a7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.367000 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cx872\" (UniqueName: \"kubernetes.io/projected/59534b10-69de-4d5d-ba59-27c4f24df1a7-kube-api-access-cx872\") pod \"openshift-config-operator-7777fb866f-9jq6t\" (UID: \"59534b10-69de-4d5d-ba59-27c4f24df1a7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.367089 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/958d5b03-aa2b-4068-b6fc-6efb125179a0-images\") pod \"machine-api-operator-5694c8668f-5v4h8\" (UID: \"958d5b03-aa2b-4068-b6fc-6efb125179a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.367186 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-client-ca\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.367219 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0d23779-3169-4e6a-9d8e-3760b45a145e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-57899\" (UID: \"a0d23779-3169-4e6a-9d8e-3760b45a145e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.367261 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e44853a-a5c9-4e24-9d55-4a8952561c98-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bd8s9\" (UID: \"6e44853a-a5c9-4e24-9d55-4a8952561c98\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.367300 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlslc\" (UniqueName: \"kubernetes.io/projected/6e44853a-a5c9-4e24-9d55-4a8952561c98-kube-api-access-nlslc\") pod \"openshift-apiserver-operator-796bbdcf4f-bd8s9\" (UID: \"6e44853a-a5c9-4e24-9d55-4a8952561c98\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.367355 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-config\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.367409 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59534b10-69de-4d5d-ba59-27c4f24df1a7-serving-cert\") pod \"openshift-config-operator-7777fb866f-9jq6t\" (UID: \"59534b10-69de-4d5d-ba59-27c4f24df1a7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.368226 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.369117 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.379774 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-t86lg"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.380561 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-t86lg" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.381472 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-gvb66"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.381985 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-gvb66" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.389356 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-cn7bz"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.389743 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.457185 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.457573 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.457716 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.458228 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.458438 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.458706 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.458817 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.458915 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.459093 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.459194 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.459282 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.459375 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.459437 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.459689 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.459915 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.460086 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.461618 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dzqlc"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.462047 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.462073 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.462546 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.467562 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-v987t"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.468674 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.469043 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.473770 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnzm6\" (UniqueName: \"kubernetes.io/projected/958d5b03-aa2b-4068-b6fc-6efb125179a0-kube-api-access-qnzm6\") pod \"machine-api-operator-5694c8668f-5v4h8\" (UID: \"958d5b03-aa2b-4068-b6fc-6efb125179a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.473810 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e44853a-a5c9-4e24-9d55-4a8952561c98-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bd8s9\" (UID: \"6e44853a-a5c9-4e24-9d55-4a8952561c98\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.473837 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlzvz\" (UniqueName: \"kubernetes.io/projected/4b13e137-25af-400d-a3c6-496fa5138392-kube-api-access-tlzvz\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.473942 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.475430 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.475517 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-trusted-ca-bundle\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.475584 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4b13e137-25af-400d-a3c6-496fa5138392-audit-dir\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.475615 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3e6497ca-0523-4658-8cbd-2340ed601e77-audit-dir\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.475669 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.475724 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/958d5b03-aa2b-4068-b6fc-6efb125179a0-config\") pod \"machine-api-operator-5694c8668f-5v4h8\" (UID: \"958d5b03-aa2b-4068-b6fc-6efb125179a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.475764 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5887047-3348-478d-a1d5-694762ac0dec-serving-cert\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.475818 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh5np\" (UniqueName: \"kubernetes.io/projected/f5887047-3348-478d-a1d5-694762ac0dec-kube-api-access-xh5np\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.475842 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.475890 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/958d5b03-aa2b-4068-b6fc-6efb125179a0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5v4h8\" (UID: \"958d5b03-aa2b-4068-b6fc-6efb125179a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.475915 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.475978 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgl29\" (UniqueName: \"kubernetes.io/projected/a0d23779-3169-4e6a-9d8e-3760b45a145e-kube-api-access-sgl29\") pod \"cluster-samples-operator-665b6dd947-57899\" (UID: \"a0d23779-3169-4e6a-9d8e-3760b45a145e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476008 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4b13e137-25af-400d-a3c6-496fa5138392-encryption-config\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476027 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c69pf\" (UniqueName: \"kubernetes.io/projected/3e6497ca-0523-4658-8cbd-2340ed601e77-kube-api-access-c69pf\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476078 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-config\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476103 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-audit\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476154 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476177 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4b13e137-25af-400d-a3c6-496fa5138392-etcd-client\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476220 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-audit-policies\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476248 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476298 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/59534b10-69de-4d5d-ba59-27c4f24df1a7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9jq6t\" (UID: \"59534b10-69de-4d5d-ba59-27c4f24df1a7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476325 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cx872\" (UniqueName: \"kubernetes.io/projected/59534b10-69de-4d5d-ba59-27c4f24df1a7-kube-api-access-cx872\") pod \"openshift-config-operator-7777fb866f-9jq6t\" (UID: \"59534b10-69de-4d5d-ba59-27c4f24df1a7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476344 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476398 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/958d5b03-aa2b-4068-b6fc-6efb125179a0-images\") pod \"machine-api-operator-5694c8668f-5v4h8\" (UID: \"958d5b03-aa2b-4068-b6fc-6efb125179a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476425 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-client-ca\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476479 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0d23779-3169-4e6a-9d8e-3760b45a145e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-57899\" (UID: \"a0d23779-3169-4e6a-9d8e-3760b45a145e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476502 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476521 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476555 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e44853a-a5c9-4e24-9d55-4a8952561c98-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bd8s9\" (UID: \"6e44853a-a5c9-4e24-9d55-4a8952561c98\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476581 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlslc\" (UniqueName: \"kubernetes.io/projected/6e44853a-a5c9-4e24-9d55-4a8952561c98-kube-api-access-nlslc\") pod \"openshift-apiserver-operator-796bbdcf4f-bd8s9\" (UID: \"6e44853a-a5c9-4e24-9d55-4a8952561c98\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476633 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476656 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476702 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476730 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-config\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476752 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4b13e137-25af-400d-a3c6-496fa5138392-node-pullsecrets\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476757 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476800 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b13e137-25af-400d-a3c6-496fa5138392-serving-cert\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476873 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-image-import-ca\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476893 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476905 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476916 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-etcd-serving-ca\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.476975 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.477003 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59534b10-69de-4d5d-ba59-27c4f24df1a7-serving-cert\") pod \"openshift-config-operator-7777fb866f-9jq6t\" (UID: \"59534b10-69de-4d5d-ba59-27c4f24df1a7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.478281 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.500912 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.501135 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.501193 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.501246 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.501398 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.501476 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.501542 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.501610 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.501666 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.501724 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.501769 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.501816 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.502637 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.503125 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.503406 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.503513 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.503732 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.503970 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.504139 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.504160 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.504286 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.505098 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.505116 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.505190 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.505241 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.505290 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.505334 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.505357 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.505391 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.505650 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.505700 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.505779 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.505872 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.505943 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.506910 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-99jjd"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.507347 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2l4wc"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.507756 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ptw97"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.508189 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.508269 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.508350 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.508424 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.508515 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.508580 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.509552 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.509635 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.509736 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.509815 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.509886 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.509973 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.510095 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.510171 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.510249 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.510546 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.510617 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.510721 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.510815 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.510920 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.511031 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.511859 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-99jjd" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.512127 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-2l4wc" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.512140 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.512258 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-ptw97" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.516590 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.516879 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.516996 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.517108 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.517251 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.517429 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.517524 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.517653 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.516901 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.518011 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.518492 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-klcvg"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.519348 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.519449 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.518145 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.518267 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.518304 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.518364 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.518558 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.518608 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.518798 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.520844 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d8qrs"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.521252 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.521421 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-klcvg" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.526847 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.527811 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.533815 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-client-ca\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.535153 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e44853a-a5c9-4e24-9d55-4a8952561c98-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bd8s9\" (UID: \"6e44853a-a5c9-4e24-9d55-4a8952561c98\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.535693 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/59534b10-69de-4d5d-ba59-27c4f24df1a7-available-featuregates\") pod \"openshift-config-operator-7777fb866f-9jq6t\" (UID: \"59534b10-69de-4d5d-ba59-27c4f24df1a7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.536012 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/958d5b03-aa2b-4068-b6fc-6efb125179a0-images\") pod \"machine-api-operator-5694c8668f-5v4h8\" (UID: \"958d5b03-aa2b-4068-b6fc-6efb125179a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.537158 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e44853a-a5c9-4e24-9d55-4a8952561c98-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bd8s9\" (UID: \"6e44853a-a5c9-4e24-9d55-4a8952561c98\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.537800 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.544022 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-config\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.546537 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.549033 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.550220 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5887047-3348-478d-a1d5-694762ac0dec-serving-cert\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.550658 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/59534b10-69de-4d5d-ba59-27c4f24df1a7-serving-cert\") pod \"openshift-config-operator-7777fb866f-9jq6t\" (UID: \"59534b10-69de-4d5d-ba59-27c4f24df1a7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.548362 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/958d5b03-aa2b-4068-b6fc-6efb125179a0-config\") pod \"machine-api-operator-5694c8668f-5v4h8\" (UID: \"958d5b03-aa2b-4068-b6fc-6efb125179a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.551136 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.552784 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/958d5b03-aa2b-4068-b6fc-6efb125179a0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5v4h8\" (UID: \"958d5b03-aa2b-4068-b6fc-6efb125179a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.556620 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.557349 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.557407 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.559246 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.559889 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0d23779-3169-4e6a-9d8e-3760b45a145e-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-57899\" (UID: \"a0d23779-3169-4e6a-9d8e-3760b45a145e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.569260 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.577227 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.577871 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4b13e137-25af-400d-a3c6-496fa5138392-encryption-config\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.577907 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c69pf\" (UniqueName: \"kubernetes.io/projected/3e6497ca-0523-4658-8cbd-2340ed601e77-kube-api-access-c69pf\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.577924 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wnfts"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.577939 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-config\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.577981 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-audit\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578005 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4b13e137-25af-400d-a3c6-496fa5138392-etcd-client\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578028 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578053 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-audit-policies\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578077 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578107 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578133 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578166 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578191 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578212 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578237 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4b13e137-25af-400d-a3c6-496fa5138392-node-pullsecrets\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578259 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b13e137-25af-400d-a3c6-496fa5138392-serving-cert\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578330 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578346 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-image-import-ca\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578391 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578418 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-etcd-serving-ca\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578443 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578498 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlzvz\" (UniqueName: \"kubernetes.io/projected/4b13e137-25af-400d-a3c6-496fa5138392-kube-api-access-tlzvz\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578523 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-trusted-ca-bundle\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578558 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4b13e137-25af-400d-a3c6-496fa5138392-audit-dir\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578582 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3e6497ca-0523-4658-8cbd-2340ed601e77-audit-dir\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578606 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578664 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.578695 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.579033 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.579610 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.580038 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.580440 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.580742 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.581314 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.581753 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.581772 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.581785 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.581816 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rkpgm"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.581835 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bxp9r"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.581853 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4b13e137-25af-400d-a3c6-496fa5138392-node-pullsecrets\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.581887 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.582457 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4b13e137-25af-400d-a3c6-496fa5138392-encryption-config\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.583022 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wnfts" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.583390 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.583467 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.583045 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-config\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.583705 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.583668 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.580047 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.584256 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-image-import-ca\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.584392 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-audit-policies\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.584514 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-etcd-serving-ca\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.584904 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b13e137-25af-400d-a3c6-496fa5138392-serving-cert\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.585114 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.585332 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-audit\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.585856 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4b13e137-25af-400d-a3c6-496fa5138392-trusted-ca-bundle\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.585909 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4b13e137-25af-400d-a3c6-496fa5138392-audit-dir\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.585946 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3e6497ca-0523-4658-8cbd-2340ed601e77-audit-dir\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.586140 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.586301 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.586821 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5v4h8"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.587272 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.587657 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.588341 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4b13e137-25af-400d-a3c6-496fa5138392-etcd-client\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.588536 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.588795 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.589841 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.594122 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-mgrh5"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.595197 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.596693 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-v987t"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.598069 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-47dws"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.599077 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.599435 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.600477 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.600939 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gfcf"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.601739 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dzqlc"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.602812 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-t86lg"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.603728 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-758r2"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.604391 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-758r2" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.605090 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.606132 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.607280 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-gvb66"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.608540 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.609259 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.609630 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.610877 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.611771 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.612994 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ptw97"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.615020 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d8qrs"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.616464 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2l4wc"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.618656 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.621014 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-fgvlq"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.622199 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-99jjd"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.622203 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-fgvlq" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.623567 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-v9npv"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.624136 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-v9npv" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.625091 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.626262 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.627307 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nv4tc"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.631156 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.633900 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-klcvg"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.634778 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.637755 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mgrh5"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.638581 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.639567 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-fgvlq"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.640756 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wnfts"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.642202 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.643385 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.644694 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.646049 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5bnff"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.647225 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.647434 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.650299 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-758r2"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.650360 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5bnff"] Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.653433 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.669408 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.690073 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.709439 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.730351 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.750391 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.769345 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.790292 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.809197 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.856138 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnzm6\" (UniqueName: \"kubernetes.io/projected/958d5b03-aa2b-4068-b6fc-6efb125179a0-kube-api-access-qnzm6\") pod \"machine-api-operator-5694c8668f-5v4h8\" (UID: \"958d5b03-aa2b-4068-b6fc-6efb125179a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.870170 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.875763 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh5np\" (UniqueName: \"kubernetes.io/projected/f5887047-3348-478d-a1d5-694762ac0dec-kube-api-access-xh5np\") pod \"controller-manager-879f6c89f-rkpgm\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.891108 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.909759 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.930240 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.947392 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.949728 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.969978 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.973107 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 02 01:43:23 crc kubenswrapper[4775]: I1002 01:43:23.991392 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.015484 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.031477 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.050784 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.071013 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.093307 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.130821 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.149814 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.194082 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlslc\" (UniqueName: \"kubernetes.io/projected/6e44853a-a5c9-4e24-9d55-4a8952561c98-kube-api-access-nlslc\") pod \"openshift-apiserver-operator-796bbdcf4f-bd8s9\" (UID: \"6e44853a-a5c9-4e24-9d55-4a8952561c98\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.198615 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rkpgm"] Oct 02 01:43:24 crc kubenswrapper[4775]: W1002 01:43:24.212983 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5887047_3348_478d_a1d5_694762ac0dec.slice/crio-4fa8a584d1c579889578ba5f50b0476374d34aef04d0ba436b5387285bb6ae2e WatchSource:0}: Error finding container 4fa8a584d1c579889578ba5f50b0476374d34aef04d0ba436b5387285bb6ae2e: Status 404 returned error can't find the container with id 4fa8a584d1c579889578ba5f50b0476374d34aef04d0ba436b5387285bb6ae2e Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.232014 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cx872\" (UniqueName: \"kubernetes.io/projected/59534b10-69de-4d5d-ba59-27c4f24df1a7-kube-api-access-cx872\") pod \"openshift-config-operator-7777fb866f-9jq6t\" (UID: \"59534b10-69de-4d5d-ba59-27c4f24df1a7\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.250558 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.252072 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgl29\" (UniqueName: \"kubernetes.io/projected/a0d23779-3169-4e6a-9d8e-3760b45a145e-kube-api-access-sgl29\") pod \"cluster-samples-operator-665b6dd947-57899\" (UID: \"a0d23779-3169-4e6a-9d8e-3760b45a145e\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.259315 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.266815 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5v4h8"] Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.270722 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.280850 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.288294 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.288360 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67aa1649-ec3b-4f32-96fa-d9aece180709-metrics-certs\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.288435 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/64a00282-301f-4947-9795-990b7b0ef4fd-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.288489 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/64a00282-301f-4947-9795-990b7b0ef4fd-audit-policies\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.288523 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdjcm\" (UniqueName: \"kubernetes.io/projected/64a00282-301f-4947-9795-990b7b0ef4fd-kube-api-access-sdjcm\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.288559 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dbe09708-d737-4915-bafb-40d9b6d7897d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-w4snq\" (UID: \"dbe09708-d737-4915-bafb-40d9b6d7897d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.288594 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/064dc547-a0cb-4008-8a6c-f4eb14ab288f-auth-proxy-config\") pod \"machine-approver-56656f9798-6g5xq\" (UID: \"064dc547-a0cb-4008-8a6c-f4eb14ab288f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" Oct 02 01:43:24 crc kubenswrapper[4775]: E1002 01:43:24.288632 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:24.788617076 +0000 UTC m=+141.955361116 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.288670 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/64a00282-301f-4947-9795-990b7b0ef4fd-audit-dir\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.288738 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-registry-tls\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.288771 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dbe09708-d737-4915-bafb-40d9b6d7897d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-w4snq\" (UID: \"dbe09708-d737-4915-bafb-40d9b6d7897d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.288833 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m47n4\" (UniqueName: \"kubernetes.io/projected/63dc8fd2-1f20-4944-a8f1-783700e192f0-kube-api-access-m47n4\") pod \"downloads-7954f5f757-gvb66\" (UID: \"63dc8fd2-1f20-4944-a8f1-783700e192f0\") " pod="openshift-console/downloads-7954f5f757-gvb66" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.290580 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/064dc547-a0cb-4008-8a6c-f4eb14ab288f-config\") pod \"machine-approver-56656f9798-6g5xq\" (UID: \"064dc547-a0cb-4008-8a6c-f4eb14ab288f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.290620 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b746ad9f-7c63-4945-adce-aa2e0311ba30-serving-cert\") pod \"route-controller-manager-6576b87f9c-whx2n\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.290675 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/adc13fc6-a2d8-44fd-ab25-075857884cad-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.290706 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qndkx\" (UniqueName: \"kubernetes.io/projected/67aa1649-ec3b-4f32-96fa-d9aece180709-kube-api-access-qndkx\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.290739 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b746ad9f-7c63-4945-adce-aa2e0311ba30-config\") pod \"route-controller-manager-6576b87f9c-whx2n\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.290785 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64a00282-301f-4947-9795-990b7b0ef4fd-serving-cert\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.290816 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/adc13fc6-a2d8-44fd-ab25-075857884cad-registry-certificates\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.290856 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8g6d\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-kube-api-access-d8g6d\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.290889 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/67aa1649-ec3b-4f32-96fa-d9aece180709-default-certificate\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.290919 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/064dc547-a0cb-4008-8a6c-f4eb14ab288f-machine-approver-tls\") pod \"machine-approver-56656f9798-6g5xq\" (UID: \"064dc547-a0cb-4008-8a6c-f4eb14ab288f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.290995 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67aa1649-ec3b-4f32-96fa-d9aece180709-service-ca-bundle\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291054 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50774a36-1b88-4b62-98ee-2af94e5f217e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-brt8k\" (UID: \"50774a36-1b88-4b62-98ee-2af94e5f217e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291088 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aec43d0e-fdc8-453c-bb71-0445e899102d-serving-cert\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291120 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b746ad9f-7c63-4945-adce-aa2e0311ba30-client-ca\") pod \"route-controller-manager-6576b87f9c-whx2n\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291168 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adc13fc6-a2d8-44fd-ab25-075857884cad-trusted-ca\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291245 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50774a36-1b88-4b62-98ee-2af94e5f217e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-brt8k\" (UID: \"50774a36-1b88-4b62-98ee-2af94e5f217e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291282 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpjsg\" (UniqueName: \"kubernetes.io/projected/50774a36-1b88-4b62-98ee-2af94e5f217e-kube-api-access-tpjsg\") pod \"openshift-controller-manager-operator-756b6f6bc6-brt8k\" (UID: \"50774a36-1b88-4b62-98ee-2af94e5f217e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291337 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/64a00282-301f-4947-9795-990b7b0ef4fd-etcd-client\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291391 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/dbe09708-d737-4915-bafb-40d9b6d7897d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-w4snq\" (UID: \"dbe09708-d737-4915-bafb-40d9b6d7897d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291425 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfdjx\" (UniqueName: \"kubernetes.io/projected/a6754f46-92c3-422e-95d9-638efe436d8e-kube-api-access-xfdjx\") pod \"dns-operator-744455d44c-t86lg\" (UID: \"a6754f46-92c3-422e-95d9-638efe436d8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-t86lg" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291478 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/64a00282-301f-4947-9795-990b7b0ef4fd-encryption-config\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291663 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xswnr\" (UniqueName: \"kubernetes.io/projected/aec43d0e-fdc8-453c-bb71-0445e899102d-kube-api-access-xswnr\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291724 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlrmc\" (UniqueName: \"kubernetes.io/projected/b746ad9f-7c63-4945-adce-aa2e0311ba30-kube-api-access-rlrmc\") pod \"route-controller-manager-6576b87f9c-whx2n\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291762 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dflwc\" (UniqueName: \"kubernetes.io/projected/dbe09708-d737-4915-bafb-40d9b6d7897d-kube-api-access-dflwc\") pod \"cluster-image-registry-operator-dc59b4c8b-w4snq\" (UID: \"dbe09708-d737-4915-bafb-40d9b6d7897d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291793 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64a00282-301f-4947-9795-990b7b0ef4fd-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291839 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-bound-sa-token\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291870 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/67aa1649-ec3b-4f32-96fa-d9aece180709-stats-auth\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291903 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aec43d0e-fdc8-453c-bb71-0445e899102d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.291948 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/adc13fc6-a2d8-44fd-ab25-075857884cad-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.292023 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aec43d0e-fdc8-453c-bb71-0445e899102d-service-ca-bundle\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.289979 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.292709 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a6754f46-92c3-422e-95d9-638efe436d8e-metrics-tls\") pod \"dns-operator-744455d44c-t86lg\" (UID: \"a6754f46-92c3-422e-95d9-638efe436d8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-t86lg" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.293552 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aec43d0e-fdc8-453c-bb71-0445e899102d-config\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.293603 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j6b7\" (UniqueName: \"kubernetes.io/projected/064dc547-a0cb-4008-8a6c-f4eb14ab288f-kube-api-access-4j6b7\") pod \"machine-approver-56656f9798-6g5xq\" (UID: \"064dc547-a0cb-4008-8a6c-f4eb14ab288f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.300564 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.313322 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.332904 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.350902 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.377153 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.390486 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.394341 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:24 crc kubenswrapper[4775]: E1002 01:43:24.394592 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:24.894559814 +0000 UTC m=+142.061303874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.394686 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4nqd\" (UniqueName: \"kubernetes.io/projected/ff918ab3-33fd-44f2-b3ba-9f419f027b10-kube-api-access-j4nqd\") pod \"machine-config-controller-84d6567774-lg72k\" (UID: \"ff918ab3-33fd-44f2-b3ba-9f419f027b10\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.394730 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3dffe31-ef37-4ecd-b72a-26d8541a331f-config\") pod \"kube-apiserver-operator-766d6c64bb-plfk4\" (UID: \"d3dffe31-ef37-4ecd-b72a-26d8541a331f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.394762 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-etcd-client\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.394811 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/127b156f-fe4c-40da-a576-e5b4c2ecbfdf-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-npmlx\" (UID: \"127b156f-fe4c-40da-a576-e5b4c2ecbfdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.394842 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8354382a-de78-4880-b344-358d49362ea8-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gzh7f\" (UID: \"8354382a-de78-4880-b344-358d49362ea8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.394874 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/064dc547-a0cb-4008-8a6c-f4eb14ab288f-config\") pod \"machine-approver-56656f9798-6g5xq\" (UID: \"064dc547-a0cb-4008-8a6c-f4eb14ab288f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.394905 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/adc13fc6-a2d8-44fd-ab25-075857884cad-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.394934 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ff918ab3-33fd-44f2-b3ba-9f419f027b10-proxy-tls\") pod \"machine-config-controller-84d6567774-lg72k\" (UID: \"ff918ab3-33fd-44f2-b3ba-9f419f027b10\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395002 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qndkx\" (UniqueName: \"kubernetes.io/projected/67aa1649-ec3b-4f32-96fa-d9aece180709-kube-api-access-qndkx\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395036 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbrvw\" (UniqueName: \"kubernetes.io/projected/115575fa-2b4f-4b7e-bdd8-46c85da3855d-kube-api-access-rbrvw\") pod \"control-plane-machine-set-operator-78cbb6b69f-wnfts\" (UID: \"115575fa-2b4f-4b7e-bdd8-46c85da3855d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wnfts" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395077 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7962f6fc-2505-49ff-b14a-a90eed3e7ba5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qglml\" (UID: \"7962f6fc-2505-49ff-b14a-a90eed3e7ba5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395111 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ff918ab3-33fd-44f2-b3ba-9f419f027b10-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lg72k\" (UID: \"ff918ab3-33fd-44f2-b3ba-9f419f027b10\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395142 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64a00282-301f-4947-9795-990b7b0ef4fd-serving-cert\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395173 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nghmr\" (UniqueName: \"kubernetes.io/projected/7c0b4bf0-5b95-4ce8-946c-47e89af5d655-kube-api-access-nghmr\") pod \"catalog-operator-68c6474976-tgx4s\" (UID: \"7c0b4bf0-5b95-4ce8-946c-47e89af5d655\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395203 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3dffe31-ef37-4ecd-b72a-26d8541a331f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-plfk4\" (UID: \"d3dffe31-ef37-4ecd-b72a-26d8541a331f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395236 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/adc13fc6-a2d8-44fd-ab25-075857884cad-registry-certificates\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395268 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-config-volume\") pod \"collect-profiles-29322810-q7gnj\" (UID: \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395299 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8sw5\" (UniqueName: \"kubernetes.io/projected/ad595bf8-fed5-4d02-8e6b-fbc3742e0ec2-kube-api-access-h8sw5\") pod \"migrator-59844c95c7-klcvg\" (UID: \"ad595bf8-fed5-4d02-8e6b-fbc3742e0ec2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-klcvg" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395329 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91116a6b-4da7-40db-9633-40a4fb1dcd7e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cng5\" (UID: \"91116a6b-4da7-40db-9633-40a4fb1dcd7e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395364 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8g6d\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-kube-api-access-d8g6d\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395395 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/67aa1649-ec3b-4f32-96fa-d9aece180709-default-certificate\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395424 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/064dc547-a0cb-4008-8a6c-f4eb14ab288f-machine-approver-tls\") pod \"machine-approver-56656f9798-6g5xq\" (UID: \"064dc547-a0cb-4008-8a6c-f4eb14ab288f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395458 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce1437a0-173e-4e15-bc0f-ca00b67e9b2e-webhook-cert\") pod \"packageserver-d55dfcdfc-pm5r9\" (UID: \"ce1437a0-173e-4e15-bc0f-ca00b67e9b2e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395488 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67aa1649-ec3b-4f32-96fa-d9aece180709-service-ca-bundle\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395518 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/395ae874-ba30-40f2-957c-c49098b29a74-metrics-tls\") pod \"dns-default-fgvlq\" (UID: \"395ae874-ba30-40f2-957c-c49098b29a74\") " pod="openshift-dns/dns-default-fgvlq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395548 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ce1437a0-173e-4e15-bc0f-ca00b67e9b2e-tmpfs\") pod \"packageserver-d55dfcdfc-pm5r9\" (UID: \"ce1437a0-173e-4e15-bc0f-ca00b67e9b2e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395578 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tcqs\" (UniqueName: \"kubernetes.io/projected/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-kube-api-access-6tcqs\") pod \"collect-profiles-29322810-q7gnj\" (UID: \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395611 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-etcd-service-ca\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395643 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adc13fc6-a2d8-44fd-ab25-075857884cad-trusted-ca\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395673 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50774a36-1b88-4b62-98ee-2af94e5f217e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-brt8k\" (UID: \"50774a36-1b88-4b62-98ee-2af94e5f217e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395704 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpjsg\" (UniqueName: \"kubernetes.io/projected/50774a36-1b88-4b62-98ee-2af94e5f217e-kube-api-access-tpjsg\") pod \"openshift-controller-manager-operator-756b6f6bc6-brt8k\" (UID: \"50774a36-1b88-4b62-98ee-2af94e5f217e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395734 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aec43d0e-fdc8-453c-bb71-0445e899102d-serving-cert\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395771 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph7dl\" (UniqueName: \"kubernetes.io/projected/91116a6b-4da7-40db-9633-40a4fb1dcd7e-kube-api-access-ph7dl\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cng5\" (UID: \"91116a6b-4da7-40db-9633-40a4fb1dcd7e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395810 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/115575fa-2b4f-4b7e-bdd8-46c85da3855d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-wnfts\" (UID: \"115575fa-2b4f-4b7e-bdd8-46c85da3855d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wnfts" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395841 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-oauth-config\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395875 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7526f918-73a6-4470-9d8a-b79b418a8f36-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ph4s2\" (UID: \"7526f918-73a6-4470-9d8a-b79b418a8f36\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395905 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-plugins-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395934 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/127b156f-fe4c-40da-a576-e5b4c2ecbfdf-config\") pod \"kube-controller-manager-operator-78b949d7b-npmlx\" (UID: \"127b156f-fe4c-40da-a576-e5b4c2ecbfdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.395989 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/64a00282-301f-4947-9795-990b7b0ef4fd-encryption-config\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396024 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4xck\" (UniqueName: \"kubernetes.io/projected/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-kube-api-access-g4xck\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396064 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xswnr\" (UniqueName: \"kubernetes.io/projected/aec43d0e-fdc8-453c-bb71-0445e899102d-kube-api-access-xswnr\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396100 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7526f918-73a6-4470-9d8a-b79b418a8f36-trusted-ca\") pod \"ingress-operator-5b745b69d9-ph4s2\" (UID: \"7526f918-73a6-4470-9d8a-b79b418a8f36\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396136 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb-images\") pod \"machine-config-operator-74547568cd-lgwbw\" (UID: \"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396166 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-secret-volume\") pod \"collect-profiles-29322810-q7gnj\" (UID: \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396212 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7f011434-93f3-4b0c-9ff7-cfbc3f333b83-node-bootstrap-token\") pod \"machine-config-server-v9npv\" (UID: \"7f011434-93f3-4b0c-9ff7-cfbc3f333b83\") " pod="openshift-machine-config-operator/machine-config-server-v9npv" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396248 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dflwc\" (UniqueName: \"kubernetes.io/projected/dbe09708-d737-4915-bafb-40d9b6d7897d-kube-api-access-dflwc\") pod \"cluster-image-registry-operator-dc59b4c8b-w4snq\" (UID: \"dbe09708-d737-4915-bafb-40d9b6d7897d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396281 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlrmc\" (UniqueName: \"kubernetes.io/projected/b746ad9f-7c63-4945-adce-aa2e0311ba30-kube-api-access-rlrmc\") pod \"route-controller-manager-6576b87f9c-whx2n\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396311 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7c0b4bf0-5b95-4ce8-946c-47e89af5d655-profile-collector-cert\") pod \"catalog-operator-68c6474976-tgx4s\" (UID: \"7c0b4bf0-5b95-4ce8-946c-47e89af5d655\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396344 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4-trusted-ca\") pod \"console-operator-58897d9998-v987t\" (UID: \"1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4\") " pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396374 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1d1cb0fe-95a4-4e16-b5f5-c3847c712081-cert\") pod \"ingress-canary-758r2\" (UID: \"1d1cb0fe-95a4-4e16-b5f5-c3847c712081\") " pod="openshift-ingress-canary/ingress-canary-758r2" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396411 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fz5g\" (UniqueName: \"kubernetes.io/projected/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-kube-api-access-7fz5g\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396456 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91116a6b-4da7-40db-9633-40a4fb1dcd7e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cng5\" (UID: \"91116a6b-4da7-40db-9633-40a4fb1dcd7e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396488 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2e1b3210-4726-450d-8da5-766903e6d91b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d8qrs\" (UID: \"2e1b3210-4726-450d-8da5-766903e6d91b\") " pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396522 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4-serving-cert\") pod \"console-operator-58897d9998-v987t\" (UID: \"1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4\") " pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396554 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j6b7\" (UniqueName: \"kubernetes.io/projected/064dc547-a0cb-4008-8a6c-f4eb14ab288f-kube-api-access-4j6b7\") pod \"machine-approver-56656f9798-6g5xq\" (UID: \"064dc547-a0cb-4008-8a6c-f4eb14ab288f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396589 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67aa1649-ec3b-4f32-96fa-d9aece180709-metrics-certs\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396634 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/64a00282-301f-4947-9795-990b7b0ef4fd-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396667 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-trusted-ca-bundle\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396701 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwgf6\" (UniqueName: \"kubernetes.io/projected/36f09b4c-9db6-4dd3-809e-bf096ac656b2-kube-api-access-fwgf6\") pod \"package-server-manager-789f6589d5-krqxw\" (UID: \"36f09b4c-9db6-4dd3-809e-bf096ac656b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396735 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q76sv\" (UniqueName: \"kubernetes.io/projected/1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4-kube-api-access-q76sv\") pod \"console-operator-58897d9998-v987t\" (UID: \"1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4\") " pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396776 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7jpl\" (UniqueName: \"kubernetes.io/projected/8354382a-de78-4880-b344-358d49362ea8-kube-api-access-p7jpl\") pod \"olm-operator-6b444d44fb-gzh7f\" (UID: \"8354382a-de78-4880-b344-358d49362ea8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396817 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-registry-tls\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396862 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dbe09708-d737-4915-bafb-40d9b6d7897d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-w4snq\" (UID: \"dbe09708-d737-4915-bafb-40d9b6d7897d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396898 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/064dc547-a0cb-4008-8a6c-f4eb14ab288f-auth-proxy-config\") pod \"machine-approver-56656f9798-6g5xq\" (UID: \"064dc547-a0cb-4008-8a6c-f4eb14ab288f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.396929 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/64a00282-301f-4947-9795-990b7b0ef4fd-audit-dir\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397019 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/971eaf50-b0a8-4339-a2d1-6d2b9df9bc27-serving-cert\") pod \"service-ca-operator-777779d784-99jjd\" (UID: \"971eaf50-b0a8-4339-a2d1-6d2b9df9bc27\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99jjd" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397037 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/064dc547-a0cb-4008-8a6c-f4eb14ab288f-config\") pod \"machine-approver-56656f9798-6g5xq\" (UID: \"064dc547-a0cb-4008-8a6c-f4eb14ab288f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397062 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-config\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397098 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/127b156f-fe4c-40da-a576-e5b4c2ecbfdf-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-npmlx\" (UID: \"127b156f-fe4c-40da-a576-e5b4c2ecbfdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397136 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb-proxy-tls\") pod \"machine-config-operator-74547568cd-lgwbw\" (UID: \"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397169 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7f011434-93f3-4b0c-9ff7-cfbc3f333b83-certs\") pod \"machine-config-server-v9npv\" (UID: \"7f011434-93f3-4b0c-9ff7-cfbc3f333b83\") " pod="openshift-machine-config-operator/machine-config-server-v9npv" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397205 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7c0b4bf0-5b95-4ce8-946c-47e89af5d655-srv-cert\") pod \"catalog-operator-68c6474976-tgx4s\" (UID: \"7c0b4bf0-5b95-4ce8-946c-47e89af5d655\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397237 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8354382a-de78-4880-b344-358d49362ea8-srv-cert\") pod \"olm-operator-6b444d44fb-gzh7f\" (UID: \"8354382a-de78-4880-b344-358d49362ea8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397278 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m47n4\" (UniqueName: \"kubernetes.io/projected/63dc8fd2-1f20-4944-a8f1-783700e192f0-kube-api-access-m47n4\") pod \"downloads-7954f5f757-gvb66\" (UID: \"63dc8fd2-1f20-4944-a8f1-783700e192f0\") " pod="openshift-console/downloads-7954f5f757-gvb66" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397318 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg7j5\" (UniqueName: \"kubernetes.io/projected/971eaf50-b0a8-4339-a2d1-6d2b9df9bc27-kube-api-access-jg7j5\") pod \"service-ca-operator-777779d784-99jjd\" (UID: \"971eaf50-b0a8-4339-a2d1-6d2b9df9bc27\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99jjd" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397351 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-socket-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397391 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfx54\" (UniqueName: \"kubernetes.io/projected/7f011434-93f3-4b0c-9ff7-cfbc3f333b83-kube-api-access-wfx54\") pod \"machine-config-server-v9npv\" (UID: \"7f011434-93f3-4b0c-9ff7-cfbc3f333b83\") " pod="openshift-machine-config-operator/machine-config-server-v9npv" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397426 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7526f918-73a6-4470-9d8a-b79b418a8f36-metrics-tls\") pod \"ingress-operator-5b745b69d9-ph4s2\" (UID: \"7526f918-73a6-4470-9d8a-b79b418a8f36\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397467 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b746ad9f-7c63-4945-adce-aa2e0311ba30-serving-cert\") pod \"route-controller-manager-6576b87f9c-whx2n\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397500 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-service-ca\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397533 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4lcd\" (UniqueName: \"kubernetes.io/projected/ce1437a0-173e-4e15-bc0f-ca00b67e9b2e-kube-api-access-l4lcd\") pod \"packageserver-d55dfcdfc-pm5r9\" (UID: \"ce1437a0-173e-4e15-bc0f-ca00b67e9b2e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397568 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djzsp\" (UniqueName: \"kubernetes.io/projected/1159e0d6-a6e9-40f9-8f4e-607ea07bb47e-kube-api-access-djzsp\") pod \"service-ca-9c57cc56f-ptw97\" (UID: \"1159e0d6-a6e9-40f9-8f4e-607ea07bb47e\") " pod="openshift-service-ca/service-ca-9c57cc56f-ptw97" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397605 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b746ad9f-7c63-4945-adce-aa2e0311ba30-config\") pod \"route-controller-manager-6576b87f9c-whx2n\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397646 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-oauth-serving-cert\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397678 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6hzq\" (UniqueName: \"kubernetes.io/projected/2bc2207d-9081-475d-be54-85e725ea3819-kube-api-access-t6hzq\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397707 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3dffe31-ef37-4ecd-b72a-26d8541a331f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-plfk4\" (UID: \"d3dffe31-ef37-4ecd-b72a-26d8541a331f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397738 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1159e0d6-a6e9-40f9-8f4e-607ea07bb47e-signing-cabundle\") pod \"service-ca-9c57cc56f-ptw97\" (UID: \"1159e0d6-a6e9-40f9-8f4e-607ea07bb47e\") " pod="openshift-service-ca/service-ca-9c57cc56f-ptw97" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397770 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-csi-data-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397799 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-config\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397829 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1159e0d6-a6e9-40f9-8f4e-607ea07bb47e-signing-key\") pod \"service-ca-9c57cc56f-ptw97\" (UID: \"1159e0d6-a6e9-40f9-8f4e-607ea07bb47e\") " pod="openshift-service-ca/service-ca-9c57cc56f-ptw97" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397869 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50774a36-1b88-4b62-98ee-2af94e5f217e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-brt8k\" (UID: \"50774a36-1b88-4b62-98ee-2af94e5f217e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397904 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-serving-cert\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.397938 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz9kg\" (UniqueName: \"kubernetes.io/projected/1d1cb0fe-95a4-4e16-b5f5-c3847c712081-kube-api-access-cz9kg\") pod \"ingress-canary-758r2\" (UID: \"1d1cb0fe-95a4-4e16-b5f5-c3847c712081\") " pod="openshift-ingress-canary/ingress-canary-758r2" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398006 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lgwbw\" (UID: \"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398045 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b746ad9f-7c63-4945-adce-aa2e0311ba30-client-ca\") pod \"route-controller-manager-6576b87f9c-whx2n\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398085 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/36f09b4c-9db6-4dd3-809e-bf096ac656b2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-krqxw\" (UID: \"36f09b4c-9db6-4dd3-809e-bf096ac656b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398122 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/64a00282-301f-4947-9795-990b7b0ef4fd-etcd-client\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398157 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7962f6fc-2505-49ff-b14a-a90eed3e7ba5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qglml\" (UID: \"7962f6fc-2505-49ff-b14a-a90eed3e7ba5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398195 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/dbe09708-d737-4915-bafb-40d9b6d7897d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-w4snq\" (UID: \"dbe09708-d737-4915-bafb-40d9b6d7897d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398233 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfdjx\" (UniqueName: \"kubernetes.io/projected/a6754f46-92c3-422e-95d9-638efe436d8e-kube-api-access-xfdjx\") pod \"dns-operator-744455d44c-t86lg\" (UID: \"a6754f46-92c3-422e-95d9-638efe436d8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-t86lg" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398268 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/395ae874-ba30-40f2-957c-c49098b29a74-config-volume\") pod \"dns-default-fgvlq\" (UID: \"395ae874-ba30-40f2-957c-c49098b29a74\") " pod="openshift-dns/dns-default-fgvlq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398305 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4-config\") pod \"console-operator-58897d9998-v987t\" (UID: \"1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4\") " pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398377 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-bound-sa-token\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398413 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/67aa1649-ec3b-4f32-96fa-d9aece180709-stats-auth\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398446 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64a00282-301f-4947-9795-990b7b0ef4fd-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398482 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-mountpoint-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398514 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-registration-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398544 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-etcd-ca\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398579 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aec43d0e-fdc8-453c-bb71-0445e899102d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398615 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b99c4b20-1a6d-426f-b9b3-ead1b1cc4993-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2l4wc\" (UID: \"b99c4b20-1a6d-426f-b9b3-ead1b1cc4993\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2l4wc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398650 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e1b3210-4726-450d-8da5-766903e6d91b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d8qrs\" (UID: \"2e1b3210-4726-450d-8da5-766903e6d91b\") " pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398684 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzgzq\" (UniqueName: \"kubernetes.io/projected/2e1b3210-4726-450d-8da5-766903e6d91b-kube-api-access-gzgzq\") pod \"marketplace-operator-79b997595-d8qrs\" (UID: \"2e1b3210-4726-450d-8da5-766903e6d91b\") " pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398728 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/adc13fc6-a2d8-44fd-ab25-075857884cad-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.398767 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d5pq\" (UniqueName: \"kubernetes.io/projected/7526f918-73a6-4470-9d8a-b79b418a8f36-kube-api-access-2d5pq\") pod \"ingress-operator-5b745b69d9-ph4s2\" (UID: \"7526f918-73a6-4470-9d8a-b79b418a8f36\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.399298 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/adc13fc6-a2d8-44fd-ab25-075857884cad-registry-certificates\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.402132 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/adc13fc6-a2d8-44fd-ab25-075857884cad-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.403023 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/67aa1649-ec3b-4f32-96fa-d9aece180709-default-certificate\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.403335 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/67aa1649-ec3b-4f32-96fa-d9aece180709-service-ca-bundle\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.412251 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/64a00282-301f-4947-9795-990b7b0ef4fd-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.412445 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8s7ct\" (UniqueName: \"kubernetes.io/projected/382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb-kube-api-access-8s7ct\") pod \"machine-config-operator-74547568cd-lgwbw\" (UID: \"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.412479 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g74m\" (UniqueName: \"kubernetes.io/projected/b99c4b20-1a6d-426f-b9b3-ead1b1cc4993-kube-api-access-6g74m\") pod \"multus-admission-controller-857f4d67dd-2l4wc\" (UID: \"b99c4b20-1a6d-426f-b9b3-ead1b1cc4993\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2l4wc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.412529 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aec43d0e-fdc8-453c-bb71-0445e899102d-config\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.412552 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aec43d0e-fdc8-453c-bb71-0445e899102d-service-ca-bundle\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.412573 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a6754f46-92c3-422e-95d9-638efe436d8e-metrics-tls\") pod \"dns-operator-744455d44c-t86lg\" (UID: \"a6754f46-92c3-422e-95d9-638efe436d8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-t86lg" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.412596 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7962f6fc-2505-49ff-b14a-a90eed3e7ba5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qglml\" (UID: \"7962f6fc-2505-49ff-b14a-a90eed3e7ba5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.412620 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce1437a0-173e-4e15-bc0f-ca00b67e9b2e-apiservice-cert\") pod \"packageserver-d55dfcdfc-pm5r9\" (UID: \"ce1437a0-173e-4e15-bc0f-ca00b67e9b2e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.412638 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-serving-cert\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.412663 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.412694 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brqtr\" (UniqueName: \"kubernetes.io/projected/395ae874-ba30-40f2-957c-c49098b29a74-kube-api-access-brqtr\") pod \"dns-default-fgvlq\" (UID: \"395ae874-ba30-40f2-957c-c49098b29a74\") " pod="openshift-dns/dns-default-fgvlq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.412714 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/64a00282-301f-4947-9795-990b7b0ef4fd-audit-policies\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.412733 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdjcm\" (UniqueName: \"kubernetes.io/projected/64a00282-301f-4947-9795-990b7b0ef4fd-kube-api-access-sdjcm\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.412749 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/971eaf50-b0a8-4339-a2d1-6d2b9df9bc27-config\") pod \"service-ca-operator-777779d784-99jjd\" (UID: \"971eaf50-b0a8-4339-a2d1-6d2b9df9bc27\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99jjd" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.413095 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/67aa1649-ec3b-4f32-96fa-d9aece180709-stats-auth\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.413218 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adc13fc6-a2d8-44fd-ab25-075857884cad-trusted-ca\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.413289 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b746ad9f-7c63-4945-adce-aa2e0311ba30-client-ca\") pod \"route-controller-manager-6576b87f9c-whx2n\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.413511 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aec43d0e-fdc8-453c-bb71-0445e899102d-serving-cert\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:24 crc kubenswrapper[4775]: E1002 01:43:24.413629 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:24.913616661 +0000 UTC m=+142.080360701 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.413767 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/adc13fc6-a2d8-44fd-ab25-075857884cad-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.414379 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/64a00282-301f-4947-9795-990b7b0ef4fd-audit-dir\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.414810 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aec43d0e-fdc8-453c-bb71-0445e899102d-config\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.414997 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/064dc547-a0cb-4008-8a6c-f4eb14ab288f-auth-proxy-config\") pod \"machine-approver-56656f9798-6g5xq\" (UID: \"064dc547-a0cb-4008-8a6c-f4eb14ab288f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.415091 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dbe09708-d737-4915-bafb-40d9b6d7897d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-w4snq\" (UID: \"dbe09708-d737-4915-bafb-40d9b6d7897d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.415779 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/064dc547-a0cb-4008-8a6c-f4eb14ab288f-machine-approver-tls\") pod \"machine-approver-56656f9798-6g5xq\" (UID: \"064dc547-a0cb-4008-8a6c-f4eb14ab288f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.416165 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.416545 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b746ad9f-7c63-4945-adce-aa2e0311ba30-serving-cert\") pod \"route-controller-manager-6576b87f9c-whx2n\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.416901 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dbe09708-d737-4915-bafb-40d9b6d7897d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-w4snq\" (UID: \"dbe09708-d737-4915-bafb-40d9b6d7897d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.418009 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aec43d0e-fdc8-453c-bb71-0445e899102d-service-ca-bundle\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.418620 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/dbe09708-d737-4915-bafb-40d9b6d7897d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-w4snq\" (UID: \"dbe09708-d737-4915-bafb-40d9b6d7897d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.419095 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a6754f46-92c3-422e-95d9-638efe436d8e-metrics-tls\") pod \"dns-operator-744455d44c-t86lg\" (UID: \"a6754f46-92c3-422e-95d9-638efe436d8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-t86lg" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.419385 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/64a00282-301f-4947-9795-990b7b0ef4fd-serving-cert\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.420023 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/67aa1649-ec3b-4f32-96fa-d9aece180709-metrics-certs\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.420291 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/64a00282-301f-4947-9795-990b7b0ef4fd-encryption-config\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.420818 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/64a00282-301f-4947-9795-990b7b0ef4fd-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.420870 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/64a00282-301f-4947-9795-990b7b0ef4fd-audit-policies\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.421089 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-registry-tls\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.421747 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b746ad9f-7c63-4945-adce-aa2e0311ba30-config\") pod \"route-controller-manager-6576b87f9c-whx2n\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.422287 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aec43d0e-fdc8-453c-bb71-0445e899102d-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.426791 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/50774a36-1b88-4b62-98ee-2af94e5f217e-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-brt8k\" (UID: \"50774a36-1b88-4b62-98ee-2af94e5f217e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.429387 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/50774a36-1b88-4b62-98ee-2af94e5f217e-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-brt8k\" (UID: \"50774a36-1b88-4b62-98ee-2af94e5f217e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.430000 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.440143 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/64a00282-301f-4947-9795-990b7b0ef4fd-etcd-client\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.449665 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.471387 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.498420 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.510218 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516248 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516377 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-serving-cert\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516401 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz9kg\" (UniqueName: \"kubernetes.io/projected/1d1cb0fe-95a4-4e16-b5f5-c3847c712081-kube-api-access-cz9kg\") pod \"ingress-canary-758r2\" (UID: \"1d1cb0fe-95a4-4e16-b5f5-c3847c712081\") " pod="openshift-ingress-canary/ingress-canary-758r2" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516417 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lgwbw\" (UID: \"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516438 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/36f09b4c-9db6-4dd3-809e-bf096ac656b2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-krqxw\" (UID: \"36f09b4c-9db6-4dd3-809e-bf096ac656b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516456 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7962f6fc-2505-49ff-b14a-a90eed3e7ba5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qglml\" (UID: \"7962f6fc-2505-49ff-b14a-a90eed3e7ba5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516481 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/395ae874-ba30-40f2-957c-c49098b29a74-config-volume\") pod \"dns-default-fgvlq\" (UID: \"395ae874-ba30-40f2-957c-c49098b29a74\") " pod="openshift-dns/dns-default-fgvlq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516497 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4-config\") pod \"console-operator-58897d9998-v987t\" (UID: \"1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4\") " pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516518 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-mountpoint-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516536 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-registration-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516551 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-etcd-ca\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516566 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b99c4b20-1a6d-426f-b9b3-ead1b1cc4993-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2l4wc\" (UID: \"b99c4b20-1a6d-426f-b9b3-ead1b1cc4993\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2l4wc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516580 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e1b3210-4726-450d-8da5-766903e6d91b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d8qrs\" (UID: \"2e1b3210-4726-450d-8da5-766903e6d91b\") " pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516599 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzgzq\" (UniqueName: \"kubernetes.io/projected/2e1b3210-4726-450d-8da5-766903e6d91b-kube-api-access-gzgzq\") pod \"marketplace-operator-79b997595-d8qrs\" (UID: \"2e1b3210-4726-450d-8da5-766903e6d91b\") " pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516623 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d5pq\" (UniqueName: \"kubernetes.io/projected/7526f918-73a6-4470-9d8a-b79b418a8f36-kube-api-access-2d5pq\") pod \"ingress-operator-5b745b69d9-ph4s2\" (UID: \"7526f918-73a6-4470-9d8a-b79b418a8f36\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516643 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8s7ct\" (UniqueName: \"kubernetes.io/projected/382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb-kube-api-access-8s7ct\") pod \"machine-config-operator-74547568cd-lgwbw\" (UID: \"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516657 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g74m\" (UniqueName: \"kubernetes.io/projected/b99c4b20-1a6d-426f-b9b3-ead1b1cc4993-kube-api-access-6g74m\") pod \"multus-admission-controller-857f4d67dd-2l4wc\" (UID: \"b99c4b20-1a6d-426f-b9b3-ead1b1cc4993\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2l4wc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516681 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7962f6fc-2505-49ff-b14a-a90eed3e7ba5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qglml\" (UID: \"7962f6fc-2505-49ff-b14a-a90eed3e7ba5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516696 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce1437a0-173e-4e15-bc0f-ca00b67e9b2e-apiservice-cert\") pod \"packageserver-d55dfcdfc-pm5r9\" (UID: \"ce1437a0-173e-4e15-bc0f-ca00b67e9b2e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516714 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-serving-cert\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516744 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brqtr\" (UniqueName: \"kubernetes.io/projected/395ae874-ba30-40f2-957c-c49098b29a74-kube-api-access-brqtr\") pod \"dns-default-fgvlq\" (UID: \"395ae874-ba30-40f2-957c-c49098b29a74\") " pod="openshift-dns/dns-default-fgvlq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516767 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/971eaf50-b0a8-4339-a2d1-6d2b9df9bc27-config\") pod \"service-ca-operator-777779d784-99jjd\" (UID: \"971eaf50-b0a8-4339-a2d1-6d2b9df9bc27\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99jjd" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516796 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4nqd\" (UniqueName: \"kubernetes.io/projected/ff918ab3-33fd-44f2-b3ba-9f419f027b10-kube-api-access-j4nqd\") pod \"machine-config-controller-84d6567774-lg72k\" (UID: \"ff918ab3-33fd-44f2-b3ba-9f419f027b10\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516811 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3dffe31-ef37-4ecd-b72a-26d8541a331f-config\") pod \"kube-apiserver-operator-766d6c64bb-plfk4\" (UID: \"d3dffe31-ef37-4ecd-b72a-26d8541a331f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516826 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-etcd-client\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516849 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/127b156f-fe4c-40da-a576-e5b4c2ecbfdf-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-npmlx\" (UID: \"127b156f-fe4c-40da-a576-e5b4c2ecbfdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516863 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8354382a-de78-4880-b344-358d49362ea8-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gzh7f\" (UID: \"8354382a-de78-4880-b344-358d49362ea8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516879 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ff918ab3-33fd-44f2-b3ba-9f419f027b10-proxy-tls\") pod \"machine-config-controller-84d6567774-lg72k\" (UID: \"ff918ab3-33fd-44f2-b3ba-9f419f027b10\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516905 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbrvw\" (UniqueName: \"kubernetes.io/projected/115575fa-2b4f-4b7e-bdd8-46c85da3855d-kube-api-access-rbrvw\") pod \"control-plane-machine-set-operator-78cbb6b69f-wnfts\" (UID: \"115575fa-2b4f-4b7e-bdd8-46c85da3855d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wnfts" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516919 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7962f6fc-2505-49ff-b14a-a90eed3e7ba5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qglml\" (UID: \"7962f6fc-2505-49ff-b14a-a90eed3e7ba5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516934 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ff918ab3-33fd-44f2-b3ba-9f419f027b10-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lg72k\" (UID: \"ff918ab3-33fd-44f2-b3ba-9f419f027b10\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.516993 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nghmr\" (UniqueName: \"kubernetes.io/projected/7c0b4bf0-5b95-4ce8-946c-47e89af5d655-kube-api-access-nghmr\") pod \"catalog-operator-68c6474976-tgx4s\" (UID: \"7c0b4bf0-5b95-4ce8-946c-47e89af5d655\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517008 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3dffe31-ef37-4ecd-b72a-26d8541a331f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-plfk4\" (UID: \"d3dffe31-ef37-4ecd-b72a-26d8541a331f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517024 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-config-volume\") pod \"collect-profiles-29322810-q7gnj\" (UID: \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517041 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8sw5\" (UniqueName: \"kubernetes.io/projected/ad595bf8-fed5-4d02-8e6b-fbc3742e0ec2-kube-api-access-h8sw5\") pod \"migrator-59844c95c7-klcvg\" (UID: \"ad595bf8-fed5-4d02-8e6b-fbc3742e0ec2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-klcvg" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517058 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91116a6b-4da7-40db-9633-40a4fb1dcd7e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cng5\" (UID: \"91116a6b-4da7-40db-9633-40a4fb1dcd7e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517088 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce1437a0-173e-4e15-bc0f-ca00b67e9b2e-webhook-cert\") pod \"packageserver-d55dfcdfc-pm5r9\" (UID: \"ce1437a0-173e-4e15-bc0f-ca00b67e9b2e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517108 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/395ae874-ba30-40f2-957c-c49098b29a74-metrics-tls\") pod \"dns-default-fgvlq\" (UID: \"395ae874-ba30-40f2-957c-c49098b29a74\") " pod="openshift-dns/dns-default-fgvlq" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517131 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ce1437a0-173e-4e15-bc0f-ca00b67e9b2e-tmpfs\") pod \"packageserver-d55dfcdfc-pm5r9\" (UID: \"ce1437a0-173e-4e15-bc0f-ca00b67e9b2e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517148 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tcqs\" (UniqueName: \"kubernetes.io/projected/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-kube-api-access-6tcqs\") pod \"collect-profiles-29322810-q7gnj\" (UID: \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517164 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-etcd-service-ca\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517195 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph7dl\" (UniqueName: \"kubernetes.io/projected/91116a6b-4da7-40db-9633-40a4fb1dcd7e-kube-api-access-ph7dl\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cng5\" (UID: \"91116a6b-4da7-40db-9633-40a4fb1dcd7e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517222 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/115575fa-2b4f-4b7e-bdd8-46c85da3855d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-wnfts\" (UID: \"115575fa-2b4f-4b7e-bdd8-46c85da3855d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wnfts" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517242 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-oauth-config\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517259 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7526f918-73a6-4470-9d8a-b79b418a8f36-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ph4s2\" (UID: \"7526f918-73a6-4470-9d8a-b79b418a8f36\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517276 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-plugins-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517292 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/127b156f-fe4c-40da-a576-e5b4c2ecbfdf-config\") pod \"kube-controller-manager-operator-78b949d7b-npmlx\" (UID: \"127b156f-fe4c-40da-a576-e5b4c2ecbfdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517309 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4xck\" (UniqueName: \"kubernetes.io/projected/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-kube-api-access-g4xck\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517333 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7526f918-73a6-4470-9d8a-b79b418a8f36-trusted-ca\") pod \"ingress-operator-5b745b69d9-ph4s2\" (UID: \"7526f918-73a6-4470-9d8a-b79b418a8f36\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517351 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb-images\") pod \"machine-config-operator-74547568cd-lgwbw\" (UID: \"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517368 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-secret-volume\") pod \"collect-profiles-29322810-q7gnj\" (UID: \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517383 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7f011434-93f3-4b0c-9ff7-cfbc3f333b83-node-bootstrap-token\") pod \"machine-config-server-v9npv\" (UID: \"7f011434-93f3-4b0c-9ff7-cfbc3f333b83\") " pod="openshift-machine-config-operator/machine-config-server-v9npv" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517412 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7c0b4bf0-5b95-4ce8-946c-47e89af5d655-profile-collector-cert\") pod \"catalog-operator-68c6474976-tgx4s\" (UID: \"7c0b4bf0-5b95-4ce8-946c-47e89af5d655\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517428 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4-trusted-ca\") pod \"console-operator-58897d9998-v987t\" (UID: \"1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4\") " pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517443 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1d1cb0fe-95a4-4e16-b5f5-c3847c712081-cert\") pod \"ingress-canary-758r2\" (UID: \"1d1cb0fe-95a4-4e16-b5f5-c3847c712081\") " pod="openshift-ingress-canary/ingress-canary-758r2" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517458 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fz5g\" (UniqueName: \"kubernetes.io/projected/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-kube-api-access-7fz5g\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517481 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91116a6b-4da7-40db-9633-40a4fb1dcd7e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cng5\" (UID: \"91116a6b-4da7-40db-9633-40a4fb1dcd7e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517484 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-mountpoint-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517497 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2e1b3210-4726-450d-8da5-766903e6d91b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d8qrs\" (UID: \"2e1b3210-4726-450d-8da5-766903e6d91b\") " pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517622 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4-serving-cert\") pod \"console-operator-58897d9998-v987t\" (UID: \"1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4\") " pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517706 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-trusted-ca-bundle\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517741 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwgf6\" (UniqueName: \"kubernetes.io/projected/36f09b4c-9db6-4dd3-809e-bf096ac656b2-kube-api-access-fwgf6\") pod \"package-server-manager-789f6589d5-krqxw\" (UID: \"36f09b4c-9db6-4dd3-809e-bf096ac656b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517767 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q76sv\" (UniqueName: \"kubernetes.io/projected/1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4-kube-api-access-q76sv\") pod \"console-operator-58897d9998-v987t\" (UID: \"1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4\") " pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517807 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7jpl\" (UniqueName: \"kubernetes.io/projected/8354382a-de78-4880-b344-358d49362ea8-kube-api-access-p7jpl\") pod \"olm-operator-6b444d44fb-gzh7f\" (UID: \"8354382a-de78-4880-b344-358d49362ea8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517877 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/971eaf50-b0a8-4339-a2d1-6d2b9df9bc27-serving-cert\") pod \"service-ca-operator-777779d784-99jjd\" (UID: \"971eaf50-b0a8-4339-a2d1-6d2b9df9bc27\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99jjd" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517909 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-config\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.517944 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/127b156f-fe4c-40da-a576-e5b4c2ecbfdf-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-npmlx\" (UID: \"127b156f-fe4c-40da-a576-e5b4c2ecbfdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.518006 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb-proxy-tls\") pod \"machine-config-operator-74547568cd-lgwbw\" (UID: \"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.518039 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7f011434-93f3-4b0c-9ff7-cfbc3f333b83-certs\") pod \"machine-config-server-v9npv\" (UID: \"7f011434-93f3-4b0c-9ff7-cfbc3f333b83\") " pod="openshift-machine-config-operator/machine-config-server-v9npv" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.518083 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7c0b4bf0-5b95-4ce8-946c-47e89af5d655-srv-cert\") pod \"catalog-operator-68c6474976-tgx4s\" (UID: \"7c0b4bf0-5b95-4ce8-946c-47e89af5d655\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.518107 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8354382a-de78-4880-b344-358d49362ea8-srv-cert\") pod \"olm-operator-6b444d44fb-gzh7f\" (UID: \"8354382a-de78-4880-b344-358d49362ea8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.518153 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg7j5\" (UniqueName: \"kubernetes.io/projected/971eaf50-b0a8-4339-a2d1-6d2b9df9bc27-kube-api-access-jg7j5\") pod \"service-ca-operator-777779d784-99jjd\" (UID: \"971eaf50-b0a8-4339-a2d1-6d2b9df9bc27\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99jjd" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.518191 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-socket-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.518218 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfx54\" (UniqueName: \"kubernetes.io/projected/7f011434-93f3-4b0c-9ff7-cfbc3f333b83-kube-api-access-wfx54\") pod \"machine-config-server-v9npv\" (UID: \"7f011434-93f3-4b0c-9ff7-cfbc3f333b83\") " pod="openshift-machine-config-operator/machine-config-server-v9npv" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.518254 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7526f918-73a6-4470-9d8a-b79b418a8f36-metrics-tls\") pod \"ingress-operator-5b745b69d9-ph4s2\" (UID: \"7526f918-73a6-4470-9d8a-b79b418a8f36\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.518290 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-service-ca\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.519506 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-lgwbw\" (UID: \"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" Oct 02 01:43:24 crc kubenswrapper[4775]: E1002 01:43:24.519614 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:25.01959419 +0000 UTC m=+142.186338470 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.519666 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4-config\") pod \"console-operator-58897d9998-v987t\" (UID: \"1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4\") " pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.519694 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e1b3210-4726-450d-8da5-766903e6d91b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d8qrs\" (UID: \"2e1b3210-4726-450d-8da5-766903e6d91b\") " pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.519772 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-etcd-ca\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.520084 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-registration-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.520447 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7962f6fc-2505-49ff-b14a-a90eed3e7ba5-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qglml\" (UID: \"7962f6fc-2505-49ff-b14a-a90eed3e7ba5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.520750 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/971eaf50-b0a8-4339-a2d1-6d2b9df9bc27-config\") pod \"service-ca-operator-777779d784-99jjd\" (UID: \"971eaf50-b0a8-4339-a2d1-6d2b9df9bc27\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99jjd" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.521108 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-socket-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.521603 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4lcd\" (UniqueName: \"kubernetes.io/projected/ce1437a0-173e-4e15-bc0f-ca00b67e9b2e-kube-api-access-l4lcd\") pod \"packageserver-d55dfcdfc-pm5r9\" (UID: \"ce1437a0-173e-4e15-bc0f-ca00b67e9b2e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.521640 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djzsp\" (UniqueName: \"kubernetes.io/projected/1159e0d6-a6e9-40f9-8f4e-607ea07bb47e-kube-api-access-djzsp\") pod \"service-ca-9c57cc56f-ptw97\" (UID: \"1159e0d6-a6e9-40f9-8f4e-607ea07bb47e\") " pod="openshift-service-ca/service-ca-9c57cc56f-ptw97" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.521680 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-oauth-serving-cert\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.521712 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6hzq\" (UniqueName: \"kubernetes.io/projected/2bc2207d-9081-475d-be54-85e725ea3819-kube-api-access-t6hzq\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.521741 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3dffe31-ef37-4ecd-b72a-26d8541a331f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-plfk4\" (UID: \"d3dffe31-ef37-4ecd-b72a-26d8541a331f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.521769 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1159e0d6-a6e9-40f9-8f4e-607ea07bb47e-signing-cabundle\") pod \"service-ca-9c57cc56f-ptw97\" (UID: \"1159e0d6-a6e9-40f9-8f4e-607ea07bb47e\") " pod="openshift-service-ca/service-ca-9c57cc56f-ptw97" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.521834 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-csi-data-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.521866 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-config\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.521891 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1159e0d6-a6e9-40f9-8f4e-607ea07bb47e-signing-key\") pod \"service-ca-9c57cc56f-ptw97\" (UID: \"1159e0d6-a6e9-40f9-8f4e-607ea07bb47e\") " pod="openshift-service-ca/service-ca-9c57cc56f-ptw97" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.522006 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ff918ab3-33fd-44f2-b3ba-9f419f027b10-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lg72k\" (UID: \"ff918ab3-33fd-44f2-b3ba-9f419f027b10\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.522275 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7962f6fc-2505-49ff-b14a-a90eed3e7ba5-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qglml\" (UID: \"7962f6fc-2505-49ff-b14a-a90eed3e7ba5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.523045 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb-images\") pod \"machine-config-operator-74547568cd-lgwbw\" (UID: \"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.523454 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/ce1437a0-173e-4e15-bc0f-ca00b67e9b2e-tmpfs\") pod \"packageserver-d55dfcdfc-pm5r9\" (UID: \"ce1437a0-173e-4e15-bc0f-ca00b67e9b2e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.523503 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-plugins-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.524196 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-etcd-service-ca\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.524338 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4-trusted-ca\") pod \"console-operator-58897d9998-v987t\" (UID: \"1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4\") " pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.526668 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/ff918ab3-33fd-44f2-b3ba-9f419f027b10-proxy-tls\") pod \"machine-config-controller-84d6567774-lg72k\" (UID: \"ff918ab3-33fd-44f2-b3ba-9f419f027b10\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.526683 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/1159e0d6-a6e9-40f9-8f4e-607ea07bb47e-signing-cabundle\") pod \"service-ca-9c57cc56f-ptw97\" (UID: \"1159e0d6-a6e9-40f9-8f4e-607ea07bb47e\") " pod="openshift-service-ca/service-ca-9c57cc56f-ptw97" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.526798 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2bc2207d-9081-475d-be54-85e725ea3819-csi-data-dir\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.527367 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/1159e0d6-a6e9-40f9-8f4e-607ea07bb47e-signing-key\") pod \"service-ca-9c57cc56f-ptw97\" (UID: \"1159e0d6-a6e9-40f9-8f4e-607ea07bb47e\") " pod="openshift-service-ca/service-ca-9c57cc56f-ptw97" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.527581 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-config\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.527855 4775 request.go:700] Waited for 1.006006712s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.528095 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-serving-cert\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.528784 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7c0b4bf0-5b95-4ce8-946c-47e89af5d655-profile-collector-cert\") pod \"catalog-operator-68c6474976-tgx4s\" (UID: \"7c0b4bf0-5b95-4ce8-946c-47e89af5d655\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.529188 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/971eaf50-b0a8-4339-a2d1-6d2b9df9bc27-serving-cert\") pod \"service-ca-operator-777779d784-99jjd\" (UID: \"971eaf50-b0a8-4339-a2d1-6d2b9df9bc27\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99jjd" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.529593 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.529682 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b99c4b20-1a6d-426f-b9b3-ead1b1cc4993-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-2l4wc\" (UID: \"b99c4b20-1a6d-426f-b9b3-ead1b1cc4993\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2l4wc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.530454 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4-serving-cert\") pod \"console-operator-58897d9998-v987t\" (UID: \"1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4\") " pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.530582 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-secret-volume\") pod \"collect-profiles-29322810-q7gnj\" (UID: \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.531678 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8354382a-de78-4880-b344-358d49362ea8-srv-cert\") pod \"olm-operator-6b444d44fb-gzh7f\" (UID: \"8354382a-de78-4880-b344-358d49362ea8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.531917 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8354382a-de78-4880-b344-358d49362ea8-profile-collector-cert\") pod \"olm-operator-6b444d44fb-gzh7f\" (UID: \"8354382a-de78-4880-b344-358d49362ea8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.532741 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" event={"ID":"958d5b03-aa2b-4068-b6fc-6efb125179a0","Type":"ContainerStarted","Data":"22aed9f6b23212ce4efa34308b8a1cd085767d604400638e6397138b386e7d06"} Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.532787 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" event={"ID":"958d5b03-aa2b-4068-b6fc-6efb125179a0","Type":"ContainerStarted","Data":"6a2e72f8bf06b03afebf4b750d220a30ca21d7fbaaace4e62af8ea6a5315a122"} Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.533303 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-etcd-client\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.537255 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb-proxy-tls\") pod \"machine-config-operator-74547568cd-lgwbw\" (UID: \"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.538729 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" event={"ID":"f5887047-3348-478d-a1d5-694762ac0dec","Type":"ContainerStarted","Data":"42121b862c9a4f5b5dfd180b378dfe19328d9c496506c9ce1daee560238b75d2"} Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.538792 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" event={"ID":"f5887047-3348-478d-a1d5-694762ac0dec","Type":"ContainerStarted","Data":"4fa8a584d1c579889578ba5f50b0476374d34aef04d0ba436b5387285bb6ae2e"} Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.549719 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.560516 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2e1b3210-4726-450d-8da5-766903e6d91b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d8qrs\" (UID: \"2e1b3210-4726-450d-8da5-766903e6d91b\") " pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.569759 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.589353 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.610140 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.623088 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: E1002 01:43:24.623503 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:25.12348709 +0000 UTC m=+142.290231130 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.630691 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.644840 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/127b156f-fe4c-40da-a576-e5b4c2ecbfdf-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-npmlx\" (UID: \"127b156f-fe4c-40da-a576-e5b4c2ecbfdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.650486 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.654707 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/127b156f-fe4c-40da-a576-e5b4c2ecbfdf-config\") pod \"kube-controller-manager-operator-78b949d7b-npmlx\" (UID: \"127b156f-fe4c-40da-a576-e5b4c2ecbfdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.671310 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.689658 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.695366 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7c0b4bf0-5b95-4ce8-946c-47e89af5d655-srv-cert\") pod \"catalog-operator-68c6474976-tgx4s\" (UID: \"7c0b4bf0-5b95-4ce8-946c-47e89af5d655\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.708971 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.722167 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t"] Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.724103 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:24 crc kubenswrapper[4775]: E1002 01:43:24.724300 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:25.224274103 +0000 UTC m=+142.391018153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.724796 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: E1002 01:43:24.725101 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:25.225088396 +0000 UTC m=+142.391832436 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.733384 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 02 01:43:24 crc kubenswrapper[4775]: W1002 01:43:24.734097 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59534b10_69de_4d5d_ba59_27c4f24df1a7.slice/crio-720a8ca4dc3400ce84e9a033a89ecfb81997030ed58387358dba93bc2aa26aab WatchSource:0}: Error finding container 720a8ca4dc3400ce84e9a033a89ecfb81997030ed58387358dba93bc2aa26aab: Status 404 returned error can't find the container with id 720a8ca4dc3400ce84e9a033a89ecfb81997030ed58387358dba93bc2aa26aab Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.734808 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899"] Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.746874 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7526f918-73a6-4470-9d8a-b79b418a8f36-metrics-tls\") pod \"ingress-operator-5b745b69d9-ph4s2\" (UID: \"7526f918-73a6-4470-9d8a-b79b418a8f36\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.778748 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.780850 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.786906 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7526f918-73a6-4470-9d8a-b79b418a8f36-trusted-ca\") pod \"ingress-operator-5b745b69d9-ph4s2\" (UID: \"7526f918-73a6-4470-9d8a-b79b418a8f36\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.791717 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.793984 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-config-volume\") pod \"collect-profiles-29322810-q7gnj\" (UID: \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.817621 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9"] Oct 02 01:43:24 crc kubenswrapper[4775]: W1002 01:43:24.823536 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e44853a_a5c9_4e24_9d55_4a8952561c98.slice/crio-df1c9466d29fd9f41972e0f355dda57ef9bc40a4e9cea5d021726c48e1a57905 WatchSource:0}: Error finding container df1c9466d29fd9f41972e0f355dda57ef9bc40a4e9cea5d021726c48e1a57905: Status 404 returned error can't find the container with id df1c9466d29fd9f41972e0f355dda57ef9bc40a4e9cea5d021726c48e1a57905 Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.829718 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:24 crc kubenswrapper[4775]: E1002 01:43:24.830362 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:25.330320103 +0000 UTC m=+142.497064183 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.836619 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c69pf\" (UniqueName: \"kubernetes.io/projected/3e6497ca-0523-4658-8cbd-2340ed601e77-kube-api-access-c69pf\") pod \"oauth-openshift-558db77b4-7gfcf\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.849447 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.855523 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlzvz\" (UniqueName: \"kubernetes.io/projected/4b13e137-25af-400d-a3c6-496fa5138392-kube-api-access-tlzvz\") pod \"apiserver-76f77b778f-47dws\" (UID: \"4b13e137-25af-400d-a3c6-496fa5138392\") " pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.856068 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/115575fa-2b4f-4b7e-bdd8-46c85da3855d-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-wnfts\" (UID: \"115575fa-2b4f-4b7e-bdd8-46c85da3855d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wnfts" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.870468 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.881048 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ce1437a0-173e-4e15-bc0f-ca00b67e9b2e-webhook-cert\") pod \"packageserver-d55dfcdfc-pm5r9\" (UID: \"ce1437a0-173e-4e15-bc0f-ca00b67e9b2e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.881101 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ce1437a0-173e-4e15-bc0f-ca00b67e9b2e-apiservice-cert\") pod \"packageserver-d55dfcdfc-pm5r9\" (UID: \"ce1437a0-173e-4e15-bc0f-ca00b67e9b2e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.890017 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.910311 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.919535 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.929585 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.930232 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.931694 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:24 crc kubenswrapper[4775]: E1002 01:43:24.932095 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:25.432078723 +0000 UTC m=+142.598822763 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.936038 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/36f09b4c-9db6-4dd3-809e-bf096ac656b2-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-krqxw\" (UID: \"36f09b4c-9db6-4dd3-809e-bf096ac656b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.950049 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.970788 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.978123 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/91116a6b-4da7-40db-9633-40a4fb1dcd7e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cng5\" (UID: \"91116a6b-4da7-40db-9633-40a4fb1dcd7e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5" Oct 02 01:43:24 crc kubenswrapper[4775]: I1002 01:43:24.991522 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.010417 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.020741 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d3dffe31-ef37-4ecd-b72a-26d8541a331f-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-plfk4\" (UID: \"d3dffe31-ef37-4ecd-b72a-26d8541a331f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.030394 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.033207 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.033421 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:25.53339067 +0000 UTC m=+142.700134710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.033617 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.034102 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:25.5340877 +0000 UTC m=+142.700831750 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.050025 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.060125 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3dffe31-ef37-4ecd-b72a-26d8541a331f-config\") pod \"kube-apiserver-operator-766d6c64bb-plfk4\" (UID: \"d3dffe31-ef37-4ecd-b72a-26d8541a331f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.070003 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.092089 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.096217 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/91116a6b-4da7-40db-9633-40a4fb1dcd7e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cng5\" (UID: \"91116a6b-4da7-40db-9633-40a4fb1dcd7e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.109845 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.129268 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.130493 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-47dws"] Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.135522 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.135548 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-oauth-serving-cert\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.135714 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:25.635675284 +0000 UTC m=+142.802419384 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.136164 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.136618 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:25.63660173 +0000 UTC m=+142.803345790 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.149226 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.154685 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-serving-cert\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.169514 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.177262 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-oauth-config\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.183828 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gfcf"] Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.189548 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.195727 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-service-ca\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.212104 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.226906 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-config\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.238221 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.238679 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:25.738564506 +0000 UTC m=+142.905308566 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.238948 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.239410 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:25.739387379 +0000 UTC m=+142.906131509 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.241997 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.249396 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.254385 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-trusted-ca-bundle\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.270303 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.294460 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.312504 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.317761 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1d1cb0fe-95a4-4e16-b5f5-c3847c712081-cert\") pod \"ingress-canary-758r2\" (UID: \"1d1cb0fe-95a4-4e16-b5f5-c3847c712081\") " pod="openshift-ingress-canary/ingress-canary-758r2" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.330614 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.340067 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.340593 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:25.840577923 +0000 UTC m=+143.007321963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.350387 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.351201 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/395ae874-ba30-40f2-957c-c49098b29a74-config-volume\") pod \"dns-default-fgvlq\" (UID: \"395ae874-ba30-40f2-957c-c49098b29a74\") " pod="openshift-dns/dns-default-fgvlq" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.369533 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.390739 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.398078 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/395ae874-ba30-40f2-957c-c49098b29a74-metrics-tls\") pod \"dns-default-fgvlq\" (UID: \"395ae874-ba30-40f2-957c-c49098b29a74\") " pod="openshift-dns/dns-default-fgvlq" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.410259 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.429536 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.434213 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/7f011434-93f3-4b0c-9ff7-cfbc3f333b83-certs\") pod \"machine-config-server-v9npv\" (UID: \"7f011434-93f3-4b0c-9ff7-cfbc3f333b83\") " pod="openshift-machine-config-operator/machine-config-server-v9npv" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.441687 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.442089 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:25.942074126 +0000 UTC m=+143.108818156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.449031 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.459649 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/7f011434-93f3-4b0c-9ff7-cfbc3f333b83-node-bootstrap-token\") pod \"machine-config-server-v9npv\" (UID: \"7f011434-93f3-4b0c-9ff7-cfbc3f333b83\") " pod="openshift-machine-config-operator/machine-config-server-v9npv" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.468761 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.489125 4775 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.509543 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.528833 4775 request.go:700] Waited for 1.416866538s due to client-side throttling, not priority and fairness, request: PATCH:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns-operator/pods/dns-operator-744455d44c-t86lg/status Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.542482 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.542629 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.042606311 +0000 UTC m=+143.209350351 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.543058 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.543326 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.043319291 +0000 UTC m=+143.210063321 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.544662 4775 generic.go:334] "Generic (PLEG): container finished" podID="59534b10-69de-4d5d-ba59-27c4f24df1a7" containerID="e12ac7c6b5d128cb0b08f1d06989b02b46e1601b1a43aa1b9a5a5fb519f3c53d" exitCode=0 Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.544706 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" event={"ID":"59534b10-69de-4d5d-ba59-27c4f24df1a7","Type":"ContainerDied","Data":"e12ac7c6b5d128cb0b08f1d06989b02b46e1601b1a43aa1b9a5a5fb519f3c53d"} Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.544758 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" event={"ID":"59534b10-69de-4d5d-ba59-27c4f24df1a7","Type":"ContainerStarted","Data":"720a8ca4dc3400ce84e9a033a89ecfb81997030ed58387358dba93bc2aa26aab"} Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.547212 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" event={"ID":"958d5b03-aa2b-4068-b6fc-6efb125179a0","Type":"ContainerStarted","Data":"479785ddaf15b49ca851b6c2ab7b309586a9190f988d8eb5f9acb32ecdf49468"} Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.548796 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899" event={"ID":"a0d23779-3169-4e6a-9d8e-3760b45a145e","Type":"ContainerStarted","Data":"a0eab07f4e37dda3ef31956793755fe1e6e17b974fdb09b39f68e4ef79434e6d"} Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.548833 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899" event={"ID":"a0d23779-3169-4e6a-9d8e-3760b45a145e","Type":"ContainerStarted","Data":"aa71f7f831896854c478c5bb4fc38f655401169399f28d7ec49a26e1004a161d"} Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.548844 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899" event={"ID":"a0d23779-3169-4e6a-9d8e-3760b45a145e","Type":"ContainerStarted","Data":"5590440597b54a3c781df985277f4dc9daf9f3e7bd456592c9b0fd78dc72b9fe"} Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.549668 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" event={"ID":"3e6497ca-0523-4658-8cbd-2340ed601e77","Type":"ContainerStarted","Data":"a1537427439d99ae2e6f244c1fc513b34dded7d7d394eac4542383d63474a42c"} Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.551129 4775 generic.go:334] "Generic (PLEG): container finished" podID="4b13e137-25af-400d-a3c6-496fa5138392" containerID="cbce01370d487f5f889b53d957f18315f3f44a64f90c99108ced1f4e945c6def" exitCode=0 Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.551159 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-47dws" event={"ID":"4b13e137-25af-400d-a3c6-496fa5138392","Type":"ContainerDied","Data":"cbce01370d487f5f889b53d957f18315f3f44a64f90c99108ced1f4e945c6def"} Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.551179 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-47dws" event={"ID":"4b13e137-25af-400d-a3c6-496fa5138392","Type":"ContainerStarted","Data":"6c8fcf8b57b29c40562e2a5e970d9324e85917e31ea2acb207f017608e0d3d01"} Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.552277 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9" event={"ID":"6e44853a-a5c9-4e24-9d55-4a8952561c98","Type":"ContainerStarted","Data":"a493fa92c9774cc9b593895b9f81c3f3056f413c7b80ce7acbf7dadab2c99bde"} Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.552307 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9" event={"ID":"6e44853a-a5c9-4e24-9d55-4a8952561c98","Type":"ContainerStarted","Data":"df1c9466d29fd9f41972e0f355dda57ef9bc40a4e9cea5d021726c48e1a57905"} Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.552387 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.561336 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.571408 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qndkx\" (UniqueName: \"kubernetes.io/projected/67aa1649-ec3b-4f32-96fa-d9aece180709-kube-api-access-qndkx\") pod \"router-default-5444994796-cn7bz\" (UID: \"67aa1649-ec3b-4f32-96fa-d9aece180709\") " pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.584517 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8g6d\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-kube-api-access-d8g6d\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.616589 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dflwc\" (UniqueName: \"kubernetes.io/projected/dbe09708-d737-4915-bafb-40d9b6d7897d-kube-api-access-dflwc\") pod \"cluster-image-registry-operator-dc59b4c8b-w4snq\" (UID: \"dbe09708-d737-4915-bafb-40d9b6d7897d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.626437 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j6b7\" (UniqueName: \"kubernetes.io/projected/064dc547-a0cb-4008-8a6c-f4eb14ab288f-kube-api-access-4j6b7\") pod \"machine-approver-56656f9798-6g5xq\" (UID: \"064dc547-a0cb-4008-8a6c-f4eb14ab288f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.643661 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.643840 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.143816705 +0000 UTC m=+143.310560735 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.644237 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.646926 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.146910833 +0000 UTC m=+143.313654873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.666044 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlrmc\" (UniqueName: \"kubernetes.io/projected/b746ad9f-7c63-4945-adce-aa2e0311ba30-kube-api-access-rlrmc\") pod \"route-controller-manager-6576b87f9c-whx2n\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.668099 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" Oct 02 01:43:25 crc kubenswrapper[4775]: W1002 01:43:25.682620 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod064dc547_a0cb_4008_8a6c_f4eb14ab288f.slice/crio-559bc0aff12f9ec164935ad3bd1f8952f9a69f3152ac3707dadc412e649c623e WatchSource:0}: Error finding container 559bc0aff12f9ec164935ad3bd1f8952f9a69f3152ac3707dadc412e649c623e: Status 404 returned error can't find the container with id 559bc0aff12f9ec164935ad3bd1f8952f9a69f3152ac3707dadc412e649c623e Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.683791 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpjsg\" (UniqueName: \"kubernetes.io/projected/50774a36-1b88-4b62-98ee-2af94e5f217e-kube-api-access-tpjsg\") pod \"openshift-controller-manager-operator-756b6f6bc6-brt8k\" (UID: \"50774a36-1b88-4b62-98ee-2af94e5f217e\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.689424 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xswnr\" (UniqueName: \"kubernetes.io/projected/aec43d0e-fdc8-453c-bb71-0445e899102d-kube-api-access-xswnr\") pod \"authentication-operator-69f744f599-bxp9r\" (UID: \"aec43d0e-fdc8-453c-bb71-0445e899102d\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.707604 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfdjx\" (UniqueName: \"kubernetes.io/projected/a6754f46-92c3-422e-95d9-638efe436d8e-kube-api-access-xfdjx\") pod \"dns-operator-744455d44c-t86lg\" (UID: \"a6754f46-92c3-422e-95d9-638efe436d8e\") " pod="openshift-dns-operator/dns-operator-744455d44c-t86lg" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.708508 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.727265 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-bound-sa-token\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.744645 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m47n4\" (UniqueName: \"kubernetes.io/projected/63dc8fd2-1f20-4944-a8f1-783700e192f0-kube-api-access-m47n4\") pod \"downloads-7954f5f757-gvb66\" (UID: \"63dc8fd2-1f20-4944-a8f1-783700e192f0\") " pod="openshift-console/downloads-7954f5f757-gvb66" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.745257 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.745428 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.2454 +0000 UTC m=+143.412144040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.745469 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.745777 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.245770211 +0000 UTC m=+143.412514251 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.768700 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdjcm\" (UniqueName: \"kubernetes.io/projected/64a00282-301f-4947-9795-990b7b0ef4fd-kube-api-access-sdjcm\") pod \"apiserver-7bbb656c7d-kqzhr\" (UID: \"64a00282-301f-4947-9795-990b7b0ef4fd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.788932 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dbe09708-d737-4915-bafb-40d9b6d7897d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-w4snq\" (UID: \"dbe09708-d737-4915-bafb-40d9b6d7897d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.803976 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g74m\" (UniqueName: \"kubernetes.io/projected/b99c4b20-1a6d-426f-b9b3-ead1b1cc4993-kube-api-access-6g74m\") pod \"multus-admission-controller-857f4d67dd-2l4wc\" (UID: \"b99c4b20-1a6d-426f-b9b3-ead1b1cc4993\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-2l4wc" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.812764 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.828496 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbrvw\" (UniqueName: \"kubernetes.io/projected/115575fa-2b4f-4b7e-bdd8-46c85da3855d-kube-api-access-rbrvw\") pod \"control-plane-machine-set-operator-78cbb6b69f-wnfts\" (UID: \"115575fa-2b4f-4b7e-bdd8-46c85da3855d\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wnfts" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.838490 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.845606 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.846309 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.846431 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.346414849 +0000 UTC m=+143.513158889 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.846604 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.846999 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.346988345 +0000 UTC m=+143.513732375 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.855208 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7962f6fc-2505-49ff-b14a-a90eed3e7ba5-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-qglml\" (UID: \"7962f6fc-2505-49ff-b14a-a90eed3e7ba5\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.862233 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.871286 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brqtr\" (UniqueName: \"kubernetes.io/projected/395ae874-ba30-40f2-957c-c49098b29a74-kube-api-access-brqtr\") pod \"dns-default-fgvlq\" (UID: \"395ae874-ba30-40f2-957c-c49098b29a74\") " pod="openshift-dns/dns-default-fgvlq" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.876781 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.881031 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-fgvlq" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.886991 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz9kg\" (UniqueName: \"kubernetes.io/projected/1d1cb0fe-95a4-4e16-b5f5-c3847c712081-kube-api-access-cz9kg\") pod \"ingress-canary-758r2\" (UID: \"1d1cb0fe-95a4-4e16-b5f5-c3847c712081\") " pod="openshift-ingress-canary/ingress-canary-758r2" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.921513 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4nqd\" (UniqueName: \"kubernetes.io/projected/ff918ab3-33fd-44f2-b3ba-9f419f027b10-kube-api-access-j4nqd\") pod \"machine-config-controller-84d6567774-lg72k\" (UID: \"ff918ab3-33fd-44f2-b3ba-9f419f027b10\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.942607 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7jpl\" (UniqueName: \"kubernetes.io/projected/8354382a-de78-4880-b344-358d49362ea8-kube-api-access-p7jpl\") pod \"olm-operator-6b444d44fb-gzh7f\" (UID: \"8354382a-de78-4880-b344-358d49362ea8\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.948751 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:25 crc kubenswrapper[4775]: E1002 01:43:25.949169 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.449155667 +0000 UTC m=+143.615899707 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.953521 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwgf6\" (UniqueName: \"kubernetes.io/projected/36f09b4c-9db6-4dd3-809e-bf096ac656b2-kube-api-access-fwgf6\") pod \"package-server-manager-789f6589d5-krqxw\" (UID: \"36f09b4c-9db6-4dd3-809e-bf096ac656b2\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.988163 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q76sv\" (UniqueName: \"kubernetes.io/projected/1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4-kube-api-access-q76sv\") pod \"console-operator-58897d9998-v987t\" (UID: \"1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4\") " pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:25 crc kubenswrapper[4775]: I1002 01:43:25.995158 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-t86lg" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.004316 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-gvb66" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.021252 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.028822 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d5pq\" (UniqueName: \"kubernetes.io/projected/7526f918-73a6-4470-9d8a-b79b418a8f36-kube-api-access-2d5pq\") pod \"ingress-operator-5b745b69d9-ph4s2\" (UID: \"7526f918-73a6-4470-9d8a-b79b418a8f36\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.028830 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzgzq\" (UniqueName: \"kubernetes.io/projected/2e1b3210-4726-450d-8da5-766903e6d91b-kube-api-access-gzgzq\") pod \"marketplace-operator-79b997595-d8qrs\" (UID: \"2e1b3210-4726-450d-8da5-766903e6d91b\") " pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.049187 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.049639 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.066141 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.066649 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:26 crc kubenswrapper[4775]: E1002 01:43:26.067081 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.567066812 +0000 UTC m=+143.733810852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.067415 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-2l4wc" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.072811 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8s7ct\" (UniqueName: \"kubernetes.io/projected/382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb-kube-api-access-8s7ct\") pod \"machine-config-operator-74547568cd-lgwbw\" (UID: \"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.090612 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.095408 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/127b156f-fe4c-40da-a576-e5b4c2ecbfdf-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-npmlx\" (UID: \"127b156f-fe4c-40da-a576-e5b4c2ecbfdf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.097109 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.101297 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg7j5\" (UniqueName: \"kubernetes.io/projected/971eaf50-b0a8-4339-a2d1-6d2b9df9bc27-kube-api-access-jg7j5\") pod \"service-ca-operator-777779d784-99jjd\" (UID: \"971eaf50-b0a8-4339-a2d1-6d2b9df9bc27\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-99jjd" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.110680 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfx54\" (UniqueName: \"kubernetes.io/projected/7f011434-93f3-4b0c-9ff7-cfbc3f333b83-kube-api-access-wfx54\") pod \"machine-config-server-v9npv\" (UID: \"7f011434-93f3-4b0c-9ff7-cfbc3f333b83\") " pod="openshift-machine-config-operator/machine-config-server-v9npv" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.126672 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wnfts" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.129266 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nghmr\" (UniqueName: \"kubernetes.io/projected/7c0b4bf0-5b95-4ce8-946c-47e89af5d655-kube-api-access-nghmr\") pod \"catalog-operator-68c6474976-tgx4s\" (UID: \"7c0b4bf0-5b95-4ce8-946c-47e89af5d655\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.132426 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3dffe31-ef37-4ecd-b72a-26d8541a331f-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-plfk4\" (UID: \"d3dffe31-ef37-4ecd-b72a-26d8541a331f\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.141204 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.153039 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8sw5\" (UniqueName: \"kubernetes.io/projected/ad595bf8-fed5-4d02-8e6b-fbc3742e0ec2-kube-api-access-h8sw5\") pod \"migrator-59844c95c7-klcvg\" (UID: \"ad595bf8-fed5-4d02-8e6b-fbc3742e0ec2\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-klcvg" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.157369 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.171706 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph7dl\" (UniqueName: \"kubernetes.io/projected/91116a6b-4da7-40db-9633-40a4fb1dcd7e-kube-api-access-ph7dl\") pod \"kube-storage-version-migrator-operator-b67b599dd-9cng5\" (UID: \"91116a6b-4da7-40db-9633-40a4fb1dcd7e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.171913 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:26 crc kubenswrapper[4775]: E1002 01:43:26.172246 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.672231338 +0000 UTC m=+143.838975378 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.173181 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-758r2" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.187727 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7526f918-73a6-4470-9d8a-b79b418a8f36-bound-sa-token\") pod \"ingress-operator-5b745b69d9-ph4s2\" (UID: \"7526f918-73a6-4470-9d8a-b79b418a8f36\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.191188 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-v9npv" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.221472 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fz5g\" (UniqueName: \"kubernetes.io/projected/71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8-kube-api-access-7fz5g\") pod \"etcd-operator-b45778765-dzqlc\" (UID: \"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.269930 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tcqs\" (UniqueName: \"kubernetes.io/projected/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-kube-api-access-6tcqs\") pod \"collect-profiles-29322810-q7gnj\" (UID: \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.273223 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:26 crc kubenswrapper[4775]: E1002 01:43:26.273572 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.773560276 +0000 UTC m=+143.940304316 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.277961 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4lcd\" (UniqueName: \"kubernetes.io/projected/ce1437a0-173e-4e15-bc0f-ca00b67e9b2e-kube-api-access-l4lcd\") pod \"packageserver-d55dfcdfc-pm5r9\" (UID: \"ce1437a0-173e-4e15-bc0f-ca00b67e9b2e\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.307686 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djzsp\" (UniqueName: \"kubernetes.io/projected/1159e0d6-a6e9-40f9-8f4e-607ea07bb47e-kube-api-access-djzsp\") pod \"service-ca-9c57cc56f-ptw97\" (UID: \"1159e0d6-a6e9-40f9-8f4e-607ea07bb47e\") " pod="openshift-service-ca/service-ca-9c57cc56f-ptw97" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.308998 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4xck\" (UniqueName: \"kubernetes.io/projected/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-kube-api-access-g4xck\") pod \"console-f9d7485db-mgrh5\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.321156 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.326397 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.335601 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6hzq\" (UniqueName: \"kubernetes.io/projected/2bc2207d-9081-475d-be54-85e725ea3819-kube-api-access-t6hzq\") pod \"csi-hostpathplugin-5bnff\" (UID: \"2bc2207d-9081-475d-be54-85e725ea3819\") " pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.361135 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-99jjd" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.375739 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-ptw97" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.376511 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:26 crc kubenswrapper[4775]: E1002 01:43:26.376637 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.876619953 +0000 UTC m=+144.043363993 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.376679 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:26 crc kubenswrapper[4775]: E1002 01:43:26.376989 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.876982203 +0000 UTC m=+144.043726243 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.382442 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-klcvg" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.406003 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.415342 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.427293 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.438122 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.449846 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.476623 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.479019 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:26 crc kubenswrapper[4775]: E1002 01:43:26.479363 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:26.97934069 +0000 UTC m=+144.146084730 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.506028 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-bxp9r"] Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.519496 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k"] Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.519690 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-5bnff" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.582680 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cn7bz" event={"ID":"67aa1649-ec3b-4f32-96fa-d9aece180709","Type":"ContainerStarted","Data":"63f75358a35322ee0e978c8a899c7793f537bb490386140f7fbbeee805062b81"} Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.582724 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-cn7bz" event={"ID":"67aa1649-ec3b-4f32-96fa-d9aece180709","Type":"ContainerStarted","Data":"a1c3829b55264c153d73bbc0bdd3828844fd2ea9a4530b52263f6d55f8d9bf85"} Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.583794 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:26 crc kubenswrapper[4775]: E1002 01:43:26.584118 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:27.084106894 +0000 UTC m=+144.250850934 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.644478 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" event={"ID":"064dc547-a0cb-4008-8a6c-f4eb14ab288f","Type":"ContainerStarted","Data":"274e1d83947495b32d066f2a43f387b8ab959147ccbee348943b2f595cab452c"} Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.644540 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" event={"ID":"064dc547-a0cb-4008-8a6c-f4eb14ab288f","Type":"ContainerStarted","Data":"559bc0aff12f9ec164935ad3bd1f8952f9a69f3152ac3707dadc412e649c623e"} Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.661755 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-v9npv" event={"ID":"7f011434-93f3-4b0c-9ff7-cfbc3f333b83","Type":"ContainerStarted","Data":"a8fe9c7f283fd18970c4e427c6591fe8fbe5213bfa809a08ea97620a5be89060"} Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.685484 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:26 crc kubenswrapper[4775]: E1002 01:43:26.685704 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:27.185674239 +0000 UTC m=+144.352418289 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.685901 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:26 crc kubenswrapper[4775]: E1002 01:43:26.686215 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:27.186203084 +0000 UTC m=+144.352947124 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.690878 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" event={"ID":"59534b10-69de-4d5d-ba59-27c4f24df1a7","Type":"ContainerStarted","Data":"435a9e73d95301094488e582ef01e5ec8140bebf186f910b41d19fa617865417"} Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.691863 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.695192 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" event={"ID":"3e6497ca-0523-4658-8cbd-2340ed601e77","Type":"ContainerStarted","Data":"e8e4ef8485f00b0fe3e9d34f3acdf5738e0cbfe250c212cfa53e36d31e54d9a4"} Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.695352 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.704729 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-47dws" event={"ID":"4b13e137-25af-400d-a3c6-496fa5138392","Type":"ContainerStarted","Data":"2a2289a3117fa7888caac3fe0368e166657cb69a1400a4d0788b411c2692d6ef"} Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.711120 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.788185 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:26 crc kubenswrapper[4775]: E1002 01:43:26.789293 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:27.289277011 +0000 UTC m=+144.456021051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.893911 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:26 crc kubenswrapper[4775]: E1002 01:43:26.894257 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:27.394241621 +0000 UTC m=+144.560985661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.955086 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.995861 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:26 crc kubenswrapper[4775]: E1002 01:43:26.996038 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:27.496015321 +0000 UTC m=+144.662759361 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:26 crc kubenswrapper[4775]: I1002 01:43:26.996118 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:26 crc kubenswrapper[4775]: E1002 01:43:26.996443 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:27.496435843 +0000 UTC m=+144.663179883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.097112 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:27 crc kubenswrapper[4775]: E1002 01:43:27.097293 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:27.597272487 +0000 UTC m=+144.764016527 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.097420 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:27 crc kubenswrapper[4775]: E1002 01:43:27.097706 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:27.597698509 +0000 UTC m=+144.764442549 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.159707 4775 patch_prober.go:28] interesting pod/router-default-5444994796-cn7bz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:43:27 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Oct 02 01:43:27 crc kubenswrapper[4775]: [+]process-running ok Oct 02 01:43:27 crc kubenswrapper[4775]: healthz check failed Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.159758 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7bz" podUID="67aa1649-ec3b-4f32-96fa-d9aece180709" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.199124 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:27 crc kubenswrapper[4775]: E1002 01:43:27.199655 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:27.699629814 +0000 UTC m=+144.866373854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.283377 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-fgvlq"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.291427 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-cn7bz" podStartSLOduration=122.291411132 podStartE2EDuration="2m2.291411132s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:27.290306751 +0000 UTC m=+144.457050791" watchObservedRunningTime="2025-10-02 01:43:27.291411132 +0000 UTC m=+144.458155172" Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.305240 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:27 crc kubenswrapper[4775]: E1002 01:43:27.305584 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:27.805572042 +0000 UTC m=+144.972316072 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:27 crc kubenswrapper[4775]: W1002 01:43:27.305879 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod395ae874_ba30_40f2_957c_c49098b29a74.slice/crio-3cb5b4fd9d574e82d0eb48c3ddecb44d33d309ad4d7fb2bed7169000fc031e55 WatchSource:0}: Error finding container 3cb5b4fd9d574e82d0eb48c3ddecb44d33d309ad4d7fb2bed7169000fc031e55: Status 404 returned error can't find the container with id 3cb5b4fd9d574e82d0eb48c3ddecb44d33d309ad4d7fb2bed7169000fc031e55 Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.403629 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.406764 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:27 crc kubenswrapper[4775]: E1002 01:43:27.407107 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:27.907092395 +0000 UTC m=+145.073836435 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.436598 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.459156 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-57899" podStartSLOduration=123.459142723 podStartE2EDuration="2m3.459142723s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:27.429732784 +0000 UTC m=+144.596476834" watchObservedRunningTime="2025-10-02 01:43:27.459142723 +0000 UTC m=+144.625886763" Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.459311 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-5v4h8" podStartSLOduration=122.459307088 podStartE2EDuration="2m2.459307088s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:27.457136196 +0000 UTC m=+144.623880236" watchObservedRunningTime="2025-10-02 01:43:27.459307088 +0000 UTC m=+144.626051128" Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.509903 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:27 crc kubenswrapper[4775]: E1002 01:43:27.510332 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:28.010316616 +0000 UTC m=+145.177060656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.615522 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:27 crc kubenswrapper[4775]: E1002 01:43:27.615967 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:28.115935985 +0000 UTC m=+145.282680025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.667106 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" podStartSLOduration=123.667086578 podStartE2EDuration="2m3.667086578s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:27.628839469 +0000 UTC m=+144.795583509" watchObservedRunningTime="2025-10-02 01:43:27.667086578 +0000 UTC m=+144.833830618" Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.707704 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fgvlq" event={"ID":"395ae874-ba30-40f2-957c-c49098b29a74","Type":"ContainerStarted","Data":"3cb5b4fd9d574e82d0eb48c3ddecb44d33d309ad4d7fb2bed7169000fc031e55"} Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.719240 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:27 crc kubenswrapper[4775]: E1002 01:43:27.719669 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:28.21965668 +0000 UTC m=+145.386400720 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.722730 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-v9npv" event={"ID":"7f011434-93f3-4b0c-9ff7-cfbc3f333b83","Type":"ContainerStarted","Data":"b046727917daaa2e3625913232df3483db393f2dffa5535d1a105ba9a1a1f2af"} Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.724476 4775 patch_prober.go:28] interesting pod/router-default-5444994796-cn7bz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:43:27 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Oct 02 01:43:27 crc kubenswrapper[4775]: [+]process-running ok Oct 02 01:43:27 crc kubenswrapper[4775]: healthz check failed Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.724508 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7bz" podUID="67aa1649-ec3b-4f32-96fa-d9aece180709" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.725390 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" event={"ID":"64a00282-301f-4947-9795-990b7b0ef4fd","Type":"ContainerStarted","Data":"8561e41a6e33f49a7781aaf96557d454c1f25a2637a53d3fed9e692aa18782e4"} Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.726910 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k" event={"ID":"50774a36-1b88-4b62-98ee-2af94e5f217e","Type":"ContainerStarted","Data":"1a806e95ffa8591c8585cb69d50f0e1c4c4a6b2d1f1ddabf882519923cc124e4"} Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.740779 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" event={"ID":"aec43d0e-fdc8-453c-bb71-0445e899102d","Type":"ContainerStarted","Data":"78b8d095cb7f28cf86dee9dc3d197969aa73ba115b4c0c4f254b9a7254da42af"} Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.759410 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-gvb66"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.790563 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d8qrs"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.820313 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:27 crc kubenswrapper[4775]: E1002 01:43:27.821567 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:28.321550304 +0000 UTC m=+145.488294344 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.822388 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:27 crc kubenswrapper[4775]: E1002 01:43:27.823847 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:28.323833038 +0000 UTC m=+145.490577078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.830067 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dzqlc"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.845411 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.848762 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.854489 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-99jjd"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.858788 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-2l4wc"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.863405 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.865463 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.868596 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.869743 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" podStartSLOduration=123.868606401 podStartE2EDuration="2m3.868606401s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:27.858748853 +0000 UTC m=+145.025492893" watchObservedRunningTime="2025-10-02 01:43:27.868606401 +0000 UTC m=+145.035350441" Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.873528 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.888503 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-758r2"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.915432 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-ptw97"] Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.923725 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:27 crc kubenswrapper[4775]: E1002 01:43:27.924266 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:28.424101096 +0000 UTC m=+145.590845136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:27 crc kubenswrapper[4775]: I1002 01:43:27.929643 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw"] Oct 02 01:43:27 crc kubenswrapper[4775]: W1002 01:43:27.958265 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb99c4b20_1a6d_426f_b9b3_ead1b1cc4993.slice/crio-649d8af341d5c7d7b343e6541987ea0cb4afc8ee51f3dc770f514cfa13ace054 WatchSource:0}: Error finding container 649d8af341d5c7d7b343e6541987ea0cb4afc8ee51f3dc770f514cfa13ace054: Status 404 returned error can't find the container with id 649d8af341d5c7d7b343e6541987ea0cb4afc8ee51f3dc770f514cfa13ace054 Oct 02 01:43:27 crc kubenswrapper[4775]: W1002 01:43:27.961431 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbe09708_d737_4915_bafb_40d9b6d7897d.slice/crio-8fa7add70f8fac5ad2f05dea153dbda51dd71f6dd1c6d176cf6e0c3777a36c68 WatchSource:0}: Error finding container 8fa7add70f8fac5ad2f05dea153dbda51dd71f6dd1c6d176cf6e0c3777a36c68: Status 404 returned error can't find the container with id 8fa7add70f8fac5ad2f05dea153dbda51dd71f6dd1c6d176cf6e0c3777a36c68 Oct 02 01:43:27 crc kubenswrapper[4775]: W1002 01:43:27.985433 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb746ad9f_7c63_4945_adce_aa2e0311ba30.slice/crio-3980ebabae2123f7cf0036e5ef6635edf5377a716ebaa09ff829c7f2d1128938 WatchSource:0}: Error finding container 3980ebabae2123f7cf0036e5ef6635edf5377a716ebaa09ff829c7f2d1128938: Status 404 returned error can't find the container with id 3980ebabae2123f7cf0036e5ef6635edf5377a716ebaa09ff829c7f2d1128938 Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.025863 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:28 crc kubenswrapper[4775]: E1002 01:43:28.026343 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:28.526326819 +0000 UTC m=+145.693070849 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.033549 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" podStartSLOduration=124.033529272 podStartE2EDuration="2m4.033529272s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:28.021301118 +0000 UTC m=+145.188045158" watchObservedRunningTime="2025-10-02 01:43:28.033529272 +0000 UTC m=+145.200273312" Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.034910 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9"] Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.061644 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bd8s9" podStartSLOduration=124.061625635 podStartE2EDuration="2m4.061625635s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:28.061312146 +0000 UTC m=+145.228056196" watchObservedRunningTime="2025-10-02 01:43:28.061625635 +0000 UTC m=+145.228369675" Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.129932 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:28 crc kubenswrapper[4775]: E1002 01:43:28.130264 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:28.63024191 +0000 UTC m=+145.796985950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.130577 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:28 crc kubenswrapper[4775]: E1002 01:43:28.130938 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:28.630925849 +0000 UTC m=+145.797669889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.151188 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj"] Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.175020 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-v987t"] Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.187903 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-klcvg"] Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.194754 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw"] Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.198496 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wnfts"] Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.201123 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4"] Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.203300 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-t86lg"] Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.211056 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5bnff"] Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.231469 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:28 crc kubenswrapper[4775]: E1002 01:43:28.232080 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:28.732057131 +0000 UTC m=+145.898801171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.235659 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2"] Oct 02 01:43:28 crc kubenswrapper[4775]: W1002 01:43:28.248204 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad595bf8_fed5_4d02_8e6b_fbc3742e0ec2.slice/crio-ac8560c35f1c77b83c51229fd2dae7906f99a92d339a3a364fb0f6ab6e8f24d1 WatchSource:0}: Error finding container ac8560c35f1c77b83c51229fd2dae7906f99a92d339a3a364fb0f6ab6e8f24d1: Status 404 returned error can't find the container with id ac8560c35f1c77b83c51229fd2dae7906f99a92d339a3a364fb0f6ab6e8f24d1 Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.272217 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s"] Oct 02 01:43:28 crc kubenswrapper[4775]: W1002 01:43:28.285236 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7526f918_73a6_4470_9d8a_b79b418a8f36.slice/crio-98f3d5ca805cc8b3309539b9b0b9ae7835a752961bb05b47cc9391b28ecb6a2a WatchSource:0}: Error finding container 98f3d5ca805cc8b3309539b9b0b9ae7835a752961bb05b47cc9391b28ecb6a2a: Status 404 returned error can't find the container with id 98f3d5ca805cc8b3309539b9b0b9ae7835a752961bb05b47cc9391b28ecb6a2a Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.297910 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mgrh5"] Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.335928 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:28 crc kubenswrapper[4775]: E1002 01:43:28.336376 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:28.836345373 +0000 UTC m=+146.003089413 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.356725 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5"] Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.437076 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:28 crc kubenswrapper[4775]: E1002 01:43:28.437224 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:28.937202617 +0000 UTC m=+146.103946657 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.437869 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:28 crc kubenswrapper[4775]: E1002 01:43:28.438237 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:28.938223276 +0000 UTC m=+146.104967316 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.540806 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:28 crc kubenswrapper[4775]: E1002 01:43:28.541140 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:29.041123748 +0000 UTC m=+146.207867788 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.643198 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:28 crc kubenswrapper[4775]: E1002 01:43:28.644058 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:29.144040991 +0000 UTC m=+146.310785031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.719125 4775 patch_prober.go:28] interesting pod/router-default-5444994796-cn7bz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:43:28 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Oct 02 01:43:28 crc kubenswrapper[4775]: [+]process-running ok Oct 02 01:43:28 crc kubenswrapper[4775]: healthz check failed Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.719216 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7bz" podUID="67aa1649-ec3b-4f32-96fa-d9aece180709" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.749189 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:28 crc kubenswrapper[4775]: E1002 01:43:28.749628 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:29.249600317 +0000 UTC m=+146.416344357 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.749790 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:28 crc kubenswrapper[4775]: E1002 01:43:28.751365 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:29.251356136 +0000 UTC m=+146.418100176 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.793127 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-t86lg" event={"ID":"a6754f46-92c3-422e-95d9-638efe436d8e","Type":"ContainerStarted","Data":"a94ba7d25ecad20b2c9859c58ddf0c650752aa6e56d709796a4acf7c6d3ef20e"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.797038 4775 generic.go:334] "Generic (PLEG): container finished" podID="64a00282-301f-4947-9795-990b7b0ef4fd" containerID="c75f5cfb99bdb404f10e3ca6bc969820b382c89ce164c0d8c39bf5dfbaff7b9a" exitCode=0 Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.797108 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" event={"ID":"64a00282-301f-4947-9795-990b7b0ef4fd","Type":"ContainerDied","Data":"c75f5cfb99bdb404f10e3ca6bc969820b382c89ce164c0d8c39bf5dfbaff7b9a"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.840845 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" event={"ID":"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8","Type":"ContainerStarted","Data":"879b20c5d3bab9cc88801b395d29b669a9bb55d5dd95d76f42a033da9d289e39"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.840911 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" event={"ID":"71b696d1-0e63-48e1-a5d2-f4f4ef64c9f8","Type":"ContainerStarted","Data":"c00e0e331d4e2f47f10e6a75c8b13d8c0a5715f3315e51473222227a39a49e2a"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.852923 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:28 crc kubenswrapper[4775]: E1002 01:43:28.853529 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:29.353494247 +0000 UTC m=+146.520238287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.879845 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" event={"ID":"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb","Type":"ContainerStarted","Data":"82ade8d77224b7a575d677f1864cd332d4656b9d6dab25140a54078a5faf47ce"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.883238 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-ptw97" event={"ID":"1159e0d6-a6e9-40f9-8f4e-607ea07bb47e","Type":"ContainerStarted","Data":"dc60d7ddb2084b45c071bc1985907b4ab1b8d7770a547e4b36b23ac9f31d4799"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.883274 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-ptw97" event={"ID":"1159e0d6-a6e9-40f9-8f4e-607ea07bb47e","Type":"ContainerStarted","Data":"a9a11fc7b96a172726b5c8017b742780b0ca269b0cd3e9dc1543dc769d540868"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.884979 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-758r2" event={"ID":"1d1cb0fe-95a4-4e16-b5f5-c3847c712081","Type":"ContainerStarted","Data":"3f13a59c71856953e81d6e53d2aea186817212c2cdb9f1602c5f5060c8316249"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.885004 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-758r2" event={"ID":"1d1cb0fe-95a4-4e16-b5f5-c3847c712081","Type":"ContainerStarted","Data":"01215a6071b6aea1912eaf5330b9ca663bcd5afdcdd6c7da058479828188eb70"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.891225 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx" event={"ID":"127b156f-fe4c-40da-a576-e5b4c2ecbfdf","Type":"ContainerStarted","Data":"b117314e0bf2e33705618f6e8613badcff0a60e60c6cbe0762854f20b6e0d1b9"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.907887 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml" event={"ID":"7962f6fc-2505-49ff-b14a-a90eed3e7ba5","Type":"ContainerStarted","Data":"4e30852eb026ed84ba0e17357078ea425c7d116569735f1305bc2de53dc6a5f6"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.921543 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2l4wc" event={"ID":"b99c4b20-1a6d-426f-b9b3-ead1b1cc4993","Type":"ContainerStarted","Data":"f4b3d614ceee48445b8b69c2dee260e744e8cda25b01bd2637e605a003123968"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.921588 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2l4wc" event={"ID":"b99c4b20-1a6d-426f-b9b3-ead1b1cc4993","Type":"ContainerStarted","Data":"649d8af341d5c7d7b343e6541987ea0cb4afc8ee51f3dc770f514cfa13ace054"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.928375 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fgvlq" event={"ID":"395ae874-ba30-40f2-957c-c49098b29a74","Type":"ContainerStarted","Data":"61a015d4527056f96fd8b19f61d222eb593997e31959799a7788c828d990960b"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.928588 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-fgvlq" Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.929861 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5bnff" event={"ID":"2bc2207d-9081-475d-be54-85e725ea3819","Type":"ContainerStarted","Data":"3722f8f5c4f8f5bd36b44e594a0e82a42e24822b6b8c389dfb1ba47c7b810eb4"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.931290 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" event={"ID":"ff918ab3-33fd-44f2-b3ba-9f419f027b10","Type":"ContainerStarted","Data":"6d7cbaecf801b3a86778edd602b8dfdf93c36b835ca3170dedb38e51de4db4b0"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.932899 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" event={"ID":"dbe09708-d737-4915-bafb-40d9b6d7897d","Type":"ContainerStarted","Data":"3eda8530d89117647e6ac2ffe1c04c85a504f32887721b52b4b181a50a435a88"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.932925 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" event={"ID":"dbe09708-d737-4915-bafb-40d9b6d7897d","Type":"ContainerStarted","Data":"8fa7add70f8fac5ad2f05dea153dbda51dd71f6dd1c6d176cf6e0c3777a36c68"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.934846 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k" event={"ID":"50774a36-1b88-4b62-98ee-2af94e5f217e","Type":"ContainerStarted","Data":"69a83821e9a65c17a462956d2d841034207bb8fd46b846b12a9be757c9df979e"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.938622 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mgrh5" event={"ID":"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8","Type":"ContainerStarted","Data":"44074d7cde6ac69e469c678e6d3e861ab6256f1461bbb9b6ed65bb1447519f8e"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.954440 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:28 crc kubenswrapper[4775]: E1002 01:43:28.955659 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:29.455642088 +0000 UTC m=+146.622386118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.959767 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" event={"ID":"aec43d0e-fdc8-453c-bb71-0445e899102d","Type":"ContainerStarted","Data":"22dec874ad9f0bb6ba81ffde9d8a54c41edcb2eefa76e3a566d9edc356918b64"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.963757 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5" event={"ID":"91116a6b-4da7-40db-9633-40a4fb1dcd7e","Type":"ContainerStarted","Data":"24450e547136ac92cc10a0d9394d94f7c9a68d5ce11eeef34911b6ab77aa1c35"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.966851 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-klcvg" event={"ID":"ad595bf8-fed5-4d02-8e6b-fbc3742e0ec2","Type":"ContainerStarted","Data":"ac8560c35f1c77b83c51229fd2dae7906f99a92d339a3a364fb0f6ab6e8f24d1"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.978117 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" event={"ID":"7526f918-73a6-4470-9d8a-b79b418a8f36","Type":"ContainerStarted","Data":"98f3d5ca805cc8b3309539b9b0b9ae7835a752961bb05b47cc9391b28ecb6a2a"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.987535 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" event={"ID":"8354382a-de78-4880-b344-358d49362ea8","Type":"ContainerStarted","Data":"98d1d322d9269ff8fdccdfdbd3fbcd70bc42e18cbf070e645648eddd1042d1b9"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.987583 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" event={"ID":"8354382a-de78-4880-b344-358d49362ea8","Type":"ContainerStarted","Data":"7bd35347bd9a6a730a9186eea49043471e3d95201e6729c4363f9cf391ec4327"} Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.987813 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.989655 4775 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-gzh7f container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.989703 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" podUID="8354382a-de78-4880-b344-358d49362ea8" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/healthz\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 02 01:43:28 crc kubenswrapper[4775]: I1002 01:43:28.991629 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" event={"ID":"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1","Type":"ContainerStarted","Data":"6c67550101e23727fe3d0621df9b999f7d09ded32b04617ccfac182f70f1b5ed"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.012734 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-47dws" event={"ID":"4b13e137-25af-400d-a3c6-496fa5138392","Type":"ContainerStarted","Data":"4ee2785a59a6f27ab14ea66ca70ca20c9c2dc89cd373ed333e9d0f56c408b4d3"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.020039 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" event={"ID":"2e1b3210-4726-450d-8da5-766903e6d91b","Type":"ContainerStarted","Data":"c335259f9d23a5a1f5199708b3dddfb4a2e5fcaa12f0bf36c4c41679845b2865"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.020075 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" event={"ID":"2e1b3210-4726-450d-8da5-766903e6d91b","Type":"ContainerStarted","Data":"d47590691bd098c633778e9ae8a10ac266792bfb91e92b59da1881531200e7f6"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.020846 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.030864 4775 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-d8qrs container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.030906 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" podUID="2e1b3210-4726-450d-8da5-766903e6d91b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.038034 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-gvb66" event={"ID":"63dc8fd2-1f20-4944-a8f1-783700e192f0","Type":"ContainerStarted","Data":"574d7292d974f2685bded473f5ce0393fbac587345eee3d436b9cec0e6107e95"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.038073 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-gvb66" event={"ID":"63dc8fd2-1f20-4944-a8f1-783700e192f0","Type":"ContainerStarted","Data":"aa34f1c4cf74ebf57bfcf1dd4e2b4635bb73b71b1bc12b5b25bacdb731026860"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.038761 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-gvb66" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.042585 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-gvb66 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.042616 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gvb66" podUID="63dc8fd2-1f20-4944-a8f1-783700e192f0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.055139 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:29 crc kubenswrapper[4775]: E1002 01:43:29.057249 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:29.557216292 +0000 UTC m=+146.723960332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.061479 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" event={"ID":"ce1437a0-173e-4e15-bc0f-ca00b67e9b2e","Type":"ContainerStarted","Data":"3513e99d473230eed3bdf7d8e6a31a1768e466655a992b032c753e0b10083589"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.062481 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.065429 4775 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pm5r9 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" start-of-body= Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.065464 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" podUID="ce1437a0-173e-4e15-bc0f-ca00b67e9b2e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": dial tcp 10.217.0.34:5443: connect: connection refused" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.066262 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" event={"ID":"b746ad9f-7c63-4945-adce-aa2e0311ba30","Type":"ContainerStarted","Data":"3980ebabae2123f7cf0036e5ef6635edf5377a716ebaa09ff829c7f2d1128938"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.077392 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.078819 4775 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-whx2n container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.078904 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" podUID="b746ad9f-7c63-4945-adce-aa2e0311ba30" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.135023 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-v987t" event={"ID":"1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4","Type":"ContainerStarted","Data":"0fe7ed352e78aa85f6bf44733c9d381db7adfa86a7a54367d85e775034ff1a19"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.135380 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.147321 4775 patch_prober.go:28] interesting pod/console-operator-58897d9998-v987t container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.147363 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-v987t" podUID="1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.156718 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:29 crc kubenswrapper[4775]: E1002 01:43:29.159170 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:29.659153958 +0000 UTC m=+146.825897998 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.172855 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4" event={"ID":"d3dffe31-ef37-4ecd-b72a-26d8541a331f","Type":"ContainerStarted","Data":"40279d04418b2d6abf0e3e25e066acd42d9611ba668f9414a5cde163e63a9cd1"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.232792 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-99jjd" event={"ID":"971eaf50-b0a8-4339-a2d1-6d2b9df9bc27","Type":"ContainerStarted","Data":"651e05a0bfd31dc14ebaf829e8d3341ccd7a3a99bd205aa117a53ac9bb39012d"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.245868 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" event={"ID":"064dc547-a0cb-4008-8a6c-f4eb14ab288f","Type":"ContainerStarted","Data":"58f88c9d582f22d04842370cfb3313ac3822a89deaab5ce3231bbc71a6e35087"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.248068 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw" event={"ID":"36f09b4c-9db6-4dd3-809e-bf096ac656b2","Type":"ContainerStarted","Data":"633d5f249ef62403e9507487ff25e532f361a93806e36b7740cfd6abfaba414e"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.249440 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" event={"ID":"7c0b4bf0-5b95-4ce8-946c-47e89af5d655","Type":"ContainerStarted","Data":"db9bb606566ca757903e4f03ff9e196e6fe9be4e0a888b4989aa814cc2a18552"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.250077 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.254097 4775 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-tgx4s container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.254219 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" podUID="7c0b4bf0-5b95-4ce8-946c-47e89af5d655" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.259175 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:29 crc kubenswrapper[4775]: E1002 01:43:29.260208 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:29.760192117 +0000 UTC m=+146.926936157 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.262033 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-gvb66" podStartSLOduration=125.262013678 podStartE2EDuration="2m5.262013678s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.261397551 +0000 UTC m=+146.428141591" watchObservedRunningTime="2025-10-02 01:43:29.262013678 +0000 UTC m=+146.428757718" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.263760 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wnfts" event={"ID":"115575fa-2b4f-4b7e-bdd8-46c85da3855d","Type":"ContainerStarted","Data":"2a436491b63237cfdf96a9ca11bf6d370d8eebcc572d6ed87835bac01e435daf"} Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.302918 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" podStartSLOduration=124.302901332 podStartE2EDuration="2m4.302901332s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.302208312 +0000 UTC m=+146.468952352" watchObservedRunningTime="2025-10-02 01:43:29.302901332 +0000 UTC m=+146.469645372" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.360901 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:29 crc kubenswrapper[4775]: E1002 01:43:29.363561 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:29.863543402 +0000 UTC m=+147.030287442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.424176 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-w4snq" podStartSLOduration=125.424150181 podStartE2EDuration="2m5.424150181s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.390587055 +0000 UTC m=+146.557331095" watchObservedRunningTime="2025-10-02 01:43:29.424150181 +0000 UTC m=+146.590894221" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.462397 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:29 crc kubenswrapper[4775]: E1002 01:43:29.462758 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:29.96274332 +0000 UTC m=+147.129487360 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.467649 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-dzqlc" podStartSLOduration=125.467623617 podStartE2EDuration="2m5.467623617s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.426301972 +0000 UTC m=+146.593046012" watchObservedRunningTime="2025-10-02 01:43:29.467623617 +0000 UTC m=+146.634367657" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.510983 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-bxp9r" podStartSLOduration=125.51096664 podStartE2EDuration="2m5.51096664s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.470266382 +0000 UTC m=+146.637010422" watchObservedRunningTime="2025-10-02 01:43:29.51096664 +0000 UTC m=+146.677710680" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.511629 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-47dws" podStartSLOduration=125.511624118 podStartE2EDuration="2m5.511624118s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.509324053 +0000 UTC m=+146.676068093" watchObservedRunningTime="2025-10-02 01:43:29.511624118 +0000 UTC m=+146.678368158" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.547398 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-fgvlq" podStartSLOduration=6.547380887 podStartE2EDuration="6.547380887s" podCreationTimestamp="2025-10-02 01:43:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.546274075 +0000 UTC m=+146.713018115" watchObservedRunningTime="2025-10-02 01:43:29.547380887 +0000 UTC m=+146.714124927" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.564345 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:29 crc kubenswrapper[4775]: E1002 01:43:29.564705 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:30.064689415 +0000 UTC m=+147.231433455 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.590334 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-brt8k" podStartSLOduration=125.590319778 podStartE2EDuration="2m5.590319778s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.588491526 +0000 UTC m=+146.755235556" watchObservedRunningTime="2025-10-02 01:43:29.590319778 +0000 UTC m=+146.757063818" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.624917 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" podStartSLOduration=124.624901913 podStartE2EDuration="2m4.624901913s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.623485953 +0000 UTC m=+146.790229993" watchObservedRunningTime="2025-10-02 01:43:29.624901913 +0000 UTC m=+146.791645953" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.667304 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:29 crc kubenswrapper[4775]: E1002 01:43:29.668118 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:30.168093601 +0000 UTC m=+147.334837641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.668604 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" podStartSLOduration=124.668574065 podStartE2EDuration="2m4.668574065s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.666792364 +0000 UTC m=+146.833536394" watchObservedRunningTime="2025-10-02 01:43:29.668574065 +0000 UTC m=+146.835318105" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.707880 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-v987t" podStartSLOduration=125.707862813 podStartE2EDuration="2m5.707862813s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.706506455 +0000 UTC m=+146.873250495" watchObservedRunningTime="2025-10-02 01:43:29.707862813 +0000 UTC m=+146.874606853" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.719319 4775 patch_prober.go:28] interesting pod/router-default-5444994796-cn7bz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:43:29 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Oct 02 01:43:29 crc kubenswrapper[4775]: [+]process-running ok Oct 02 01:43:29 crc kubenswrapper[4775]: healthz check failed Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.719387 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7bz" podUID="67aa1649-ec3b-4f32-96fa-d9aece180709" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.753710 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-ptw97" podStartSLOduration=124.753696175 podStartE2EDuration="2m4.753696175s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.7531598 +0000 UTC m=+146.919903840" watchObservedRunningTime="2025-10-02 01:43:29.753696175 +0000 UTC m=+146.920440205" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.770172 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:29 crc kubenswrapper[4775]: E1002 01:43:29.770494 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:30.270482699 +0000 UTC m=+147.437226739 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.786860 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" podStartSLOduration=124.78684582 podStartE2EDuration="2m4.78684582s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.784208036 +0000 UTC m=+146.950952076" watchObservedRunningTime="2025-10-02 01:43:29.78684582 +0000 UTC m=+146.953589860" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.829456 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-758r2" podStartSLOduration=6.8294415520000005 podStartE2EDuration="6.829441552s" podCreationTimestamp="2025-10-02 01:43:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.828202537 +0000 UTC m=+146.994946577" watchObservedRunningTime="2025-10-02 01:43:29.829441552 +0000 UTC m=+146.996185592" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.871406 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:29 crc kubenswrapper[4775]: E1002 01:43:29.871713 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:30.371698333 +0000 UTC m=+147.538442363 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.920649 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-99jjd" podStartSLOduration=124.920634314 podStartE2EDuration="2m4.920634314s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.878906737 +0000 UTC m=+147.045650777" watchObservedRunningTime="2025-10-02 01:43:29.920634314 +0000 UTC m=+147.087378354" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.921704 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.922078 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.944074 4775 patch_prober.go:28] interesting pod/apiserver-76f77b778f-47dws container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 02 01:43:29 crc kubenswrapper[4775]: [+]log ok Oct 02 01:43:29 crc kubenswrapper[4775]: [+]etcd ok Oct 02 01:43:29 crc kubenswrapper[4775]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 02 01:43:29 crc kubenswrapper[4775]: [+]poststarthook/generic-apiserver-start-informers ok Oct 02 01:43:29 crc kubenswrapper[4775]: [+]poststarthook/max-in-flight-filter ok Oct 02 01:43:29 crc kubenswrapper[4775]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 02 01:43:29 crc kubenswrapper[4775]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 02 01:43:29 crc kubenswrapper[4775]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 02 01:43:29 crc kubenswrapper[4775]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 02 01:43:29 crc kubenswrapper[4775]: [+]poststarthook/project.openshift.io-projectcache ok Oct 02 01:43:29 crc kubenswrapper[4775]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 02 01:43:29 crc kubenswrapper[4775]: [+]poststarthook/openshift.io-startinformers ok Oct 02 01:43:29 crc kubenswrapper[4775]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 02 01:43:29 crc kubenswrapper[4775]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 02 01:43:29 crc kubenswrapper[4775]: livez check failed Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.944121 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-47dws" podUID="4b13e137-25af-400d-a3c6-496fa5138392" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.946859 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6g5xq" podStartSLOduration=126.946848033 podStartE2EDuration="2m6.946848033s" podCreationTimestamp="2025-10-02 01:41:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.92157885 +0000 UTC m=+147.088322900" watchObservedRunningTime="2025-10-02 01:43:29.946848033 +0000 UTC m=+147.113592073" Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.975266 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:29 crc kubenswrapper[4775]: E1002 01:43:29.975617 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:30.475604584 +0000 UTC m=+147.642348624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:29 crc kubenswrapper[4775]: I1002 01:43:29.982351 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-v9npv" podStartSLOduration=6.982333854 podStartE2EDuration="6.982333854s" podCreationTimestamp="2025-10-02 01:43:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.947384278 +0000 UTC m=+147.114128328" watchObservedRunningTime="2025-10-02 01:43:29.982333854 +0000 UTC m=+147.149077894" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.076618 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:30 crc kubenswrapper[4775]: E1002 01:43:30.076941 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:30.576926221 +0000 UTC m=+147.743670261 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.178463 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:30 crc kubenswrapper[4775]: E1002 01:43:30.178829 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:30.678812485 +0000 UTC m=+147.845556525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.269928 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-v987t" event={"ID":"1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4","Type":"ContainerStarted","Data":"36748e89d5d187b6de2340b4b88f19a70866f83b43a3db65cd9a691b8726d292"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.270531 4775 patch_prober.go:28] interesting pod/console-operator-58897d9998-v987t container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.270565 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-v987t" podUID="1f2cdcf9-3f4d-48b5-bdea-b5980a87afe4" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/readyz\": dial tcp 10.217.0.16:8443: connect: connection refused" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.272020 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw" event={"ID":"36f09b4c-9db6-4dd3-809e-bf096ac656b2","Type":"ContainerStarted","Data":"0b19ac0b0ec89971ef1a203e52a9279d8be9309db6f1212619c8095cdf41e480"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.272060 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw" event={"ID":"36f09b4c-9db6-4dd3-809e-bf096ac656b2","Type":"ContainerStarted","Data":"1ca7c03d9f8a32b2dc743cc44b7327e15b4405fb6ca92070bbe68e9c14989860"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.272091 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.274644 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mgrh5" event={"ID":"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8","Type":"ContainerStarted","Data":"68d0c1e399050ecaed8b9a92889f547ea07e05b90af67879292d4c0f53ba71a1"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.275989 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-t86lg" event={"ID":"a6754f46-92c3-422e-95d9-638efe436d8e","Type":"ContainerStarted","Data":"6109502b655c0229021a27f38d11a37fa2245a9f1acf211f1d943c6a605cecf4"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.276018 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-t86lg" event={"ID":"a6754f46-92c3-422e-95d9-638efe436d8e","Type":"ContainerStarted","Data":"a4d8e598c86cb3d0359e256dd7c3e9e69b08d027dd24317d90e085cccbd9c442"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.277565 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-99jjd" event={"ID":"971eaf50-b0a8-4339-a2d1-6d2b9df9bc27","Type":"ContainerStarted","Data":"0fef0cf2cbb2d1dd4a664e96a55ee142a136e40583d96f12510057dde4813e8d"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.279050 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:30 crc kubenswrapper[4775]: E1002 01:43:30.279418 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:30.779406072 +0000 UTC m=+147.946150112 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.280282 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-fgvlq" event={"ID":"395ae874-ba30-40f2-957c-c49098b29a74","Type":"ContainerStarted","Data":"7810ac3d7dd35abcf81d6d8509944d38f90bb0b924b3de75502d1f694fa034f3"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.283421 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" event={"ID":"7526f918-73a6-4470-9d8a-b79b418a8f36","Type":"ContainerStarted","Data":"5eeb463bcfe24c8c965c5e724781b96bdd0d97fb7bed3b23d023d2f84a024eb0"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.283462 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" event={"ID":"7526f918-73a6-4470-9d8a-b79b418a8f36","Type":"ContainerStarted","Data":"aa2e6b750a31c73919ba5f5f371ee88d409a7a1d5b93c1a280f28b0f4a3a5a93"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.286064 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" event={"ID":"ce1437a0-173e-4e15-bc0f-ca00b67e9b2e","Type":"ContainerStarted","Data":"1fe40ff33690940275829e113bd6e945687a29d2603a194658b02a4da15e1989"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.288671 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" event={"ID":"7c0b4bf0-5b95-4ce8-946c-47e89af5d655","Type":"ContainerStarted","Data":"88b60b92b03aeb02ea57d94cd8941cb8d1ce9d1bd6c85b835966b1273c2dec58"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.290891 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" event={"ID":"64a00282-301f-4947-9795-990b7b0ef4fd","Type":"ContainerStarted","Data":"cf156d9584a591b3161328b881a3394e0647ec4c76f28d6911dc4e8dd81cabfd"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.292888 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-2l4wc" event={"ID":"b99c4b20-1a6d-426f-b9b3-ead1b1cc4993","Type":"ContainerStarted","Data":"61a11fb6aa9e1dc82df6c0c290cdff9ef33374e38104ab3760db45feb3f395c2"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.294273 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4" event={"ID":"d3dffe31-ef37-4ecd-b72a-26d8541a331f","Type":"ContainerStarted","Data":"178549faa0a79570ad50119ec2da0622c8e850615263c5a84b4c60d617fae14f"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.295227 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.296044 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-9jq6t" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.297539 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wnfts" event={"ID":"115575fa-2b4f-4b7e-bdd8-46c85da3855d","Type":"ContainerStarted","Data":"9b89d5235e1d95a083d1f45170c8b541a76bdc5c4dd79cb37b710c0f2da01b76"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.299000 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5bnff" event={"ID":"2bc2207d-9081-475d-be54-85e725ea3819","Type":"ContainerStarted","Data":"ca64e2a791a8309235889bc886c9e46a82ee66a2eb520161b1d70f07abb48552"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.299909 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" event={"ID":"b746ad9f-7c63-4945-adce-aa2e0311ba30","Type":"ContainerStarted","Data":"273908bae7ad3e2d8c3d077b9b42606c9b7cef5cf0db3ba510318fbe159fc240"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.302743 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" event={"ID":"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb","Type":"ContainerStarted","Data":"efc477524ca01f31df8327a3c03cc87afde4cba918672c3089503b3b56b12556"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.302781 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" event={"ID":"382aa7eb-2fa9-4897-8cf6-e2db3cdfe0bb","Type":"ContainerStarted","Data":"cacc84b593635c7f851c71359b61b3de5e27c52d0cab3d6f3d665a5ba668297a"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.304305 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx" event={"ID":"127b156f-fe4c-40da-a576-e5b4c2ecbfdf","Type":"ContainerStarted","Data":"137b0167c0e7b048e3cfd1551f43fac55a0b0f3f4ac8b1767de6645c03c487b2"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.306814 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml" event={"ID":"7962f6fc-2505-49ff-b14a-a90eed3e7ba5","Type":"ContainerStarted","Data":"62958d3661ce8da5dcb0a87f2a9e6b143211daeb3ae12fcfc9c387f122ebaf0e"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.311343 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5" event={"ID":"91116a6b-4da7-40db-9633-40a4fb1dcd7e","Type":"ContainerStarted","Data":"8aab57b1904ffe456a635f9540875138cb2a41a033a175555c49eec90fc9fa58"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.317844 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-klcvg" event={"ID":"ad595bf8-fed5-4d02-8e6b-fbc3742e0ec2","Type":"ContainerStarted","Data":"4325e5b8dc1030934cfb6ab8f0726c72b37c8adba6302840128bd460abbaa934"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.317892 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-klcvg" event={"ID":"ad595bf8-fed5-4d02-8e6b-fbc3742e0ec2","Type":"ContainerStarted","Data":"fb937c5d54e8b4d556e3dbc65e4d77e4c8442a11adfd2e4e0e22eb1fbde19bc5"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.318473 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-tgx4s" podStartSLOduration=125.318453013 podStartE2EDuration="2m5.318453013s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:29.982093877 +0000 UTC m=+147.148837917" watchObservedRunningTime="2025-10-02 01:43:30.318453013 +0000 UTC m=+147.485197053" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.319613 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" event={"ID":"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1","Type":"ContainerStarted","Data":"38e75b293fb6affc2f95a7680ce1755481bcef32a319098bd8b55afb6cb27312"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.326005 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" event={"ID":"ff918ab3-33fd-44f2-b3ba-9f419f027b10","Type":"ContainerStarted","Data":"59694e5257253da92021df501066be107edbeb706bf7c2289514a2a09d09307e"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.326075 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" event={"ID":"ff918ab3-33fd-44f2-b3ba-9f419f027b10","Type":"ContainerStarted","Data":"3713a22e89864e3c0c24f997e6c5ee31e2ae9d7201c17c9eb0b61cd2675ae453"} Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.326877 4775 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-d8qrs container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" start-of-body= Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.326926 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" podUID="2e1b3210-4726-450d-8da5-766903e6d91b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.23:8080/healthz\": dial tcp 10.217.0.23:8080: connect: connection refused" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.327228 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-gvb66 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.327277 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gvb66" podUID="63dc8fd2-1f20-4944-a8f1-783700e192f0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.353098 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-lgwbw" podStartSLOduration=125.35308411 podStartE2EDuration="2m5.35308411s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:30.352320788 +0000 UTC m=+147.519064838" watchObservedRunningTime="2025-10-02 01:43:30.35308411 +0000 UTC m=+147.519828150" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.353509 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw" podStartSLOduration=125.353503912 podStartE2EDuration="2m5.353503912s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:30.320179162 +0000 UTC m=+147.486923202" watchObservedRunningTime="2025-10-02 01:43:30.353503912 +0000 UTC m=+147.520247952" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.380978 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:30 crc kubenswrapper[4775]: E1002 01:43:30.385452 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:30.885436722 +0000 UTC m=+148.052180762 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.400383 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-gzh7f" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.406691 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-klcvg" podStartSLOduration=125.406675181 podStartE2EDuration="2m5.406675181s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:30.406447235 +0000 UTC m=+147.573191275" watchObservedRunningTime="2025-10-02 01:43:30.406675181 +0000 UTC m=+147.573419221" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.427530 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-2l4wc" podStartSLOduration=125.427514009 podStartE2EDuration="2m5.427514009s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:30.425354828 +0000 UTC m=+147.592098868" watchObservedRunningTime="2025-10-02 01:43:30.427514009 +0000 UTC m=+147.594258049" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.482148 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:30 crc kubenswrapper[4775]: E1002 01:43:30.482357 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:30.982330385 +0000 UTC m=+148.149074425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.482512 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:30 crc kubenswrapper[4775]: E1002 01:43:30.482943 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:30.982927042 +0000 UTC m=+148.149671082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.509716 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-ph4s2" podStartSLOduration=125.509696397 podStartE2EDuration="2m5.509696397s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:30.502939676 +0000 UTC m=+147.669683716" watchObservedRunningTime="2025-10-02 01:43:30.509696397 +0000 UTC m=+147.676440437" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.589208 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:30 crc kubenswrapper[4775]: E1002 01:43:30.589606 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.08958825 +0000 UTC m=+148.256332290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.607784 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-t86lg" podStartSLOduration=126.607765933 podStartE2EDuration="2m6.607765933s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:30.607284999 +0000 UTC m=+147.774029039" watchObservedRunningTime="2025-10-02 01:43:30.607765933 +0000 UTC m=+147.774509973" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.691350 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:30 crc kubenswrapper[4775]: E1002 01:43:30.691940 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.191928456 +0000 UTC m=+148.358672496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.702222 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.714774 4775 patch_prober.go:28] interesting pod/router-default-5444994796-cn7bz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:43:30 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Oct 02 01:43:30 crc kubenswrapper[4775]: [+]process-running ok Oct 02 01:43:30 crc kubenswrapper[4775]: healthz check failed Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.714822 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7bz" podUID="67aa1649-ec3b-4f32-96fa-d9aece180709" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.724854 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" podStartSLOduration=125.724838315 podStartE2EDuration="2m5.724838315s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:30.722878969 +0000 UTC m=+147.889623009" watchObservedRunningTime="2025-10-02 01:43:30.724838315 +0000 UTC m=+147.891582355" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.795446 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:30 crc kubenswrapper[4775]: E1002 01:43:30.795645 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.295629211 +0000 UTC m=+148.462373251 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.795705 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:30 crc kubenswrapper[4775]: E1002 01:43:30.796001 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.295994291 +0000 UTC m=+148.462738331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.798287 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-qglml" podStartSLOduration=125.798268946 podStartE2EDuration="2m5.798268946s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:30.794225711 +0000 UTC m=+147.960969751" watchObservedRunningTime="2025-10-02 01:43:30.798268946 +0000 UTC m=+147.965012986" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.846028 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.846079 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.898864 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:30 crc kubenswrapper[4775]: E1002 01:43:30.899271 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.399255734 +0000 UTC m=+148.565999774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.903084 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-mgrh5" podStartSLOduration=126.903071001 podStartE2EDuration="2m6.903071001s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:30.902109104 +0000 UTC m=+148.068853144" watchObservedRunningTime="2025-10-02 01:43:30.903071001 +0000 UTC m=+148.069815041" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.903986 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-npmlx" podStartSLOduration=125.903980207 podStartE2EDuration="2m5.903980207s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:30.842149913 +0000 UTC m=+148.008893953" watchObservedRunningTime="2025-10-02 01:43:30.903980207 +0000 UTC m=+148.070724257" Oct 02 01:43:30 crc kubenswrapper[4775]: I1002 01:43:30.939772 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-9cng5" podStartSLOduration=125.939757696 podStartE2EDuration="2m5.939757696s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:30.93919901 +0000 UTC m=+148.105943050" watchObservedRunningTime="2025-10-02 01:43:30.939757696 +0000 UTC m=+148.106501736" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.000425 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.000707 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.500696275 +0000 UTC m=+148.667440305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.014498 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-wnfts" podStartSLOduration=126.014481213 podStartE2EDuration="2m6.014481213s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:31.007430295 +0000 UTC m=+148.174174335" watchObservedRunningTime="2025-10-02 01:43:31.014481213 +0000 UTC m=+148.181225253" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.049614 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-plfk4" podStartSLOduration=126.049597524 podStartE2EDuration="2m6.049597524s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:31.047296609 +0000 UTC m=+148.214040649" watchObservedRunningTime="2025-10-02 01:43:31.049597524 +0000 UTC m=+148.216341554" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.080753 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" podStartSLOduration=126.080737342 podStartE2EDuration="2m6.080737342s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:31.079830866 +0000 UTC m=+148.246574916" watchObservedRunningTime="2025-10-02 01:43:31.080737342 +0000 UTC m=+148.247481372" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.102089 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.102177 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.602157346 +0000 UTC m=+148.768901386 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.102334 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.102604 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.602590338 +0000 UTC m=+148.769334378 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.137815 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lg72k" podStartSLOduration=126.137797351 podStartE2EDuration="2m6.137797351s" podCreationTimestamp="2025-10-02 01:41:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:31.101860418 +0000 UTC m=+148.268604458" watchObservedRunningTime="2025-10-02 01:43:31.137797351 +0000 UTC m=+148.304541391" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.203577 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.203972 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.703942207 +0000 UTC m=+148.870686247 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.287367 4775 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pm5r9 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.287441 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" podUID="ce1437a0-173e-4e15-bc0f-ca00b67e9b2e" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.34:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.305122 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.305469 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.805442679 +0000 UTC m=+148.972186719 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.337179 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5bnff" event={"ID":"2bc2207d-9081-475d-be54-85e725ea3819","Type":"ContainerStarted","Data":"ae9e123171c9f3d110d3a9468157b82cf2a401be5fb62dc46604295fe5aca16e"} Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.339475 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-gvb66 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.339523 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gvb66" podUID="63dc8fd2-1f20-4944-a8f1-783700e192f0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.388940 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.406375 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.406552 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.90652567 +0000 UTC m=+149.073269710 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.406641 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.406982 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:31.906975093 +0000 UTC m=+149.073719133 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.507926 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.508129 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:32.008104945 +0000 UTC m=+149.174848985 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.509181 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.510376 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:32.010360469 +0000 UTC m=+149.177104619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.564258 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pm5r9" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.616449 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.616670 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:32.116640826 +0000 UTC m=+149.283384866 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.616744 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.617075 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:32.117062478 +0000 UTC m=+149.283806518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.712226 4775 patch_prober.go:28] interesting pod/router-default-5444994796-cn7bz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:43:31 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Oct 02 01:43:31 crc kubenswrapper[4775]: [+]process-running ok Oct 02 01:43:31 crc kubenswrapper[4775]: healthz check failed Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.712503 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7bz" podUID="67aa1649-ec3b-4f32-96fa-d9aece180709" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.717623 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.717798 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:32.217773448 +0000 UTC m=+149.384517488 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.717835 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.717861 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.717891 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.718182 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:32.21817353 +0000 UTC m=+149.384917570 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.718782 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.718820 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.718931 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.723665 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.724252 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.736468 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.802145 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.820166 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.820519 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:32.320503226 +0000 UTC m=+149.487247266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.892623 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.921643 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:31 crc kubenswrapper[4775]: E1002 01:43:31.922167 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:32.422155033 +0000 UTC m=+149.588899073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.929370 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:31 crc kubenswrapper[4775]: I1002 01:43:31.945070 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.031839 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:32 crc kubenswrapper[4775]: E1002 01:43:32.031979 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:32.531937469 +0000 UTC m=+149.698681509 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.032055 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:32 crc kubenswrapper[4775]: E1002 01:43:32.032340 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:32.53233274 +0000 UTC m=+149.699076780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.032362 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-v987t" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.043322 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zzlw6"] Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.044262 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.046251 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.087263 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zzlw6"] Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.133045 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.133531 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-utilities\") pod \"certified-operators-zzlw6\" (UID: \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\") " pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.133558 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-catalog-content\") pod \"certified-operators-zzlw6\" (UID: \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\") " pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.133585 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2fxg\" (UniqueName: \"kubernetes.io/projected/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-kube-api-access-x2fxg\") pod \"certified-operators-zzlw6\" (UID: \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\") " pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:43:32 crc kubenswrapper[4775]: E1002 01:43:32.133699 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:32.633682568 +0000 UTC m=+149.800426608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.236007 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-utilities\") pod \"certified-operators-zzlw6\" (UID: \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\") " pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.236056 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-catalog-content\") pod \"certified-operators-zzlw6\" (UID: \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\") " pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.236078 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2fxg\" (UniqueName: \"kubernetes.io/projected/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-kube-api-access-x2fxg\") pod \"certified-operators-zzlw6\" (UID: \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\") " pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.236101 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:32 crc kubenswrapper[4775]: E1002 01:43:32.236700 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:32.736688183 +0000 UTC m=+149.903432223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.237472 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-utilities\") pod \"certified-operators-zzlw6\" (UID: \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\") " pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.237585 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zj54d"] Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.237688 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-catalog-content\") pod \"certified-operators-zzlw6\" (UID: \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\") " pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.238460 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.241020 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.249629 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zj54d"] Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.289174 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2fxg\" (UniqueName: \"kubernetes.io/projected/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-kube-api-access-x2fxg\") pod \"certified-operators-zzlw6\" (UID: \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\") " pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.344564 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.344783 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d1808de-113f-4223-99eb-0f72d7a1b8de-catalog-content\") pod \"community-operators-zj54d\" (UID: \"2d1808de-113f-4223-99eb-0f72d7a1b8de\") " pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.344830 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d1808de-113f-4223-99eb-0f72d7a1b8de-utilities\") pod \"community-operators-zj54d\" (UID: \"2d1808de-113f-4223-99eb-0f72d7a1b8de\") " pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.344873 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slg4j\" (UniqueName: \"kubernetes.io/projected/2d1808de-113f-4223-99eb-0f72d7a1b8de-kube-api-access-slg4j\") pod \"community-operators-zj54d\" (UID: \"2d1808de-113f-4223-99eb-0f72d7a1b8de\") " pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:43:32 crc kubenswrapper[4775]: E1002 01:43:32.344979 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:32.844963626 +0000 UTC m=+150.011707666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:32 crc kubenswrapper[4775]: W1002 01:43:32.351479 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-65362c3bc5c4b86aeebeee59b0706c53f1c863ca3362352a8c3fe0f382495816 WatchSource:0}: Error finding container 65362c3bc5c4b86aeebeee59b0706c53f1c863ca3362352a8c3fe0f382495816: Status 404 returned error can't find the container with id 65362c3bc5c4b86aeebeee59b0706c53f1c863ca3362352a8c3fe0f382495816 Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.379732 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5bnff" event={"ID":"2bc2207d-9081-475d-be54-85e725ea3819","Type":"ContainerStarted","Data":"9951e9f23dc586d140b99fba5f60e6bfb185c18d1b2feb571eb2c251d2c70c41"} Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.379778 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5bnff" event={"ID":"2bc2207d-9081-475d-be54-85e725ea3819","Type":"ContainerStarted","Data":"8ac527fe8b627aefc98398247ef21cae2130a2da6eaa57144dd7bfaa9394245a"} Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.410065 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kqzhr" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.415435 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.438046 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-5bnff" podStartSLOduration=9.438024421 podStartE2EDuration="9.438024421s" podCreationTimestamp="2025-10-02 01:43:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:32.421309439 +0000 UTC m=+149.588053479" watchObservedRunningTime="2025-10-02 01:43:32.438024421 +0000 UTC m=+149.604768461" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.440596 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6s8gr"] Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.453822 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d1808de-113f-4223-99eb-0f72d7a1b8de-utilities\") pod \"community-operators-zj54d\" (UID: \"2d1808de-113f-4223-99eb-0f72d7a1b8de\") " pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.453906 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.453940 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slg4j\" (UniqueName: \"kubernetes.io/projected/2d1808de-113f-4223-99eb-0f72d7a1b8de-kube-api-access-slg4j\") pod \"community-operators-zj54d\" (UID: \"2d1808de-113f-4223-99eb-0f72d7a1b8de\") " pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.454059 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d1808de-113f-4223-99eb-0f72d7a1b8de-catalog-content\") pod \"community-operators-zj54d\" (UID: \"2d1808de-113f-4223-99eb-0f72d7a1b8de\") " pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.454802 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d1808de-113f-4223-99eb-0f72d7a1b8de-catalog-content\") pod \"community-operators-zj54d\" (UID: \"2d1808de-113f-4223-99eb-0f72d7a1b8de\") " pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.455067 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d1808de-113f-4223-99eb-0f72d7a1b8de-utilities\") pod \"community-operators-zj54d\" (UID: \"2d1808de-113f-4223-99eb-0f72d7a1b8de\") " pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.457498 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:43:32 crc kubenswrapper[4775]: E1002 01:43:32.458543 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:32.958529319 +0000 UTC m=+150.125273359 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.463879 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6s8gr"] Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.491344 4775 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.492343 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slg4j\" (UniqueName: \"kubernetes.io/projected/2d1808de-113f-4223-99eb-0f72d7a1b8de-kube-api-access-slg4j\") pod \"community-operators-zj54d\" (UID: \"2d1808de-113f-4223-99eb-0f72d7a1b8de\") " pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:43:32 crc kubenswrapper[4775]: W1002 01:43:32.549269 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-7505824cd6030ebecbb4a68974d46f8714ca35b75110e1f1939c9886bfe73cb2 WatchSource:0}: Error finding container 7505824cd6030ebecbb4a68974d46f8714ca35b75110e1f1939c9886bfe73cb2: Status 404 returned error can't find the container with id 7505824cd6030ebecbb4a68974d46f8714ca35b75110e1f1939c9886bfe73cb2 Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.554712 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:32 crc kubenswrapper[4775]: E1002 01:43:32.554820 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:33.054796884 +0000 UTC m=+150.221540924 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.555019 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5884c55d-52c3-4bf0-bab6-d20aaf245f97-catalog-content\") pod \"certified-operators-6s8gr\" (UID: \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\") " pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.555188 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5884c55d-52c3-4bf0-bab6-d20aaf245f97-utilities\") pod \"certified-operators-6s8gr\" (UID: \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\") " pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.555236 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpc67\" (UniqueName: \"kubernetes.io/projected/5884c55d-52c3-4bf0-bab6-d20aaf245f97-kube-api-access-gpc67\") pod \"certified-operators-6s8gr\" (UID: \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\") " pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.555300 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:32 crc kubenswrapper[4775]: E1002 01:43:32.557976 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:33.057947483 +0000 UTC m=+150.224691523 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.595343 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.637606 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v6n5w"] Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.638671 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.644669 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6n5w"] Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.662594 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.662966 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5884c55d-52c3-4bf0-bab6-d20aaf245f97-catalog-content\") pod \"certified-operators-6s8gr\" (UID: \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\") " pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.663035 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5884c55d-52c3-4bf0-bab6-d20aaf245f97-utilities\") pod \"certified-operators-6s8gr\" (UID: \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\") " pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.663076 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpc67\" (UniqueName: \"kubernetes.io/projected/5884c55d-52c3-4bf0-bab6-d20aaf245f97-kube-api-access-gpc67\") pod \"certified-operators-6s8gr\" (UID: \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\") " pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:43:32 crc kubenswrapper[4775]: E1002 01:43:32.663493 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:33.163477319 +0000 UTC m=+150.330221359 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.664543 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5884c55d-52c3-4bf0-bab6-d20aaf245f97-catalog-content\") pod \"certified-operators-6s8gr\" (UID: \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\") " pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.665653 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5884c55d-52c3-4bf0-bab6-d20aaf245f97-utilities\") pod \"certified-operators-6s8gr\" (UID: \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\") " pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.690789 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpc67\" (UniqueName: \"kubernetes.io/projected/5884c55d-52c3-4bf0-bab6-d20aaf245f97-kube-api-access-gpc67\") pod \"certified-operators-6s8gr\" (UID: \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\") " pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.692093 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zzlw6"] Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.715134 4775 patch_prober.go:28] interesting pod/router-default-5444994796-cn7bz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:43:32 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Oct 02 01:43:32 crc kubenswrapper[4775]: [+]process-running ok Oct 02 01:43:32 crc kubenswrapper[4775]: healthz check failed Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.715177 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7bz" podUID="67aa1649-ec3b-4f32-96fa-d9aece180709" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.765275 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ebb5372-c005-4a86-b356-f36a3078feb2-catalog-content\") pod \"community-operators-v6n5w\" (UID: \"1ebb5372-c005-4a86-b356-f36a3078feb2\") " pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.765338 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ebb5372-c005-4a86-b356-f36a3078feb2-utilities\") pod \"community-operators-v6n5w\" (UID: \"1ebb5372-c005-4a86-b356-f36a3078feb2\") " pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.765370 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt2gz\" (UniqueName: \"kubernetes.io/projected/1ebb5372-c005-4a86-b356-f36a3078feb2-kube-api-access-zt2gz\") pod \"community-operators-v6n5w\" (UID: \"1ebb5372-c005-4a86-b356-f36a3078feb2\") " pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.765432 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:32 crc kubenswrapper[4775]: E1002 01:43:32.765720 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-02 01:43:33.265703912 +0000 UTC m=+150.432447952 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nv4tc" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.784071 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.866165 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.866318 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ebb5372-c005-4a86-b356-f36a3078feb2-utilities\") pod \"community-operators-v6n5w\" (UID: \"1ebb5372-c005-4a86-b356-f36a3078feb2\") " pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.866352 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt2gz\" (UniqueName: \"kubernetes.io/projected/1ebb5372-c005-4a86-b356-f36a3078feb2-kube-api-access-zt2gz\") pod \"community-operators-v6n5w\" (UID: \"1ebb5372-c005-4a86-b356-f36a3078feb2\") " pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.866418 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ebb5372-c005-4a86-b356-f36a3078feb2-catalog-content\") pod \"community-operators-v6n5w\" (UID: \"1ebb5372-c005-4a86-b356-f36a3078feb2\") " pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.867151 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ebb5372-c005-4a86-b356-f36a3078feb2-catalog-content\") pod \"community-operators-v6n5w\" (UID: \"1ebb5372-c005-4a86-b356-f36a3078feb2\") " pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:43:32 crc kubenswrapper[4775]: E1002 01:43:32.867458 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-02 01:43:33.367436131 +0000 UTC m=+150.534180171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.867452 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ebb5372-c005-4a86-b356-f36a3078feb2-utilities\") pod \"community-operators-v6n5w\" (UID: \"1ebb5372-c005-4a86-b356-f36a3078feb2\") " pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.886823 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt2gz\" (UniqueName: \"kubernetes.io/projected/1ebb5372-c005-4a86-b356-f36a3078feb2-kube-api-access-zt2gz\") pod \"community-operators-v6n5w\" (UID: \"1ebb5372-c005-4a86-b356-f36a3078feb2\") " pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.925175 4775 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-02T01:43:32.491575501Z","Handler":null,"Name":""} Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.930052 4775 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.930082 4775 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.954709 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.967578 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.969609 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.969648 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:32 crc kubenswrapper[4775]: I1002 01:43:32.990778 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nv4tc\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.044735 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zj54d"] Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.101932 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.106346 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.150363 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.155244 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6s8gr"] Oct 02 01:43:33 crc kubenswrapper[4775]: W1002 01:43:33.171386 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5884c55d_52c3_4bf0_bab6_d20aaf245f97.slice/crio-d5b2c629c8d3b91e14b73875c267e3a0a91e6c1bcb97838c3922184d9c073dde WatchSource:0}: Error finding container d5b2c629c8d3b91e14b73875c267e3a0a91e6c1bcb97838c3922184d9c073dde: Status 404 returned error can't find the container with id d5b2c629c8d3b91e14b73875c267e3a0a91e6c1bcb97838c3922184d9c073dde Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.222845 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6n5w"] Oct 02 01:43:33 crc kubenswrapper[4775]: W1002 01:43:33.231492 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ebb5372_c005_4a86_b356_f36a3078feb2.slice/crio-854c8e4b69276a3fef89a53c59e2eeed2eb17d1f82bece841cc4610c5a95caff WatchSource:0}: Error finding container 854c8e4b69276a3fef89a53c59e2eeed2eb17d1f82bece841cc4610c5a95caff: Status 404 returned error can't find the container with id 854c8e4b69276a3fef89a53c59e2eeed2eb17d1f82bece841cc4610c5a95caff Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.386637 4775 generic.go:334] "Generic (PLEG): container finished" podID="5884c55d-52c3-4bf0-bab6-d20aaf245f97" containerID="c111e43ee0dc208e7e1305c6a494b46925c49af0fa52a2f9658896f07b48f517" exitCode=0 Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.386707 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s8gr" event={"ID":"5884c55d-52c3-4bf0-bab6-d20aaf245f97","Type":"ContainerDied","Data":"c111e43ee0dc208e7e1305c6a494b46925c49af0fa52a2f9658896f07b48f517"} Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.386738 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s8gr" event={"ID":"5884c55d-52c3-4bf0-bab6-d20aaf245f97","Type":"ContainerStarted","Data":"d5b2c629c8d3b91e14b73875c267e3a0a91e6c1bcb97838c3922184d9c073dde"} Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.389010 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"79834eb6ab0bb42432e90099fdd8ac79edcba35c8d022cf2b8229625f8fe092c"} Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.389031 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"65362c3bc5c4b86aeebeee59b0706c53f1c863ca3362352a8c3fe0f382495816"} Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.389357 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.390859 4775 generic.go:334] "Generic (PLEG): container finished" podID="2d1808de-113f-4223-99eb-0f72d7a1b8de" containerID="16c3739cae5327a8336493d7ff12700ee47b92f838f449a7ccb697d50bac1fbc" exitCode=0 Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.391025 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zj54d" event={"ID":"2d1808de-113f-4223-99eb-0f72d7a1b8de","Type":"ContainerDied","Data":"16c3739cae5327a8336493d7ff12700ee47b92f838f449a7ccb697d50bac1fbc"} Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.391042 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zj54d" event={"ID":"2d1808de-113f-4223-99eb-0f72d7a1b8de","Type":"ContainerStarted","Data":"b48f0bcfb4276d2988743e5cb63dcf53c455135c9a9678bee700d5e26272e796"} Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.395208 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"6c26480bf039dc7b97b4433270ff97163a932e93de69fdbb2a3d75ba240ef169"} Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.395246 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"91f45d425c7b06e0354896798633aace13b85f6f21d05d5e31795e7bef912b7d"} Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.396321 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6n5w" event={"ID":"1ebb5372-c005-4a86-b356-f36a3078feb2","Type":"ContainerStarted","Data":"854c8e4b69276a3fef89a53c59e2eeed2eb17d1f82bece841cc4610c5a95caff"} Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.398285 4775 generic.go:334] "Generic (PLEG): container finished" podID="c13d75ef-3ed7-46ea-a2c6-879f1427ac41" containerID="ab4a78408ec0e7ab99fd20209e0a79003560537d24bee261cd65c21b986d4626" exitCode=0 Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.398337 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzlw6" event={"ID":"c13d75ef-3ed7-46ea-a2c6-879f1427ac41","Type":"ContainerDied","Data":"ab4a78408ec0e7ab99fd20209e0a79003560537d24bee261cd65c21b986d4626"} Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.398352 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzlw6" event={"ID":"c13d75ef-3ed7-46ea-a2c6-879f1427ac41","Type":"ContainerStarted","Data":"7fa53129fdb07b020ea7a8216ba9eb355d893cda011460e3cb8f02887aa6ac9d"} Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.400061 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6e25f66f365d272ecacca4e90236b83a20d266e8d97f7b98aeae3e22b6fac9a9"} Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.400097 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"7505824cd6030ebecbb4a68974d46f8714ca35b75110e1f1939c9886bfe73cb2"} Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.593349 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nv4tc"] Oct 02 01:43:33 crc kubenswrapper[4775]: W1002 01:43:33.603683 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadc13fc6_a2d8_44fd_ab25_075857884cad.slice/crio-edee121d93e0bedbc1dbe048021eed82daee6746e68ee586e5bc12cfbecb3c08 WatchSource:0}: Error finding container edee121d93e0bedbc1dbe048021eed82daee6746e68ee586e5bc12cfbecb3c08: Status 404 returned error can't find the container with id edee121d93e0bedbc1dbe048021eed82daee6746e68ee586e5bc12cfbecb3c08 Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.711589 4775 patch_prober.go:28] interesting pod/router-default-5444994796-cn7bz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:43:33 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Oct 02 01:43:33 crc kubenswrapper[4775]: [+]process-running ok Oct 02 01:43:33 crc kubenswrapper[4775]: healthz check failed Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.711652 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7bz" podUID="67aa1649-ec3b-4f32-96fa-d9aece180709" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:43:33 crc kubenswrapper[4775]: I1002 01:43:33.771676 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.246216 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8mr77"] Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.247713 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.249670 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.255829 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8mr77"] Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.318633 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-catalog-content\") pod \"redhat-marketplace-8mr77\" (UID: \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\") " pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.318707 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjs88\" (UniqueName: \"kubernetes.io/projected/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-kube-api-access-rjs88\") pod \"redhat-marketplace-8mr77\" (UID: \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\") " pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.318813 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-utilities\") pod \"redhat-marketplace-8mr77\" (UID: \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\") " pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.343490 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.344586 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.347462 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.347743 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.351425 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.406048 4775 generic.go:334] "Generic (PLEG): container finished" podID="1ebb5372-c005-4a86-b356-f36a3078feb2" containerID="08a69f3f6869511a79f1e2cd39c3bcd8d3953473442e6b6a941d55421b35d210" exitCode=0 Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.406120 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6n5w" event={"ID":"1ebb5372-c005-4a86-b356-f36a3078feb2","Type":"ContainerDied","Data":"08a69f3f6869511a79f1e2cd39c3bcd8d3953473442e6b6a941d55421b35d210"} Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.407873 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" event={"ID":"adc13fc6-a2d8-44fd-ab25-075857884cad","Type":"ContainerStarted","Data":"edf431bc8bde9ffadbbc3588ca4cad38ae75ab62a4b62381893a15d736d41671"} Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.407917 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" event={"ID":"adc13fc6-a2d8-44fd-ab25-075857884cad","Type":"ContainerStarted","Data":"edee121d93e0bedbc1dbe048021eed82daee6746e68ee586e5bc12cfbecb3c08"} Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.420595 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-catalog-content\") pod \"redhat-marketplace-8mr77\" (UID: \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\") " pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.420909 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjs88\" (UniqueName: \"kubernetes.io/projected/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-kube-api-access-rjs88\") pod \"redhat-marketplace-8mr77\" (UID: \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\") " pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.421089 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-utilities\") pod \"redhat-marketplace-8mr77\" (UID: \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\") " pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.421468 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-catalog-content\") pod \"redhat-marketplace-8mr77\" (UID: \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\") " pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.422233 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-utilities\") pod \"redhat-marketplace-8mr77\" (UID: \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\") " pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.442869 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjs88\" (UniqueName: \"kubernetes.io/projected/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-kube-api-access-rjs88\") pod \"redhat-marketplace-8mr77\" (UID: \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\") " pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.454439 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" podStartSLOduration=130.454423949 podStartE2EDuration="2m10.454423949s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:43:34.451482166 +0000 UTC m=+151.618226206" watchObservedRunningTime="2025-10-02 01:43:34.454423949 +0000 UTC m=+151.621167989" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.522585 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.524126 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.568729 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.625233 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.625322 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.625403 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.648667 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-78q28"] Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.652579 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.656125 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.658865 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.662708 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-78q28"] Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.716310 4775 patch_prober.go:28] interesting pod/router-default-5444994796-cn7bz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:43:34 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Oct 02 01:43:34 crc kubenswrapper[4775]: [+]process-running ok Oct 02 01:43:34 crc kubenswrapper[4775]: healthz check failed Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.716353 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7bz" podUID="67aa1649-ec3b-4f32-96fa-d9aece180709" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.813038 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8mr77"] Oct 02 01:43:34 crc kubenswrapper[4775]: W1002 01:43:34.819143 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f666cea_2dd7_4e5e_b5a2_d3a2120535d0.slice/crio-66ad110d117b33900253e43bbd91f2631ed5d7d0efb66a37f3f0943c8447edb4 WatchSource:0}: Error finding container 66ad110d117b33900253e43bbd91f2631ed5d7d0efb66a37f3f0943c8447edb4: Status 404 returned error can't find the container with id 66ad110d117b33900253e43bbd91f2631ed5d7d0efb66a37f3f0943c8447edb4 Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.827610 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9jmx\" (UniqueName: \"kubernetes.io/projected/b4ec164f-8f88-4057-8517-2727b6790a18-kube-api-access-z9jmx\") pod \"redhat-marketplace-78q28\" (UID: \"b4ec164f-8f88-4057-8517-2727b6790a18\") " pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.827709 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4ec164f-8f88-4057-8517-2727b6790a18-catalog-content\") pod \"redhat-marketplace-78q28\" (UID: \"b4ec164f-8f88-4057-8517-2727b6790a18\") " pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.827891 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4ec164f-8f88-4057-8517-2727b6790a18-utilities\") pod \"redhat-marketplace-78q28\" (UID: \"b4ec164f-8f88-4057-8517-2727b6790a18\") " pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.927855 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.928795 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4ec164f-8f88-4057-8517-2727b6790a18-utilities\") pod \"redhat-marketplace-78q28\" (UID: \"b4ec164f-8f88-4057-8517-2727b6790a18\") " pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.928829 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9jmx\" (UniqueName: \"kubernetes.io/projected/b4ec164f-8f88-4057-8517-2727b6790a18-kube-api-access-z9jmx\") pod \"redhat-marketplace-78q28\" (UID: \"b4ec164f-8f88-4057-8517-2727b6790a18\") " pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.928856 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4ec164f-8f88-4057-8517-2727b6790a18-catalog-content\") pod \"redhat-marketplace-78q28\" (UID: \"b4ec164f-8f88-4057-8517-2727b6790a18\") " pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.929389 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4ec164f-8f88-4057-8517-2727b6790a18-catalog-content\") pod \"redhat-marketplace-78q28\" (UID: \"b4ec164f-8f88-4057-8517-2727b6790a18\") " pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.929595 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4ec164f-8f88-4057-8517-2727b6790a18-utilities\") pod \"redhat-marketplace-78q28\" (UID: \"b4ec164f-8f88-4057-8517-2727b6790a18\") " pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.935593 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-47dws" Oct 02 01:43:34 crc kubenswrapper[4775]: I1002 01:43:34.954907 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9jmx\" (UniqueName: \"kubernetes.io/projected/b4ec164f-8f88-4057-8517-2727b6790a18-kube-api-access-z9jmx\") pod \"redhat-marketplace-78q28\" (UID: \"b4ec164f-8f88-4057-8517-2727b6790a18\") " pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.021003 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.116876 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.238160 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-drbkg"] Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.239322 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.251805 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.254511 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-drbkg"] Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.324979 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-78q28"] Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.333846 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ccbabf-0c44-467d-812c-79477c19e012-catalog-content\") pod \"redhat-operators-drbkg\" (UID: \"c8ccbabf-0c44-467d-812c-79477c19e012\") " pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.333940 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ccbabf-0c44-467d-812c-79477c19e012-utilities\") pod \"redhat-operators-drbkg\" (UID: \"c8ccbabf-0c44-467d-812c-79477c19e012\") " pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.333987 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7gpc\" (UniqueName: \"kubernetes.io/projected/c8ccbabf-0c44-467d-812c-79477c19e012-kube-api-access-d7gpc\") pod \"redhat-operators-drbkg\" (UID: \"c8ccbabf-0c44-467d-812c-79477c19e012\") " pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:43:35 crc kubenswrapper[4775]: W1002 01:43:35.374472 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4ec164f_8f88_4057_8517_2727b6790a18.slice/crio-34d77bd12750d3d069e4946a4daa5fb1d087e8dc9194539cb3c4c2dc47db45b3 WatchSource:0}: Error finding container 34d77bd12750d3d069e4946a4daa5fb1d087e8dc9194539cb3c4c2dc47db45b3: Status 404 returned error can't find the container with id 34d77bd12750d3d069e4946a4daa5fb1d087e8dc9194539cb3c4c2dc47db45b3 Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.416757 4775 generic.go:334] "Generic (PLEG): container finished" podID="0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" containerID="9c975c6f0575fcbd4a37bfc8d8f5c239ae0354fec3f031c733423ad259ed4eaf" exitCode=0 Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.416811 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8mr77" event={"ID":"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0","Type":"ContainerDied","Data":"9c975c6f0575fcbd4a37bfc8d8f5c239ae0354fec3f031c733423ad259ed4eaf"} Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.416835 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8mr77" event={"ID":"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0","Type":"ContainerStarted","Data":"66ad110d117b33900253e43bbd91f2631ed5d7d0efb66a37f3f0943c8447edb4"} Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.419148 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78q28" event={"ID":"b4ec164f-8f88-4057-8517-2727b6790a18","Type":"ContainerStarted","Data":"34d77bd12750d3d069e4946a4daa5fb1d087e8dc9194539cb3c4c2dc47db45b3"} Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.420627 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a","Type":"ContainerStarted","Data":"1292d12e8c8966ae9cf658d574f6338069508a0684993b899d9253b0cc3f113f"} Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.423267 4775 generic.go:334] "Generic (PLEG): container finished" podID="b9ff5f53-63d9-47b2-b1ae-703e32c35ab1" containerID="38e75b293fb6affc2f95a7680ce1755481bcef32a319098bd8b55afb6cb27312" exitCode=0 Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.423353 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" event={"ID":"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1","Type":"ContainerDied","Data":"38e75b293fb6affc2f95a7680ce1755481bcef32a319098bd8b55afb6cb27312"} Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.423558 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.435756 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ccbabf-0c44-467d-812c-79477c19e012-catalog-content\") pod \"redhat-operators-drbkg\" (UID: \"c8ccbabf-0c44-467d-812c-79477c19e012\") " pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.435834 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ccbabf-0c44-467d-812c-79477c19e012-utilities\") pod \"redhat-operators-drbkg\" (UID: \"c8ccbabf-0c44-467d-812c-79477c19e012\") " pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.435855 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7gpc\" (UniqueName: \"kubernetes.io/projected/c8ccbabf-0c44-467d-812c-79477c19e012-kube-api-access-d7gpc\") pod \"redhat-operators-drbkg\" (UID: \"c8ccbabf-0c44-467d-812c-79477c19e012\") " pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.436361 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ccbabf-0c44-467d-812c-79477c19e012-catalog-content\") pod \"redhat-operators-drbkg\" (UID: \"c8ccbabf-0c44-467d-812c-79477c19e012\") " pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.436585 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ccbabf-0c44-467d-812c-79477c19e012-utilities\") pod \"redhat-operators-drbkg\" (UID: \"c8ccbabf-0c44-467d-812c-79477c19e012\") " pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.467822 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7gpc\" (UniqueName: \"kubernetes.io/projected/c8ccbabf-0c44-467d-812c-79477c19e012-kube-api-access-d7gpc\") pod \"redhat-operators-drbkg\" (UID: \"c8ccbabf-0c44-467d-812c-79477c19e012\") " pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.575849 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.642534 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-r5qrq"] Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.644524 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.677776 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r5qrq"] Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.709342 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.712473 4775 patch_prober.go:28] interesting pod/router-default-5444994796-cn7bz container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 02 01:43:35 crc kubenswrapper[4775]: [-]has-synced failed: reason withheld Oct 02 01:43:35 crc kubenswrapper[4775]: [+]process-running ok Oct 02 01:43:35 crc kubenswrapper[4775]: healthz check failed Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.712522 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-cn7bz" podUID="67aa1649-ec3b-4f32-96fa-d9aece180709" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.742504 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574f848c-0a41-45f1-a4d1-e0b2f00328ec-catalog-content\") pod \"redhat-operators-r5qrq\" (UID: \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\") " pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.742815 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx7xs\" (UniqueName: \"kubernetes.io/projected/574f848c-0a41-45f1-a4d1-e0b2f00328ec-kube-api-access-zx7xs\") pod \"redhat-operators-r5qrq\" (UID: \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\") " pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.742860 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574f848c-0a41-45f1-a4d1-e0b2f00328ec-utilities\") pod \"redhat-operators-r5qrq\" (UID: \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\") " pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.845017 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574f848c-0a41-45f1-a4d1-e0b2f00328ec-utilities\") pod \"redhat-operators-r5qrq\" (UID: \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\") " pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.845172 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574f848c-0a41-45f1-a4d1-e0b2f00328ec-catalog-content\") pod \"redhat-operators-r5qrq\" (UID: \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\") " pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.845205 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx7xs\" (UniqueName: \"kubernetes.io/projected/574f848c-0a41-45f1-a4d1-e0b2f00328ec-kube-api-access-zx7xs\") pod \"redhat-operators-r5qrq\" (UID: \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\") " pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.846544 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574f848c-0a41-45f1-a4d1-e0b2f00328ec-utilities\") pod \"redhat-operators-r5qrq\" (UID: \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\") " pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.860612 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574f848c-0a41-45f1-a4d1-e0b2f00328ec-catalog-content\") pod \"redhat-operators-r5qrq\" (UID: \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\") " pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:43:35 crc kubenswrapper[4775]: I1002 01:43:35.866821 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx7xs\" (UniqueName: \"kubernetes.io/projected/574f848c-0a41-45f1-a4d1-e0b2f00328ec-kube-api-access-zx7xs\") pod \"redhat-operators-r5qrq\" (UID: \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\") " pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.005995 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-gvb66 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.006046 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-gvb66" podUID="63dc8fd2-1f20-4944-a8f1-783700e192f0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.006102 4775 patch_prober.go:28] interesting pod/downloads-7954f5f757-gvb66 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" start-of-body= Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.006205 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-gvb66" podUID="63dc8fd2-1f20-4944-a8f1-783700e192f0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.21:8080/\": dial tcp 10.217.0.21:8080: connect: connection refused" Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.043142 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.074915 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-drbkg"] Oct 02 01:43:36 crc kubenswrapper[4775]: W1002 01:43:36.089189 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8ccbabf_0c44_467d_812c_79477c19e012.slice/crio-05f88e0900a667b132ff7c22147aaf3b0f1352e6b1215a80ea522f0773577649 WatchSource:0}: Error finding container 05f88e0900a667b132ff7c22147aaf3b0f1352e6b1215a80ea522f0773577649: Status 404 returned error can't find the container with id 05f88e0900a667b132ff7c22147aaf3b0f1352e6b1215a80ea522f0773577649 Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.446868 4775 generic.go:334] "Generic (PLEG): container finished" podID="b4ec164f-8f88-4057-8517-2727b6790a18" containerID="ddfac46a351ec6850933d8f341312be27f08f78ce80b6b90d7b307019ebd882d" exitCode=0 Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.447318 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78q28" event={"ID":"b4ec164f-8f88-4057-8517-2727b6790a18","Type":"ContainerDied","Data":"ddfac46a351ec6850933d8f341312be27f08f78ce80b6b90d7b307019ebd882d"} Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.453391 4775 generic.go:334] "Generic (PLEG): container finished" podID="d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a" containerID="dd700299c0ff82251f47a980a7ea0c2927fabbad95902c49c99762963299a148" exitCode=0 Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.453471 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a","Type":"ContainerDied","Data":"dd700299c0ff82251f47a980a7ea0c2927fabbad95902c49c99762963299a148"} Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.459746 4775 generic.go:334] "Generic (PLEG): container finished" podID="c8ccbabf-0c44-467d-812c-79477c19e012" containerID="cf6f56dabc2f8834a410370468a3854d57aafe3f2ad97c0a6e6b71904f0227c8" exitCode=0 Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.460457 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drbkg" event={"ID":"c8ccbabf-0c44-467d-812c-79477c19e012","Type":"ContainerDied","Data":"cf6f56dabc2f8834a410370468a3854d57aafe3f2ad97c0a6e6b71904f0227c8"} Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.460480 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drbkg" event={"ID":"c8ccbabf-0c44-467d-812c-79477c19e012","Type":"ContainerStarted","Data":"05f88e0900a667b132ff7c22147aaf3b0f1352e6b1215a80ea522f0773577649"} Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.479471 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.480462 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.481637 4775 patch_prober.go:28] interesting pod/console-f9d7485db-mgrh5 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.36:8443/health\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.481695 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mgrh5" podUID="c1a3714d-b7af-49ce-9ed7-64b75dd94ce8" containerName="console" probeResult="failure" output="Get \"https://10.217.0.36:8443/health\": dial tcp 10.217.0.36:8443: connect: connection refused" Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.525558 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-r5qrq"] Oct 02 01:43:36 crc kubenswrapper[4775]: W1002 01:43:36.543361 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod574f848c_0a41_45f1_a4d1_e0b2f00328ec.slice/crio-9c909fb6d2651c20e8879e3394959c8248e03dad5ff4862ce4558924a6b986cb WatchSource:0}: Error finding container 9c909fb6d2651c20e8879e3394959c8248e03dad5ff4862ce4558924a6b986cb: Status 404 returned error can't find the container with id 9c909fb6d2651c20e8879e3394959c8248e03dad5ff4862ce4558924a6b986cb Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.715388 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.724315 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-cn7bz" Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.852405 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.976994 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6tcqs\" (UniqueName: \"kubernetes.io/projected/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-kube-api-access-6tcqs\") pod \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\" (UID: \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\") " Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.977137 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-secret-volume\") pod \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\" (UID: \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\") " Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.977175 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-config-volume\") pod \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\" (UID: \"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1\") " Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.978343 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-config-volume" (OuterVolumeSpecName: "config-volume") pod "b9ff5f53-63d9-47b2-b1ae-703e32c35ab1" (UID: "b9ff5f53-63d9-47b2-b1ae-703e32c35ab1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.983874 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-kube-api-access-6tcqs" (OuterVolumeSpecName: "kube-api-access-6tcqs") pod "b9ff5f53-63d9-47b2-b1ae-703e32c35ab1" (UID: "b9ff5f53-63d9-47b2-b1ae-703e32c35ab1"). InnerVolumeSpecName "kube-api-access-6tcqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:43:36 crc kubenswrapper[4775]: I1002 01:43:36.984343 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b9ff5f53-63d9-47b2-b1ae-703e32c35ab1" (UID: "b9ff5f53-63d9-47b2-b1ae-703e32c35ab1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.079670 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6tcqs\" (UniqueName: \"kubernetes.io/projected/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-kube-api-access-6tcqs\") on node \"crc\" DevicePath \"\"" Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.079716 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.079745 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.233627 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.233891 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.480444 4775 generic.go:334] "Generic (PLEG): container finished" podID="574f848c-0a41-45f1-a4d1-e0b2f00328ec" containerID="6ab7c03d51760ca279c06e24c53b31e641b456e24fca62f0164f7666a1e370b2" exitCode=0 Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.480525 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5qrq" event={"ID":"574f848c-0a41-45f1-a4d1-e0b2f00328ec","Type":"ContainerDied","Data":"6ab7c03d51760ca279c06e24c53b31e641b456e24fca62f0164f7666a1e370b2"} Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.480573 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5qrq" event={"ID":"574f848c-0a41-45f1-a4d1-e0b2f00328ec","Type":"ContainerStarted","Data":"9c909fb6d2651c20e8879e3394959c8248e03dad5ff4862ce4558924a6b986cb"} Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.485489 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.487126 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj" event={"ID":"b9ff5f53-63d9-47b2-b1ae-703e32c35ab1","Type":"ContainerDied","Data":"6c67550101e23727fe3d0621df9b999f7d09ded32b04617ccfac182f70f1b5ed"} Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.487160 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c67550101e23727fe3d0621df9b999f7d09ded32b04617ccfac182f70f1b5ed" Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.775868 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.891504 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a-kubelet-dir\") pod \"d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a\" (UID: \"d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a\") " Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.891575 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a-kube-api-access\") pod \"d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a\" (UID: \"d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a\") " Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.891861 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a" (UID: "d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.895516 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a" (UID: "d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.992871 4775 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 01:43:37 crc kubenswrapper[4775]: I1002 01:43:37.992907 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 01:43:38 crc kubenswrapper[4775]: I1002 01:43:38.507029 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a","Type":"ContainerDied","Data":"1292d12e8c8966ae9cf658d574f6338069508a0684993b899d9253b0cc3f113f"} Oct 02 01:43:38 crc kubenswrapper[4775]: I1002 01:43:38.507064 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1292d12e8c8966ae9cf658d574f6338069508a0684993b899d9253b0cc3f113f" Oct 02 01:43:38 crc kubenswrapper[4775]: I1002 01:43:38.507112 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 02 01:43:38 crc kubenswrapper[4775]: I1002 01:43:38.923585 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 01:43:38 crc kubenswrapper[4775]: E1002 01:43:38.923782 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9ff5f53-63d9-47b2-b1ae-703e32c35ab1" containerName="collect-profiles" Oct 02 01:43:38 crc kubenswrapper[4775]: I1002 01:43:38.923794 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9ff5f53-63d9-47b2-b1ae-703e32c35ab1" containerName="collect-profiles" Oct 02 01:43:38 crc kubenswrapper[4775]: E1002 01:43:38.923807 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a" containerName="pruner" Oct 02 01:43:38 crc kubenswrapper[4775]: I1002 01:43:38.923813 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a" containerName="pruner" Oct 02 01:43:38 crc kubenswrapper[4775]: I1002 01:43:38.923898 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9ff5f53-63d9-47b2-b1ae-703e32c35ab1" containerName="collect-profiles" Oct 02 01:43:38 crc kubenswrapper[4775]: I1002 01:43:38.923913 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f7aac8-e4f0-48d1-99d2-ae68cfcaf87a" containerName="pruner" Oct 02 01:43:38 crc kubenswrapper[4775]: I1002 01:43:38.924303 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:43:38 crc kubenswrapper[4775]: I1002 01:43:38.950610 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 02 01:43:38 crc kubenswrapper[4775]: I1002 01:43:38.950854 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 02 01:43:38 crc kubenswrapper[4775]: I1002 01:43:38.955274 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 01:43:39 crc kubenswrapper[4775]: I1002 01:43:39.108534 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57e608aa-52d0-4cba-acaf-43f045815939-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"57e608aa-52d0-4cba-acaf-43f045815939\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:43:39 crc kubenswrapper[4775]: I1002 01:43:39.108831 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57e608aa-52d0-4cba-acaf-43f045815939-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"57e608aa-52d0-4cba-acaf-43f045815939\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:43:39 crc kubenswrapper[4775]: I1002 01:43:39.209763 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57e608aa-52d0-4cba-acaf-43f045815939-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"57e608aa-52d0-4cba-acaf-43f045815939\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:43:39 crc kubenswrapper[4775]: I1002 01:43:39.209819 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57e608aa-52d0-4cba-acaf-43f045815939-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"57e608aa-52d0-4cba-acaf-43f045815939\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:43:39 crc kubenswrapper[4775]: I1002 01:43:39.210010 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57e608aa-52d0-4cba-acaf-43f045815939-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"57e608aa-52d0-4cba-acaf-43f045815939\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:43:39 crc kubenswrapper[4775]: I1002 01:43:39.224910 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57e608aa-52d0-4cba-acaf-43f045815939-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"57e608aa-52d0-4cba-acaf-43f045815939\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:43:39 crc kubenswrapper[4775]: I1002 01:43:39.277998 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:43:40 crc kubenswrapper[4775]: I1002 01:43:40.889838 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-fgvlq" Oct 02 01:43:41 crc kubenswrapper[4775]: I1002 01:43:41.931048 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:43:46 crc kubenswrapper[4775]: I1002 01:43:46.015324 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-gvb66" Oct 02 01:43:46 crc kubenswrapper[4775]: I1002 01:43:46.484195 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:46 crc kubenswrapper[4775]: I1002 01:43:46.491614 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:43:47 crc kubenswrapper[4775]: I1002 01:43:47.543795 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs\") pod \"network-metrics-daemon-x7pdc\" (UID: \"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\") " pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:43:47 crc kubenswrapper[4775]: I1002 01:43:47.550721 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ef954ceb-7b89-42cf-83dd-2e64dc9e44c1-metrics-certs\") pod \"network-metrics-daemon-x7pdc\" (UID: \"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1\") " pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:43:47 crc kubenswrapper[4775]: I1002 01:43:47.809415 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-x7pdc" Oct 02 01:43:53 crc kubenswrapper[4775]: I1002 01:43:53.161208 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:44:06 crc kubenswrapper[4775]: I1002 01:44:06.167515 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-krqxw" Oct 02 01:44:07 crc kubenswrapper[4775]: I1002 01:44:07.233809 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:44:07 crc kubenswrapper[4775]: I1002 01:44:07.233861 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:44:12 crc kubenswrapper[4775]: I1002 01:44:12.960881 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 02 01:44:21 crc kubenswrapper[4775]: E1002 01:44:21.067226 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 02 01:44:21 crc kubenswrapper[4775]: E1002 01:44:21.068012 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d7gpc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-drbkg_openshift-marketplace(c8ccbabf-0c44-467d-812c-79477c19e012): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:44:21 crc kubenswrapper[4775]: E1002 01:44:21.069547 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-drbkg" podUID="c8ccbabf-0c44-467d-812c-79477c19e012" Oct 02 01:44:22 crc kubenswrapper[4775]: E1002 01:44:22.123817 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-drbkg" podUID="c8ccbabf-0c44-467d-812c-79477c19e012" Oct 02 01:44:22 crc kubenswrapper[4775]: E1002 01:44:22.184111 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 01:44:22 crc kubenswrapper[4775]: E1002 01:44:22.184255 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-slg4j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zj54d_openshift-marketplace(2d1808de-113f-4223-99eb-0f72d7a1b8de): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:44:22 crc kubenswrapper[4775]: E1002 01:44:22.185473 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zj54d" podUID="2d1808de-113f-4223-99eb-0f72d7a1b8de" Oct 02 01:44:22 crc kubenswrapper[4775]: E1002 01:44:22.841286 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 01:44:22 crc kubenswrapper[4775]: E1002 01:44:22.841709 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z9jmx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-78q28_openshift-marketplace(b4ec164f-8f88-4057-8517-2727b6790a18): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:44:22 crc kubenswrapper[4775]: E1002 01:44:22.844225 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-78q28" podUID="b4ec164f-8f88-4057-8517-2727b6790a18" Oct 02 01:44:23 crc kubenswrapper[4775]: I1002 01:44:23.155565 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.172566 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zj54d" podUID="2d1808de-113f-4223-99eb-0f72d7a1b8de" Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.172729 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-78q28" podUID="b4ec164f-8f88-4057-8517-2727b6790a18" Oct 02 01:44:24 crc kubenswrapper[4775]: W1002 01:44:24.240793 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod57e608aa_52d0_4cba_acaf_43f045815939.slice/crio-30f230182a7708883733dd63b0c8f76d4bb0073f1e81aeb6e5f47fb6a2e148e0 WatchSource:0}: Error finding container 30f230182a7708883733dd63b0c8f76d4bb0073f1e81aeb6e5f47fb6a2e148e0: Status 404 returned error can't find the container with id 30f230182a7708883733dd63b0c8f76d4bb0073f1e81aeb6e5f47fb6a2e148e0 Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.283383 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.283544 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rjs88,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-8mr77_openshift-marketplace(0f666cea-2dd7-4e5e-b5a2-d3a2120535d0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.284837 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-8mr77" podUID="0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.296487 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.296885 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x2fxg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zzlw6_openshift-marketplace(c13d75ef-3ed7-46ea-a2c6-879f1427ac41): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.298759 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zzlw6" podUID="c13d75ef-3ed7-46ea-a2c6-879f1427ac41" Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.310316 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.310453 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zt2gz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-v6n5w_openshift-marketplace(1ebb5372-c005-4a86-b356-f36a3078feb2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.311783 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-v6n5w" podUID="1ebb5372-c005-4a86-b356-f36a3078feb2" Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.319088 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.319261 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gpc67,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-6s8gr_openshift-marketplace(5884c55d-52c3-4bf0-bab6-d20aaf245f97): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.321117 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-6s8gr" podUID="5884c55d-52c3-4bf0-bab6-d20aaf245f97" Oct 02 01:44:24 crc kubenswrapper[4775]: I1002 01:44:24.656109 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-x7pdc"] Oct 02 01:44:24 crc kubenswrapper[4775]: I1002 01:44:24.808924 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"57e608aa-52d0-4cba-acaf-43f045815939","Type":"ContainerStarted","Data":"c3c3f01120f9109bb3174ff4d006322ee901fba926c9f0a1eb50f7b164523aee"} Oct 02 01:44:24 crc kubenswrapper[4775]: I1002 01:44:24.808986 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"57e608aa-52d0-4cba-acaf-43f045815939","Type":"ContainerStarted","Data":"30f230182a7708883733dd63b0c8f76d4bb0073f1e81aeb6e5f47fb6a2e148e0"} Oct 02 01:44:24 crc kubenswrapper[4775]: I1002 01:44:24.813498 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5qrq" event={"ID":"574f848c-0a41-45f1-a4d1-e0b2f00328ec","Type":"ContainerStarted","Data":"44ffe2f6b659d2eab2fa99e2f7ba1aa535434ec993ad62ee553531963c27582a"} Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.817335 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-v6n5w" podUID="1ebb5372-c005-4a86-b356-f36a3078feb2" Oct 02 01:44:24 crc kubenswrapper[4775]: I1002 01:44:24.817809 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" event={"ID":"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1","Type":"ContainerStarted","Data":"02cbde70c12d739ac4117026b63efa119e595c2950f062d8e5b6765b948e7260"} Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.818744 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8mr77" podUID="0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.818881 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zzlw6" podUID="c13d75ef-3ed7-46ea-a2c6-879f1427ac41" Oct 02 01:44:24 crc kubenswrapper[4775]: E1002 01:44:24.819099 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-6s8gr" podUID="5884c55d-52c3-4bf0-bab6-d20aaf245f97" Oct 02 01:44:24 crc kubenswrapper[4775]: I1002 01:44:24.851839 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=46.85181609 podStartE2EDuration="46.85181609s" podCreationTimestamp="2025-10-02 01:43:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:44:24.830033503 +0000 UTC m=+201.996777593" watchObservedRunningTime="2025-10-02 01:44:24.85181609 +0000 UTC m=+202.018560140" Oct 02 01:44:25 crc kubenswrapper[4775]: I1002 01:44:25.835527 4775 generic.go:334] "Generic (PLEG): container finished" podID="574f848c-0a41-45f1-a4d1-e0b2f00328ec" containerID="44ffe2f6b659d2eab2fa99e2f7ba1aa535434ec993ad62ee553531963c27582a" exitCode=0 Oct 02 01:44:25 crc kubenswrapper[4775]: I1002 01:44:25.835627 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5qrq" event={"ID":"574f848c-0a41-45f1-a4d1-e0b2f00328ec","Type":"ContainerDied","Data":"44ffe2f6b659d2eab2fa99e2f7ba1aa535434ec993ad62ee553531963c27582a"} Oct 02 01:44:25 crc kubenswrapper[4775]: I1002 01:44:25.840507 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" event={"ID":"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1","Type":"ContainerStarted","Data":"634ed66ef36d058b8bb968d2e4ca6c06a55b609fb4854ddce8681aef0340078e"} Oct 02 01:44:25 crc kubenswrapper[4775]: I1002 01:44:25.840575 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-x7pdc" event={"ID":"ef954ceb-7b89-42cf-83dd-2e64dc9e44c1","Type":"ContainerStarted","Data":"6017996d5a20d93d0052cf667ad0d60152c7973b1d353d18bd268124f94f9fb5"} Oct 02 01:44:25 crc kubenswrapper[4775]: I1002 01:44:25.843687 4775 generic.go:334] "Generic (PLEG): container finished" podID="57e608aa-52d0-4cba-acaf-43f045815939" containerID="c3c3f01120f9109bb3174ff4d006322ee901fba926c9f0a1eb50f7b164523aee" exitCode=0 Oct 02 01:44:25 crc kubenswrapper[4775]: I1002 01:44:25.843823 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"57e608aa-52d0-4cba-acaf-43f045815939","Type":"ContainerDied","Data":"c3c3f01120f9109bb3174ff4d006322ee901fba926c9f0a1eb50f7b164523aee"} Oct 02 01:44:25 crc kubenswrapper[4775]: I1002 01:44:25.912399 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-x7pdc" podStartSLOduration=181.912366376 podStartE2EDuration="3m1.912366376s" podCreationTimestamp="2025-10-02 01:41:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:44:25.904400967 +0000 UTC m=+203.071145017" watchObservedRunningTime="2025-10-02 01:44:25.912366376 +0000 UTC m=+203.079110426" Oct 02 01:44:26 crc kubenswrapper[4775]: I1002 01:44:26.856862 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5qrq" event={"ID":"574f848c-0a41-45f1-a4d1-e0b2f00328ec","Type":"ContainerStarted","Data":"cd4e4f96b5b2f63018d81e199021fdabc9e2f086ebd3763384a6e1157dae6c57"} Oct 02 01:44:26 crc kubenswrapper[4775]: I1002 01:44:26.887764 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-r5qrq" podStartSLOduration=3.063346547 podStartE2EDuration="51.887745167s" podCreationTimestamp="2025-10-02 01:43:35 +0000 UTC" firstStartedPulling="2025-10-02 01:43:37.486877733 +0000 UTC m=+154.653621773" lastFinishedPulling="2025-10-02 01:44:26.311276363 +0000 UTC m=+203.478020393" observedRunningTime="2025-10-02 01:44:26.884926649 +0000 UTC m=+204.051670699" watchObservedRunningTime="2025-10-02 01:44:26.887745167 +0000 UTC m=+204.054489217" Oct 02 01:44:27 crc kubenswrapper[4775]: I1002 01:44:27.228757 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:44:27 crc kubenswrapper[4775]: I1002 01:44:27.406174 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57e608aa-52d0-4cba-acaf-43f045815939-kubelet-dir\") pod \"57e608aa-52d0-4cba-acaf-43f045815939\" (UID: \"57e608aa-52d0-4cba-acaf-43f045815939\") " Oct 02 01:44:27 crc kubenswrapper[4775]: I1002 01:44:27.406269 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57e608aa-52d0-4cba-acaf-43f045815939-kube-api-access\") pod \"57e608aa-52d0-4cba-acaf-43f045815939\" (UID: \"57e608aa-52d0-4cba-acaf-43f045815939\") " Oct 02 01:44:27 crc kubenswrapper[4775]: I1002 01:44:27.406476 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/57e608aa-52d0-4cba-acaf-43f045815939-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "57e608aa-52d0-4cba-acaf-43f045815939" (UID: "57e608aa-52d0-4cba-acaf-43f045815939"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:44:27 crc kubenswrapper[4775]: I1002 01:44:27.406678 4775 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57e608aa-52d0-4cba-acaf-43f045815939-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 02 01:44:27 crc kubenswrapper[4775]: I1002 01:44:27.413347 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57e608aa-52d0-4cba-acaf-43f045815939-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "57e608aa-52d0-4cba-acaf-43f045815939" (UID: "57e608aa-52d0-4cba-acaf-43f045815939"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:44:27 crc kubenswrapper[4775]: I1002 01:44:27.507375 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57e608aa-52d0-4cba-acaf-43f045815939-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 02 01:44:27 crc kubenswrapper[4775]: I1002 01:44:27.864560 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"57e608aa-52d0-4cba-acaf-43f045815939","Type":"ContainerDied","Data":"30f230182a7708883733dd63b0c8f76d4bb0073f1e81aeb6e5f47fb6a2e148e0"} Oct 02 01:44:27 crc kubenswrapper[4775]: I1002 01:44:27.864647 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30f230182a7708883733dd63b0c8f76d4bb0073f1e81aeb6e5f47fb6a2e148e0" Oct 02 01:44:27 crc kubenswrapper[4775]: I1002 01:44:27.864601 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 02 01:44:36 crc kubenswrapper[4775]: I1002 01:44:36.044300 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:44:36 crc kubenswrapper[4775]: I1002 01:44:36.045149 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:44:36 crc kubenswrapper[4775]: I1002 01:44:36.924996 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:44:36 crc kubenswrapper[4775]: I1002 01:44:36.989285 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:44:37 crc kubenswrapper[4775]: I1002 01:44:37.234001 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:44:37 crc kubenswrapper[4775]: I1002 01:44:37.234084 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:44:37 crc kubenswrapper[4775]: I1002 01:44:37.234180 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:44:37 crc kubenswrapper[4775]: I1002 01:44:37.235193 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 01:44:37 crc kubenswrapper[4775]: I1002 01:44:37.235359 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb" gracePeriod=600 Oct 02 01:44:38 crc kubenswrapper[4775]: I1002 01:44:38.214823 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r5qrq"] Oct 02 01:44:38 crc kubenswrapper[4775]: I1002 01:44:38.945835 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb" exitCode=0 Oct 02 01:44:38 crc kubenswrapper[4775]: I1002 01:44:38.945985 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb"} Oct 02 01:44:38 crc kubenswrapper[4775]: I1002 01:44:38.946621 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-r5qrq" podUID="574f848c-0a41-45f1-a4d1-e0b2f00328ec" containerName="registry-server" containerID="cri-o://cd4e4f96b5b2f63018d81e199021fdabc9e2f086ebd3763384a6e1157dae6c57" gracePeriod=2 Oct 02 01:44:41 crc kubenswrapper[4775]: I1002 01:44:41.969493 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"e819a457bcc1b0b56d9d1d4506dfc8bb5cf3853dae381f750b921e36982f1712"} Oct 02 01:44:41 crc kubenswrapper[4775]: I1002 01:44:41.971411 4775 generic.go:334] "Generic (PLEG): container finished" podID="574f848c-0a41-45f1-a4d1-e0b2f00328ec" containerID="cd4e4f96b5b2f63018d81e199021fdabc9e2f086ebd3763384a6e1157dae6c57" exitCode=0 Oct 02 01:44:41 crc kubenswrapper[4775]: I1002 01:44:41.971464 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5qrq" event={"ID":"574f848c-0a41-45f1-a4d1-e0b2f00328ec","Type":"ContainerDied","Data":"cd4e4f96b5b2f63018d81e199021fdabc9e2f086ebd3763384a6e1157dae6c57"} Oct 02 01:44:43 crc kubenswrapper[4775]: I1002 01:44:43.130861 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:44:43 crc kubenswrapper[4775]: I1002 01:44:43.330179 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574f848c-0a41-45f1-a4d1-e0b2f00328ec-utilities\") pod \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\" (UID: \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\") " Oct 02 01:44:43 crc kubenswrapper[4775]: I1002 01:44:43.330268 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574f848c-0a41-45f1-a4d1-e0b2f00328ec-catalog-content\") pod \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\" (UID: \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\") " Oct 02 01:44:43 crc kubenswrapper[4775]: I1002 01:44:43.330334 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx7xs\" (UniqueName: \"kubernetes.io/projected/574f848c-0a41-45f1-a4d1-e0b2f00328ec-kube-api-access-zx7xs\") pod \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\" (UID: \"574f848c-0a41-45f1-a4d1-e0b2f00328ec\") " Oct 02 01:44:43 crc kubenswrapper[4775]: I1002 01:44:43.331693 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/574f848c-0a41-45f1-a4d1-e0b2f00328ec-utilities" (OuterVolumeSpecName: "utilities") pod "574f848c-0a41-45f1-a4d1-e0b2f00328ec" (UID: "574f848c-0a41-45f1-a4d1-e0b2f00328ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:44:43 crc kubenswrapper[4775]: I1002 01:44:43.337579 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/574f848c-0a41-45f1-a4d1-e0b2f00328ec-kube-api-access-zx7xs" (OuterVolumeSpecName: "kube-api-access-zx7xs") pod "574f848c-0a41-45f1-a4d1-e0b2f00328ec" (UID: "574f848c-0a41-45f1-a4d1-e0b2f00328ec"). InnerVolumeSpecName "kube-api-access-zx7xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:44:43 crc kubenswrapper[4775]: I1002 01:44:43.431941 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx7xs\" (UniqueName: \"kubernetes.io/projected/574f848c-0a41-45f1-a4d1-e0b2f00328ec-kube-api-access-zx7xs\") on node \"crc\" DevicePath \"\"" Oct 02 01:44:43 crc kubenswrapper[4775]: I1002 01:44:43.432011 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/574f848c-0a41-45f1-a4d1-e0b2f00328ec-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:44:43 crc kubenswrapper[4775]: I1002 01:44:43.697320 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/574f848c-0a41-45f1-a4d1-e0b2f00328ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "574f848c-0a41-45f1-a4d1-e0b2f00328ec" (UID: "574f848c-0a41-45f1-a4d1-e0b2f00328ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:44:43 crc kubenswrapper[4775]: I1002 01:44:43.734769 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/574f848c-0a41-45f1-a4d1-e0b2f00328ec-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:44:43 crc kubenswrapper[4775]: I1002 01:44:43.984384 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-r5qrq" event={"ID":"574f848c-0a41-45f1-a4d1-e0b2f00328ec","Type":"ContainerDied","Data":"9c909fb6d2651c20e8879e3394959c8248e03dad5ff4862ce4558924a6b986cb"} Oct 02 01:44:43 crc kubenswrapper[4775]: I1002 01:44:43.984447 4775 scope.go:117] "RemoveContainer" containerID="cd4e4f96b5b2f63018d81e199021fdabc9e2f086ebd3763384a6e1157dae6c57" Oct 02 01:44:43 crc kubenswrapper[4775]: I1002 01:44:43.984786 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-r5qrq" Oct 02 01:44:44 crc kubenswrapper[4775]: I1002 01:44:44.008750 4775 scope.go:117] "RemoveContainer" containerID="44ffe2f6b659d2eab2fa99e2f7ba1aa535434ec993ad62ee553531963c27582a" Oct 02 01:44:44 crc kubenswrapper[4775]: I1002 01:44:44.036170 4775 scope.go:117] "RemoveContainer" containerID="6ab7c03d51760ca279c06e24c53b31e641b456e24fca62f0164f7666a1e370b2" Oct 02 01:44:44 crc kubenswrapper[4775]: I1002 01:44:44.037026 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-r5qrq"] Oct 02 01:44:44 crc kubenswrapper[4775]: I1002 01:44:44.050078 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-r5qrq"] Oct 02 01:44:44 crc kubenswrapper[4775]: I1002 01:44:44.989946 4775 generic.go:334] "Generic (PLEG): container finished" podID="c8ccbabf-0c44-467d-812c-79477c19e012" containerID="2f11b0584a76dab4307e3624ed58106b8ed4e7f3c2ca57957276a319f0690ca5" exitCode=0 Oct 02 01:44:44 crc kubenswrapper[4775]: I1002 01:44:44.990022 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drbkg" event={"ID":"c8ccbabf-0c44-467d-812c-79477c19e012","Type":"ContainerDied","Data":"2f11b0584a76dab4307e3624ed58106b8ed4e7f3c2ca57957276a319f0690ca5"} Oct 02 01:44:44 crc kubenswrapper[4775]: I1002 01:44:44.991886 4775 generic.go:334] "Generic (PLEG): container finished" podID="0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" containerID="ea2c0f7560d7d3ec7004ec2cbc0e88db0ff7eae8adf65345e209ac67816aff1d" exitCode=0 Oct 02 01:44:44 crc kubenswrapper[4775]: I1002 01:44:44.991934 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8mr77" event={"ID":"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0","Type":"ContainerDied","Data":"ea2c0f7560d7d3ec7004ec2cbc0e88db0ff7eae8adf65345e209ac67816aff1d"} Oct 02 01:44:44 crc kubenswrapper[4775]: I1002 01:44:44.997707 4775 generic.go:334] "Generic (PLEG): container finished" podID="b4ec164f-8f88-4057-8517-2727b6790a18" containerID="36f547bcb1a6e8098586ebe3679d9953e9f5182f44604215ed0ffc40a2e6c583" exitCode=0 Oct 02 01:44:44 crc kubenswrapper[4775]: I1002 01:44:44.997741 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78q28" event={"ID":"b4ec164f-8f88-4057-8517-2727b6790a18","Type":"ContainerDied","Data":"36f547bcb1a6e8098586ebe3679d9953e9f5182f44604215ed0ffc40a2e6c583"} Oct 02 01:44:45 crc kubenswrapper[4775]: I1002 01:44:45.003373 4775 generic.go:334] "Generic (PLEG): container finished" podID="5884c55d-52c3-4bf0-bab6-d20aaf245f97" containerID="ca4a4f9a3d7e3120b4976992fd7f6d80634d96657bc77768c8ae0a214da0d528" exitCode=0 Oct 02 01:44:45 crc kubenswrapper[4775]: I1002 01:44:45.003418 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s8gr" event={"ID":"5884c55d-52c3-4bf0-bab6-d20aaf245f97","Type":"ContainerDied","Data":"ca4a4f9a3d7e3120b4976992fd7f6d80634d96657bc77768c8ae0a214da0d528"} Oct 02 01:44:45 crc kubenswrapper[4775]: I1002 01:44:45.006112 4775 generic.go:334] "Generic (PLEG): container finished" podID="2d1808de-113f-4223-99eb-0f72d7a1b8de" containerID="2ec623203c86b2ad1073eee1eab7a31ee2a78303144ca39b8628cdd524c9b269" exitCode=0 Oct 02 01:44:45 crc kubenswrapper[4775]: I1002 01:44:45.006222 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zj54d" event={"ID":"2d1808de-113f-4223-99eb-0f72d7a1b8de","Type":"ContainerDied","Data":"2ec623203c86b2ad1073eee1eab7a31ee2a78303144ca39b8628cdd524c9b269"} Oct 02 01:44:45 crc kubenswrapper[4775]: I1002 01:44:45.008737 4775 generic.go:334] "Generic (PLEG): container finished" podID="1ebb5372-c005-4a86-b356-f36a3078feb2" containerID="2dd82b0df1465ac5c331d7e6a0db07b5ebcb927fd87a1c048bd1cddbec28ba11" exitCode=0 Oct 02 01:44:45 crc kubenswrapper[4775]: I1002 01:44:45.008811 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6n5w" event={"ID":"1ebb5372-c005-4a86-b356-f36a3078feb2","Type":"ContainerDied","Data":"2dd82b0df1465ac5c331d7e6a0db07b5ebcb927fd87a1c048bd1cddbec28ba11"} Oct 02 01:44:45 crc kubenswrapper[4775]: I1002 01:44:45.014098 4775 generic.go:334] "Generic (PLEG): container finished" podID="c13d75ef-3ed7-46ea-a2c6-879f1427ac41" containerID="ebe7fec64953fd11163fa5282af0a9cf3a6e1568a1469a787c9052026143b42f" exitCode=0 Oct 02 01:44:45 crc kubenswrapper[4775]: I1002 01:44:45.014164 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzlw6" event={"ID":"c13d75ef-3ed7-46ea-a2c6-879f1427ac41","Type":"ContainerDied","Data":"ebe7fec64953fd11163fa5282af0a9cf3a6e1568a1469a787c9052026143b42f"} Oct 02 01:44:45 crc kubenswrapper[4775]: I1002 01:44:45.811085 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="574f848c-0a41-45f1-a4d1-e0b2f00328ec" path="/var/lib/kubelet/pods/574f848c-0a41-45f1-a4d1-e0b2f00328ec/volumes" Oct 02 01:44:46 crc kubenswrapper[4775]: I1002 01:44:46.023943 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzlw6" event={"ID":"c13d75ef-3ed7-46ea-a2c6-879f1427ac41","Type":"ContainerStarted","Data":"84e601cc3683e54ef987352c8987aeb1a8e13742bfd22f4b70f473a5c26d8a4a"} Oct 02 01:44:46 crc kubenswrapper[4775]: I1002 01:44:46.030690 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drbkg" event={"ID":"c8ccbabf-0c44-467d-812c-79477c19e012","Type":"ContainerStarted","Data":"868175f2791bd40663953c18e25df7b84e007a69401f15db9feeaf4a7a2f15fc"} Oct 02 01:44:46 crc kubenswrapper[4775]: I1002 01:44:46.033056 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8mr77" event={"ID":"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0","Type":"ContainerStarted","Data":"b6581c6c4f123e1ef81dfec9083a405d3392019901bf2942712a4b738236fd9c"} Oct 02 01:44:46 crc kubenswrapper[4775]: I1002 01:44:46.035475 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78q28" event={"ID":"b4ec164f-8f88-4057-8517-2727b6790a18","Type":"ContainerStarted","Data":"7f0304b9eebd1ac2e76635991f2984ac5ac48e5fafae67622a2e5c1f2529219d"} Oct 02 01:44:46 crc kubenswrapper[4775]: I1002 01:44:46.037636 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s8gr" event={"ID":"5884c55d-52c3-4bf0-bab6-d20aaf245f97","Type":"ContainerStarted","Data":"6080885fc39c0cbc2f9f978d5f0301541ae677fccbd99e31adedbfe32b04e8c4"} Oct 02 01:44:46 crc kubenswrapper[4775]: I1002 01:44:46.038947 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6n5w" event={"ID":"1ebb5372-c005-4a86-b356-f36a3078feb2","Type":"ContainerStarted","Data":"b2c84ebeffb94c9502372752a4b3104ae617678c8d1df671388b3beb8618e5ee"} Oct 02 01:44:46 crc kubenswrapper[4775]: I1002 01:44:46.063748 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zzlw6" podStartSLOduration=2.020575423 podStartE2EDuration="1m14.063734306s" podCreationTimestamp="2025-10-02 01:43:32 +0000 UTC" firstStartedPulling="2025-10-02 01:43:33.39923821 +0000 UTC m=+150.565982250" lastFinishedPulling="2025-10-02 01:44:45.442397093 +0000 UTC m=+222.609141133" observedRunningTime="2025-10-02 01:44:46.061728961 +0000 UTC m=+223.228473001" watchObservedRunningTime="2025-10-02 01:44:46.063734306 +0000 UTC m=+223.230478346" Oct 02 01:44:46 crc kubenswrapper[4775]: I1002 01:44:46.077275 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-drbkg" podStartSLOduration=2.043140979 podStartE2EDuration="1m11.077242936s" podCreationTimestamp="2025-10-02 01:43:35 +0000 UTC" firstStartedPulling="2025-10-02 01:43:36.462002168 +0000 UTC m=+153.628746208" lastFinishedPulling="2025-10-02 01:44:45.496104105 +0000 UTC m=+222.662848165" observedRunningTime="2025-10-02 01:44:46.075082527 +0000 UTC m=+223.241826577" watchObservedRunningTime="2025-10-02 01:44:46.077242936 +0000 UTC m=+223.243986976" Oct 02 01:44:46 crc kubenswrapper[4775]: I1002 01:44:46.093055 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v6n5w" podStartSLOduration=3.031582835 podStartE2EDuration="1m14.093036929s" podCreationTimestamp="2025-10-02 01:43:32 +0000 UTC" firstStartedPulling="2025-10-02 01:43:34.408136974 +0000 UTC m=+151.574881014" lastFinishedPulling="2025-10-02 01:44:45.469591058 +0000 UTC m=+222.636335108" observedRunningTime="2025-10-02 01:44:46.089924263 +0000 UTC m=+223.256668303" watchObservedRunningTime="2025-10-02 01:44:46.093036929 +0000 UTC m=+223.259780969" Oct 02 01:44:46 crc kubenswrapper[4775]: I1002 01:44:46.120278 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8mr77" podStartSLOduration=1.874394176 podStartE2EDuration="1m12.120246704s" podCreationTimestamp="2025-10-02 01:43:34 +0000 UTC" firstStartedPulling="2025-10-02 01:43:35.426975998 +0000 UTC m=+152.593720038" lastFinishedPulling="2025-10-02 01:44:45.672828516 +0000 UTC m=+222.839572566" observedRunningTime="2025-10-02 01:44:46.118083395 +0000 UTC m=+223.284827435" watchObservedRunningTime="2025-10-02 01:44:46.120246704 +0000 UTC m=+223.286990744" Oct 02 01:44:46 crc kubenswrapper[4775]: I1002 01:44:46.135909 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-78q28" podStartSLOduration=3.086888087 podStartE2EDuration="1m12.135895983s" podCreationTimestamp="2025-10-02 01:43:34 +0000 UTC" firstStartedPulling="2025-10-02 01:43:36.449208927 +0000 UTC m=+153.615952967" lastFinishedPulling="2025-10-02 01:44:45.498216813 +0000 UTC m=+222.664960863" observedRunningTime="2025-10-02 01:44:46.132929762 +0000 UTC m=+223.299673802" watchObservedRunningTime="2025-10-02 01:44:46.135895983 +0000 UTC m=+223.302640023" Oct 02 01:44:46 crc kubenswrapper[4775]: I1002 01:44:46.147154 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6s8gr" podStartSLOduration=1.909786899 podStartE2EDuration="1m14.147139751s" podCreationTimestamp="2025-10-02 01:43:32 +0000 UTC" firstStartedPulling="2025-10-02 01:43:33.389108574 +0000 UTC m=+150.555852614" lastFinishedPulling="2025-10-02 01:44:45.626461426 +0000 UTC m=+222.793205466" observedRunningTime="2025-10-02 01:44:46.1467392 +0000 UTC m=+223.313483240" watchObservedRunningTime="2025-10-02 01:44:46.147139751 +0000 UTC m=+223.313883791" Oct 02 01:44:47 crc kubenswrapper[4775]: I1002 01:44:47.046669 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zj54d" event={"ID":"2d1808de-113f-4223-99eb-0f72d7a1b8de","Type":"ContainerStarted","Data":"3e174e58fc35e133b2319c734bdb722ea5821e22f2979a6b0e3fdc53b0c4b18b"} Oct 02 01:44:47 crc kubenswrapper[4775]: I1002 01:44:47.066873 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zj54d" podStartSLOduration=2.5854948330000003 podStartE2EDuration="1m15.066856388s" podCreationTimestamp="2025-10-02 01:43:32 +0000 UTC" firstStartedPulling="2025-10-02 01:43:33.391764729 +0000 UTC m=+150.558508769" lastFinishedPulling="2025-10-02 01:44:45.873126284 +0000 UTC m=+223.039870324" observedRunningTime="2025-10-02 01:44:47.064669158 +0000 UTC m=+224.231413198" watchObservedRunningTime="2025-10-02 01:44:47.066856388 +0000 UTC m=+224.233600428" Oct 02 01:44:52 crc kubenswrapper[4775]: I1002 01:44:52.415582 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:44:52 crc kubenswrapper[4775]: I1002 01:44:52.416106 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:44:52 crc kubenswrapper[4775]: I1002 01:44:52.465066 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:44:52 crc kubenswrapper[4775]: I1002 01:44:52.596463 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:44:52 crc kubenswrapper[4775]: I1002 01:44:52.596519 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:44:52 crc kubenswrapper[4775]: I1002 01:44:52.666331 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:44:52 crc kubenswrapper[4775]: I1002 01:44:52.784287 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:44:52 crc kubenswrapper[4775]: I1002 01:44:52.784316 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:44:52 crc kubenswrapper[4775]: I1002 01:44:52.842656 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:44:52 crc kubenswrapper[4775]: I1002 01:44:52.955729 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:44:52 crc kubenswrapper[4775]: I1002 01:44:52.955796 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:44:53 crc kubenswrapper[4775]: I1002 01:44:53.004755 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:44:53 crc kubenswrapper[4775]: I1002 01:44:53.116802 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:44:53 crc kubenswrapper[4775]: I1002 01:44:53.121812 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:44:53 crc kubenswrapper[4775]: I1002 01:44:53.123411 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:44:53 crc kubenswrapper[4775]: I1002 01:44:53.142287 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:44:53 crc kubenswrapper[4775]: I1002 01:44:53.811492 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v6n5w"] Oct 02 01:44:54 crc kubenswrapper[4775]: I1002 01:44:54.569006 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:44:54 crc kubenswrapper[4775]: I1002 01:44:54.569048 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:44:54 crc kubenswrapper[4775]: I1002 01:44:54.636640 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:44:55 crc kubenswrapper[4775]: I1002 01:44:55.022649 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:44:55 crc kubenswrapper[4775]: I1002 01:44:55.022747 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:44:55 crc kubenswrapper[4775]: I1002 01:44:55.089033 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:44:55 crc kubenswrapper[4775]: I1002 01:44:55.090032 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v6n5w" podUID="1ebb5372-c005-4a86-b356-f36a3078feb2" containerName="registry-server" containerID="cri-o://b2c84ebeffb94c9502372752a4b3104ae617678c8d1df671388b3beb8618e5ee" gracePeriod=2 Oct 02 01:44:55 crc kubenswrapper[4775]: I1002 01:44:55.134602 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:44:55 crc kubenswrapper[4775]: I1002 01:44:55.142726 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:44:55 crc kubenswrapper[4775]: I1002 01:44:55.206351 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6s8gr"] Oct 02 01:44:55 crc kubenswrapper[4775]: I1002 01:44:55.206547 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6s8gr" podUID="5884c55d-52c3-4bf0-bab6-d20aaf245f97" containerName="registry-server" containerID="cri-o://6080885fc39c0cbc2f9f978d5f0301541ae677fccbd99e31adedbfe32b04e8c4" gracePeriod=2 Oct 02 01:44:55 crc kubenswrapper[4775]: I1002 01:44:55.576686 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:44:55 crc kubenswrapper[4775]: I1002 01:44:55.576944 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:44:55 crc kubenswrapper[4775]: I1002 01:44:55.596139 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gfcf"] Oct 02 01:44:55 crc kubenswrapper[4775]: I1002 01:44:55.663925 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:44:56 crc kubenswrapper[4775]: I1002 01:44:56.144611 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:44:57 crc kubenswrapper[4775]: I1002 01:44:57.101694 4775 generic.go:334] "Generic (PLEG): container finished" podID="5884c55d-52c3-4bf0-bab6-d20aaf245f97" containerID="6080885fc39c0cbc2f9f978d5f0301541ae677fccbd99e31adedbfe32b04e8c4" exitCode=0 Oct 02 01:44:57 crc kubenswrapper[4775]: I1002 01:44:57.101783 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s8gr" event={"ID":"5884c55d-52c3-4bf0-bab6-d20aaf245f97","Type":"ContainerDied","Data":"6080885fc39c0cbc2f9f978d5f0301541ae677fccbd99e31adedbfe32b04e8c4"} Oct 02 01:44:57 crc kubenswrapper[4775]: I1002 01:44:57.609327 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-78q28"] Oct 02 01:44:57 crc kubenswrapper[4775]: I1002 01:44:57.610144 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-78q28" podUID="b4ec164f-8f88-4057-8517-2727b6790a18" containerName="registry-server" containerID="cri-o://7f0304b9eebd1ac2e76635991f2984ac5ac48e5fafae67622a2e5c1f2529219d" gracePeriod=2 Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.108011 4775 generic.go:334] "Generic (PLEG): container finished" podID="1ebb5372-c005-4a86-b356-f36a3078feb2" containerID="b2c84ebeffb94c9502372752a4b3104ae617678c8d1df671388b3beb8618e5ee" exitCode=0 Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.108163 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6n5w" event={"ID":"1ebb5372-c005-4a86-b356-f36a3078feb2","Type":"ContainerDied","Data":"b2c84ebeffb94c9502372752a4b3104ae617678c8d1df671388b3beb8618e5ee"} Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.350315 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.428368 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ebb5372-c005-4a86-b356-f36a3078feb2-catalog-content\") pod \"1ebb5372-c005-4a86-b356-f36a3078feb2\" (UID: \"1ebb5372-c005-4a86-b356-f36a3078feb2\") " Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.428652 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ebb5372-c005-4a86-b356-f36a3078feb2-utilities\") pod \"1ebb5372-c005-4a86-b356-f36a3078feb2\" (UID: \"1ebb5372-c005-4a86-b356-f36a3078feb2\") " Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.428705 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zt2gz\" (UniqueName: \"kubernetes.io/projected/1ebb5372-c005-4a86-b356-f36a3078feb2-kube-api-access-zt2gz\") pod \"1ebb5372-c005-4a86-b356-f36a3078feb2\" (UID: \"1ebb5372-c005-4a86-b356-f36a3078feb2\") " Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.430137 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ebb5372-c005-4a86-b356-f36a3078feb2-utilities" (OuterVolumeSpecName: "utilities") pod "1ebb5372-c005-4a86-b356-f36a3078feb2" (UID: "1ebb5372-c005-4a86-b356-f36a3078feb2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.434083 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ebb5372-c005-4a86-b356-f36a3078feb2-kube-api-access-zt2gz" (OuterVolumeSpecName: "kube-api-access-zt2gz") pod "1ebb5372-c005-4a86-b356-f36a3078feb2" (UID: "1ebb5372-c005-4a86-b356-f36a3078feb2"). InnerVolumeSpecName "kube-api-access-zt2gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.480512 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.530285 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zt2gz\" (UniqueName: \"kubernetes.io/projected/1ebb5372-c005-4a86-b356-f36a3078feb2-kube-api-access-zt2gz\") on node \"crc\" DevicePath \"\"" Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.530312 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ebb5372-c005-4a86-b356-f36a3078feb2-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.631776 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5884c55d-52c3-4bf0-bab6-d20aaf245f97-utilities\") pod \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\" (UID: \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\") " Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.632002 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpc67\" (UniqueName: \"kubernetes.io/projected/5884c55d-52c3-4bf0-bab6-d20aaf245f97-kube-api-access-gpc67\") pod \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\" (UID: \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\") " Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.632067 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5884c55d-52c3-4bf0-bab6-d20aaf245f97-catalog-content\") pod \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\" (UID: \"5884c55d-52c3-4bf0-bab6-d20aaf245f97\") " Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.632868 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5884c55d-52c3-4bf0-bab6-d20aaf245f97-utilities" (OuterVolumeSpecName: "utilities") pod "5884c55d-52c3-4bf0-bab6-d20aaf245f97" (UID: "5884c55d-52c3-4bf0-bab6-d20aaf245f97"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.635453 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5884c55d-52c3-4bf0-bab6-d20aaf245f97-kube-api-access-gpc67" (OuterVolumeSpecName: "kube-api-access-gpc67") pod "5884c55d-52c3-4bf0-bab6-d20aaf245f97" (UID: "5884c55d-52c3-4bf0-bab6-d20aaf245f97"). InnerVolumeSpecName "kube-api-access-gpc67". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.637522 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5884c55d-52c3-4bf0-bab6-d20aaf245f97-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:44:58 crc kubenswrapper[4775]: I1002 01:44:58.637559 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpc67\" (UniqueName: \"kubernetes.io/projected/5884c55d-52c3-4bf0-bab6-d20aaf245f97-kube-api-access-gpc67\") on node \"crc\" DevicePath \"\"" Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.062520 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ebb5372-c005-4a86-b356-f36a3078feb2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ebb5372-c005-4a86-b356-f36a3078feb2" (UID: "1ebb5372-c005-4a86-b356-f36a3078feb2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.114863 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6n5w" Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.114882 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6n5w" event={"ID":"1ebb5372-c005-4a86-b356-f36a3078feb2","Type":"ContainerDied","Data":"854c8e4b69276a3fef89a53c59e2eeed2eb17d1f82bece841cc4610c5a95caff"} Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.114924 4775 scope.go:117] "RemoveContainer" containerID="b2c84ebeffb94c9502372752a4b3104ae617678c8d1df671388b3beb8618e5ee" Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.117284 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6s8gr" event={"ID":"5884c55d-52c3-4bf0-bab6-d20aaf245f97","Type":"ContainerDied","Data":"d5b2c629c8d3b91e14b73875c267e3a0a91e6c1bcb97838c3922184d9c073dde"} Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.117382 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6s8gr" Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.131096 4775 scope.go:117] "RemoveContainer" containerID="2dd82b0df1465ac5c331d7e6a0db07b5ebcb927fd87a1c048bd1cddbec28ba11" Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.150751 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ebb5372-c005-4a86-b356-f36a3078feb2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.163191 4775 scope.go:117] "RemoveContainer" containerID="08a69f3f6869511a79f1e2cd39c3bcd8d3953473442e6b6a941d55421b35d210" Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.173030 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v6n5w"] Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.176281 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v6n5w"] Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.206634 4775 scope.go:117] "RemoveContainer" containerID="6080885fc39c0cbc2f9f978d5f0301541ae677fccbd99e31adedbfe32b04e8c4" Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.233729 4775 scope.go:117] "RemoveContainer" containerID="ca4a4f9a3d7e3120b4976992fd7f6d80634d96657bc77768c8ae0a214da0d528" Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.255750 4775 scope.go:117] "RemoveContainer" containerID="c111e43ee0dc208e7e1305c6a494b46925c49af0fa52a2f9658896f07b48f517" Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.400172 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5884c55d-52c3-4bf0-bab6-d20aaf245f97-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5884c55d-52c3-4bf0-bab6-d20aaf245f97" (UID: "5884c55d-52c3-4bf0-bab6-d20aaf245f97"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.443764 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6s8gr"] Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.456170 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6s8gr"] Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.457364 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5884c55d-52c3-4bf0-bab6-d20aaf245f97-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.778608 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ebb5372-c005-4a86-b356-f36a3078feb2" path="/var/lib/kubelet/pods/1ebb5372-c005-4a86-b356-f36a3078feb2/volumes" Oct 02 01:44:59 crc kubenswrapper[4775]: I1002 01:44:59.780308 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5884c55d-52c3-4bf0-bab6-d20aaf245f97" path="/var/lib/kubelet/pods/5884c55d-52c3-4bf0-bab6-d20aaf245f97/volumes" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.128703 4775 generic.go:334] "Generic (PLEG): container finished" podID="b4ec164f-8f88-4057-8517-2727b6790a18" containerID="7f0304b9eebd1ac2e76635991f2984ac5ac48e5fafae67622a2e5c1f2529219d" exitCode=0 Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.128852 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78q28" event={"ID":"b4ec164f-8f88-4057-8517-2727b6790a18","Type":"ContainerDied","Data":"7f0304b9eebd1ac2e76635991f2984ac5ac48e5fafae67622a2e5c1f2529219d"} Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.139220 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2"] Oct 02 01:45:00 crc kubenswrapper[4775]: E1002 01:45:00.139606 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ebb5372-c005-4a86-b356-f36a3078feb2" containerName="extract-content" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.139648 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ebb5372-c005-4a86-b356-f36a3078feb2" containerName="extract-content" Oct 02 01:45:00 crc kubenswrapper[4775]: E1002 01:45:00.139677 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ebb5372-c005-4a86-b356-f36a3078feb2" containerName="extract-utilities" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.139694 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ebb5372-c005-4a86-b356-f36a3078feb2" containerName="extract-utilities" Oct 02 01:45:00 crc kubenswrapper[4775]: E1002 01:45:00.139721 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5884c55d-52c3-4bf0-bab6-d20aaf245f97" containerName="extract-content" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.139737 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5884c55d-52c3-4bf0-bab6-d20aaf245f97" containerName="extract-content" Oct 02 01:45:00 crc kubenswrapper[4775]: E1002 01:45:00.139758 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5884c55d-52c3-4bf0-bab6-d20aaf245f97" containerName="registry-server" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.139773 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5884c55d-52c3-4bf0-bab6-d20aaf245f97" containerName="registry-server" Oct 02 01:45:00 crc kubenswrapper[4775]: E1002 01:45:00.139793 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ebb5372-c005-4a86-b356-f36a3078feb2" containerName="registry-server" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.139808 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ebb5372-c005-4a86-b356-f36a3078feb2" containerName="registry-server" Oct 02 01:45:00 crc kubenswrapper[4775]: E1002 01:45:00.139832 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574f848c-0a41-45f1-a4d1-e0b2f00328ec" containerName="extract-utilities" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.139848 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="574f848c-0a41-45f1-a4d1-e0b2f00328ec" containerName="extract-utilities" Oct 02 01:45:00 crc kubenswrapper[4775]: E1002 01:45:00.139869 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57e608aa-52d0-4cba-acaf-43f045815939" containerName="pruner" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.139885 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="57e608aa-52d0-4cba-acaf-43f045815939" containerName="pruner" Oct 02 01:45:00 crc kubenswrapper[4775]: E1002 01:45:00.139904 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574f848c-0a41-45f1-a4d1-e0b2f00328ec" containerName="registry-server" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.139920 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="574f848c-0a41-45f1-a4d1-e0b2f00328ec" containerName="registry-server" Oct 02 01:45:00 crc kubenswrapper[4775]: E1002 01:45:00.139948 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5884c55d-52c3-4bf0-bab6-d20aaf245f97" containerName="extract-utilities" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.140034 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5884c55d-52c3-4bf0-bab6-d20aaf245f97" containerName="extract-utilities" Oct 02 01:45:00 crc kubenswrapper[4775]: E1002 01:45:00.140055 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574f848c-0a41-45f1-a4d1-e0b2f00328ec" containerName="extract-content" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.140073 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="574f848c-0a41-45f1-a4d1-e0b2f00328ec" containerName="extract-content" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.140359 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ebb5372-c005-4a86-b356-f36a3078feb2" containerName="registry-server" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.140399 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5884c55d-52c3-4bf0-bab6-d20aaf245f97" containerName="registry-server" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.140433 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="574f848c-0a41-45f1-a4d1-e0b2f00328ec" containerName="registry-server" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.140451 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="57e608aa-52d0-4cba-acaf-43f045815939" containerName="pruner" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.141360 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.145279 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.145458 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.153273 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2"] Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.268303 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqmxr\" (UniqueName: \"kubernetes.io/projected/406d7724-813a-4299-90d3-c48c9f5b139e-kube-api-access-cqmxr\") pod \"collect-profiles-29322825-pjpz2\" (UID: \"406d7724-813a-4299-90d3-c48c9f5b139e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.268371 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/406d7724-813a-4299-90d3-c48c9f5b139e-secret-volume\") pod \"collect-profiles-29322825-pjpz2\" (UID: \"406d7724-813a-4299-90d3-c48c9f5b139e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.268403 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/406d7724-813a-4299-90d3-c48c9f5b139e-config-volume\") pod \"collect-profiles-29322825-pjpz2\" (UID: \"406d7724-813a-4299-90d3-c48c9f5b139e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.370022 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqmxr\" (UniqueName: \"kubernetes.io/projected/406d7724-813a-4299-90d3-c48c9f5b139e-kube-api-access-cqmxr\") pod \"collect-profiles-29322825-pjpz2\" (UID: \"406d7724-813a-4299-90d3-c48c9f5b139e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.370136 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/406d7724-813a-4299-90d3-c48c9f5b139e-secret-volume\") pod \"collect-profiles-29322825-pjpz2\" (UID: \"406d7724-813a-4299-90d3-c48c9f5b139e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.370181 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/406d7724-813a-4299-90d3-c48c9f5b139e-config-volume\") pod \"collect-profiles-29322825-pjpz2\" (UID: \"406d7724-813a-4299-90d3-c48c9f5b139e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.371289 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/406d7724-813a-4299-90d3-c48c9f5b139e-config-volume\") pod \"collect-profiles-29322825-pjpz2\" (UID: \"406d7724-813a-4299-90d3-c48c9f5b139e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.375182 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/406d7724-813a-4299-90d3-c48c9f5b139e-secret-volume\") pod \"collect-profiles-29322825-pjpz2\" (UID: \"406d7724-813a-4299-90d3-c48c9f5b139e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.388283 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqmxr\" (UniqueName: \"kubernetes.io/projected/406d7724-813a-4299-90d3-c48c9f5b139e-kube-api-access-cqmxr\") pod \"collect-profiles-29322825-pjpz2\" (UID: \"406d7724-813a-4299-90d3-c48c9f5b139e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.459428 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" Oct 02 01:45:00 crc kubenswrapper[4775]: I1002 01:45:00.688918 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2"] Oct 02 01:45:00 crc kubenswrapper[4775]: W1002 01:45:00.697792 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod406d7724_813a_4299_90d3_c48c9f5b139e.slice/crio-9c3b3b423ab697c7aa3ca5ff5dbe083bd3b3c24d37af3038eeb63ae8185f0f58 WatchSource:0}: Error finding container 9c3b3b423ab697c7aa3ca5ff5dbe083bd3b3c24d37af3038eeb63ae8185f0f58: Status 404 returned error can't find the container with id 9c3b3b423ab697c7aa3ca5ff5dbe083bd3b3c24d37af3038eeb63ae8185f0f58 Oct 02 01:45:01 crc kubenswrapper[4775]: I1002 01:45:01.136108 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" event={"ID":"406d7724-813a-4299-90d3-c48c9f5b139e","Type":"ContainerStarted","Data":"9c3b3b423ab697c7aa3ca5ff5dbe083bd3b3c24d37af3038eeb63ae8185f0f58"} Oct 02 01:45:01 crc kubenswrapper[4775]: I1002 01:45:01.499526 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:45:01 crc kubenswrapper[4775]: I1002 01:45:01.584095 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4ec164f-8f88-4057-8517-2727b6790a18-catalog-content\") pod \"b4ec164f-8f88-4057-8517-2727b6790a18\" (UID: \"b4ec164f-8f88-4057-8517-2727b6790a18\") " Oct 02 01:45:01 crc kubenswrapper[4775]: I1002 01:45:01.584151 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9jmx\" (UniqueName: \"kubernetes.io/projected/b4ec164f-8f88-4057-8517-2727b6790a18-kube-api-access-z9jmx\") pod \"b4ec164f-8f88-4057-8517-2727b6790a18\" (UID: \"b4ec164f-8f88-4057-8517-2727b6790a18\") " Oct 02 01:45:01 crc kubenswrapper[4775]: I1002 01:45:01.584205 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4ec164f-8f88-4057-8517-2727b6790a18-utilities\") pod \"b4ec164f-8f88-4057-8517-2727b6790a18\" (UID: \"b4ec164f-8f88-4057-8517-2727b6790a18\") " Oct 02 01:45:01 crc kubenswrapper[4775]: I1002 01:45:01.585155 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4ec164f-8f88-4057-8517-2727b6790a18-utilities" (OuterVolumeSpecName: "utilities") pod "b4ec164f-8f88-4057-8517-2727b6790a18" (UID: "b4ec164f-8f88-4057-8517-2727b6790a18"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:45:01 crc kubenswrapper[4775]: I1002 01:45:01.590034 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4ec164f-8f88-4057-8517-2727b6790a18-kube-api-access-z9jmx" (OuterVolumeSpecName: "kube-api-access-z9jmx") pod "b4ec164f-8f88-4057-8517-2727b6790a18" (UID: "b4ec164f-8f88-4057-8517-2727b6790a18"). InnerVolumeSpecName "kube-api-access-z9jmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:45:01 crc kubenswrapper[4775]: I1002 01:45:01.596341 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4ec164f-8f88-4057-8517-2727b6790a18-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4ec164f-8f88-4057-8517-2727b6790a18" (UID: "b4ec164f-8f88-4057-8517-2727b6790a18"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:45:01 crc kubenswrapper[4775]: I1002 01:45:01.685495 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4ec164f-8f88-4057-8517-2727b6790a18-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:01 crc kubenswrapper[4775]: I1002 01:45:01.685529 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9jmx\" (UniqueName: \"kubernetes.io/projected/b4ec164f-8f88-4057-8517-2727b6790a18-kube-api-access-z9jmx\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:01 crc kubenswrapper[4775]: I1002 01:45:01.685542 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4ec164f-8f88-4057-8517-2727b6790a18-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:02 crc kubenswrapper[4775]: I1002 01:45:02.149132 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" event={"ID":"406d7724-813a-4299-90d3-c48c9f5b139e","Type":"ContainerStarted","Data":"9ff66b564c8f4c4ce9b4c0bb0f26bd91e314c00fafbb2218aefca1f7f91394cc"} Oct 02 01:45:02 crc kubenswrapper[4775]: I1002 01:45:02.156411 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-78q28" event={"ID":"b4ec164f-8f88-4057-8517-2727b6790a18","Type":"ContainerDied","Data":"34d77bd12750d3d069e4946a4daa5fb1d087e8dc9194539cb3c4c2dc47db45b3"} Oct 02 01:45:02 crc kubenswrapper[4775]: I1002 01:45:02.156475 4775 scope.go:117] "RemoveContainer" containerID="7f0304b9eebd1ac2e76635991f2984ac5ac48e5fafae67622a2e5c1f2529219d" Oct 02 01:45:02 crc kubenswrapper[4775]: I1002 01:45:02.156482 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-78q28" Oct 02 01:45:02 crc kubenswrapper[4775]: I1002 01:45:02.165199 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" podStartSLOduration=2.165181116 podStartE2EDuration="2.165181116s" podCreationTimestamp="2025-10-02 01:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:45:02.16313473 +0000 UTC m=+239.329878770" watchObservedRunningTime="2025-10-02 01:45:02.165181116 +0000 UTC m=+239.331925156" Oct 02 01:45:02 crc kubenswrapper[4775]: I1002 01:45:02.178198 4775 scope.go:117] "RemoveContainer" containerID="36f547bcb1a6e8098586ebe3679d9953e9f5182f44604215ed0ffc40a2e6c583" Oct 02 01:45:02 crc kubenswrapper[4775]: I1002 01:45:02.179912 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-78q28"] Oct 02 01:45:02 crc kubenswrapper[4775]: I1002 01:45:02.183455 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-78q28"] Oct 02 01:45:02 crc kubenswrapper[4775]: I1002 01:45:02.191222 4775 scope.go:117] "RemoveContainer" containerID="ddfac46a351ec6850933d8f341312be27f08f78ce80b6b90d7b307019ebd882d" Oct 02 01:45:03 crc kubenswrapper[4775]: I1002 01:45:03.163367 4775 generic.go:334] "Generic (PLEG): container finished" podID="406d7724-813a-4299-90d3-c48c9f5b139e" containerID="9ff66b564c8f4c4ce9b4c0bb0f26bd91e314c00fafbb2218aefca1f7f91394cc" exitCode=0 Oct 02 01:45:03 crc kubenswrapper[4775]: I1002 01:45:03.163416 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" event={"ID":"406d7724-813a-4299-90d3-c48c9f5b139e","Type":"ContainerDied","Data":"9ff66b564c8f4c4ce9b4c0bb0f26bd91e314c00fafbb2218aefca1f7f91394cc"} Oct 02 01:45:03 crc kubenswrapper[4775]: I1002 01:45:03.771040 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4ec164f-8f88-4057-8517-2727b6790a18" path="/var/lib/kubelet/pods/b4ec164f-8f88-4057-8517-2727b6790a18/volumes" Oct 02 01:45:04 crc kubenswrapper[4775]: I1002 01:45:04.395429 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" Oct 02 01:45:04 crc kubenswrapper[4775]: I1002 01:45:04.517205 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/406d7724-813a-4299-90d3-c48c9f5b139e-config-volume\") pod \"406d7724-813a-4299-90d3-c48c9f5b139e\" (UID: \"406d7724-813a-4299-90d3-c48c9f5b139e\") " Oct 02 01:45:04 crc kubenswrapper[4775]: I1002 01:45:04.517321 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqmxr\" (UniqueName: \"kubernetes.io/projected/406d7724-813a-4299-90d3-c48c9f5b139e-kube-api-access-cqmxr\") pod \"406d7724-813a-4299-90d3-c48c9f5b139e\" (UID: \"406d7724-813a-4299-90d3-c48c9f5b139e\") " Oct 02 01:45:04 crc kubenswrapper[4775]: I1002 01:45:04.517420 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/406d7724-813a-4299-90d3-c48c9f5b139e-secret-volume\") pod \"406d7724-813a-4299-90d3-c48c9f5b139e\" (UID: \"406d7724-813a-4299-90d3-c48c9f5b139e\") " Oct 02 01:45:04 crc kubenswrapper[4775]: I1002 01:45:04.518017 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/406d7724-813a-4299-90d3-c48c9f5b139e-config-volume" (OuterVolumeSpecName: "config-volume") pod "406d7724-813a-4299-90d3-c48c9f5b139e" (UID: "406d7724-813a-4299-90d3-c48c9f5b139e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:45:04 crc kubenswrapper[4775]: I1002 01:45:04.520829 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/406d7724-813a-4299-90d3-c48c9f5b139e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "406d7724-813a-4299-90d3-c48c9f5b139e" (UID: "406d7724-813a-4299-90d3-c48c9f5b139e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:45:04 crc kubenswrapper[4775]: I1002 01:45:04.521049 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/406d7724-813a-4299-90d3-c48c9f5b139e-kube-api-access-cqmxr" (OuterVolumeSpecName: "kube-api-access-cqmxr") pod "406d7724-813a-4299-90d3-c48c9f5b139e" (UID: "406d7724-813a-4299-90d3-c48c9f5b139e"). InnerVolumeSpecName "kube-api-access-cqmxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:45:04 crc kubenswrapper[4775]: I1002 01:45:04.618744 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/406d7724-813a-4299-90d3-c48c9f5b139e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:04 crc kubenswrapper[4775]: I1002 01:45:04.618786 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/406d7724-813a-4299-90d3-c48c9f5b139e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:04 crc kubenswrapper[4775]: I1002 01:45:04.618802 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqmxr\" (UniqueName: \"kubernetes.io/projected/406d7724-813a-4299-90d3-c48c9f5b139e-kube-api-access-cqmxr\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:05 crc kubenswrapper[4775]: I1002 01:45:05.177853 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" event={"ID":"406d7724-813a-4299-90d3-c48c9f5b139e","Type":"ContainerDied","Data":"9c3b3b423ab697c7aa3ca5ff5dbe083bd3b3c24d37af3038eeb63ae8185f0f58"} Oct 02 01:45:05 crc kubenswrapper[4775]: I1002 01:45:05.177892 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9c3b3b423ab697c7aa3ca5ff5dbe083bd3b3c24d37af3038eeb63ae8185f0f58" Oct 02 01:45:05 crc kubenswrapper[4775]: I1002 01:45:05.177919 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2" Oct 02 01:45:20 crc kubenswrapper[4775]: I1002 01:45:20.636006 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" podUID="3e6497ca-0523-4658-8cbd-2340ed601e77" containerName="oauth-openshift" containerID="cri-o://e8e4ef8485f00b0fe3e9d34f3acdf5738e0cbfe250c212cfa53e36d31e54d9a4" gracePeriod=15 Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.143813 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.178079 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-748578cd96-4jndj"] Oct 02 01:45:21 crc kubenswrapper[4775]: E1002 01:45:21.178329 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="406d7724-813a-4299-90d3-c48c9f5b139e" containerName="collect-profiles" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.178343 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="406d7724-813a-4299-90d3-c48c9f5b139e" containerName="collect-profiles" Oct 02 01:45:21 crc kubenswrapper[4775]: E1002 01:45:21.178363 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e6497ca-0523-4658-8cbd-2340ed601e77" containerName="oauth-openshift" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.178371 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e6497ca-0523-4658-8cbd-2340ed601e77" containerName="oauth-openshift" Oct 02 01:45:21 crc kubenswrapper[4775]: E1002 01:45:21.178382 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ec164f-8f88-4057-8517-2727b6790a18" containerName="extract-utilities" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.178391 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ec164f-8f88-4057-8517-2727b6790a18" containerName="extract-utilities" Oct 02 01:45:21 crc kubenswrapper[4775]: E1002 01:45:21.178403 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ec164f-8f88-4057-8517-2727b6790a18" containerName="extract-content" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.178411 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ec164f-8f88-4057-8517-2727b6790a18" containerName="extract-content" Oct 02 01:45:21 crc kubenswrapper[4775]: E1002 01:45:21.178425 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4ec164f-8f88-4057-8517-2727b6790a18" containerName="registry-server" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.178433 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4ec164f-8f88-4057-8517-2727b6790a18" containerName="registry-server" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.178558 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4ec164f-8f88-4057-8517-2727b6790a18" containerName="registry-server" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.178572 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e6497ca-0523-4658-8cbd-2340ed601e77" containerName="oauth-openshift" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.178585 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="406d7724-813a-4299-90d3-c48c9f5b139e" containerName="collect-profiles" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.179049 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.204328 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-748578cd96-4jndj"] Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.244077 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-error\") pod \"3e6497ca-0523-4658-8cbd-2340ed601e77\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.244152 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-idp-0-file-data\") pod \"3e6497ca-0523-4658-8cbd-2340ed601e77\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.244193 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-session\") pod \"3e6497ca-0523-4658-8cbd-2340ed601e77\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.244226 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-audit-policies\") pod \"3e6497ca-0523-4658-8cbd-2340ed601e77\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.244259 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-ocp-branding-template\") pod \"3e6497ca-0523-4658-8cbd-2340ed601e77\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.244281 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-trusted-ca-bundle\") pod \"3e6497ca-0523-4658-8cbd-2340ed601e77\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.244302 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3e6497ca-0523-4658-8cbd-2340ed601e77-audit-dir\") pod \"3e6497ca-0523-4658-8cbd-2340ed601e77\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.244328 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-serving-cert\") pod \"3e6497ca-0523-4658-8cbd-2340ed601e77\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.244351 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-router-certs\") pod \"3e6497ca-0523-4658-8cbd-2340ed601e77\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.244376 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c69pf\" (UniqueName: \"kubernetes.io/projected/3e6497ca-0523-4658-8cbd-2340ed601e77-kube-api-access-c69pf\") pod \"3e6497ca-0523-4658-8cbd-2340ed601e77\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.244399 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-login\") pod \"3e6497ca-0523-4658-8cbd-2340ed601e77\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.244443 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-service-ca\") pod \"3e6497ca-0523-4658-8cbd-2340ed601e77\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.244468 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-cliconfig\") pod \"3e6497ca-0523-4658-8cbd-2340ed601e77\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.244491 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-provider-selection\") pod \"3e6497ca-0523-4658-8cbd-2340ed601e77\" (UID: \"3e6497ca-0523-4658-8cbd-2340ed601e77\") " Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.245390 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-cliconfig\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.245479 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-user-template-error\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.245509 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms4q9\" (UniqueName: \"kubernetes.io/projected/24a9abfe-477b-4e31-9f27-7876dd371348-kube-api-access-ms4q9\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.245561 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-serving-cert\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.245613 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-router-certs\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.245703 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-user-template-login\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.245738 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.245779 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.245818 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.245857 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/24a9abfe-477b-4e31-9f27-7876dd371348-audit-dir\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.245885 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-service-ca\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.245992 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.246044 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-session\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.246091 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/24a9abfe-477b-4e31-9f27-7876dd371348-audit-policies\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.246581 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "3e6497ca-0523-4658-8cbd-2340ed601e77" (UID: "3e6497ca-0523-4658-8cbd-2340ed601e77"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.246663 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "3e6497ca-0523-4658-8cbd-2340ed601e77" (UID: "3e6497ca-0523-4658-8cbd-2340ed601e77"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.246723 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3e6497ca-0523-4658-8cbd-2340ed601e77-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "3e6497ca-0523-4658-8cbd-2340ed601e77" (UID: "3e6497ca-0523-4658-8cbd-2340ed601e77"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.247111 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "3e6497ca-0523-4658-8cbd-2340ed601e77" (UID: "3e6497ca-0523-4658-8cbd-2340ed601e77"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.247263 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "3e6497ca-0523-4658-8cbd-2340ed601e77" (UID: "3e6497ca-0523-4658-8cbd-2340ed601e77"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.251779 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "3e6497ca-0523-4658-8cbd-2340ed601e77" (UID: "3e6497ca-0523-4658-8cbd-2340ed601e77"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.252153 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e6497ca-0523-4658-8cbd-2340ed601e77-kube-api-access-c69pf" (OuterVolumeSpecName: "kube-api-access-c69pf") pod "3e6497ca-0523-4658-8cbd-2340ed601e77" (UID: "3e6497ca-0523-4658-8cbd-2340ed601e77"). InnerVolumeSpecName "kube-api-access-c69pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.252329 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "3e6497ca-0523-4658-8cbd-2340ed601e77" (UID: "3e6497ca-0523-4658-8cbd-2340ed601e77"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.252642 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "3e6497ca-0523-4658-8cbd-2340ed601e77" (UID: "3e6497ca-0523-4658-8cbd-2340ed601e77"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.252936 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "3e6497ca-0523-4658-8cbd-2340ed601e77" (UID: "3e6497ca-0523-4658-8cbd-2340ed601e77"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.253314 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "3e6497ca-0523-4658-8cbd-2340ed601e77" (UID: "3e6497ca-0523-4658-8cbd-2340ed601e77"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.253482 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "3e6497ca-0523-4658-8cbd-2340ed601e77" (UID: "3e6497ca-0523-4658-8cbd-2340ed601e77"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.253600 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "3e6497ca-0523-4658-8cbd-2340ed601e77" (UID: "3e6497ca-0523-4658-8cbd-2340ed601e77"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.253839 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "3e6497ca-0523-4658-8cbd-2340ed601e77" (UID: "3e6497ca-0523-4658-8cbd-2340ed601e77"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.284779 4775 generic.go:334] "Generic (PLEG): container finished" podID="3e6497ca-0523-4658-8cbd-2340ed601e77" containerID="e8e4ef8485f00b0fe3e9d34f3acdf5738e0cbfe250c212cfa53e36d31e54d9a4" exitCode=0 Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.284822 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" event={"ID":"3e6497ca-0523-4658-8cbd-2340ed601e77","Type":"ContainerDied","Data":"e8e4ef8485f00b0fe3e9d34f3acdf5738e0cbfe250c212cfa53e36d31e54d9a4"} Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.284854 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" event={"ID":"3e6497ca-0523-4658-8cbd-2340ed601e77","Type":"ContainerDied","Data":"a1537427439d99ae2e6f244c1fc513b34dded7d7d394eac4542383d63474a42c"} Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.284853 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-7gfcf" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.284875 4775 scope.go:117] "RemoveContainer" containerID="e8e4ef8485f00b0fe3e9d34f3acdf5738e0cbfe250c212cfa53e36d31e54d9a4" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.314652 4775 scope.go:117] "RemoveContainer" containerID="e8e4ef8485f00b0fe3e9d34f3acdf5738e0cbfe250c212cfa53e36d31e54d9a4" Oct 02 01:45:21 crc kubenswrapper[4775]: E1002 01:45:21.315206 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8e4ef8485f00b0fe3e9d34f3acdf5738e0cbfe250c212cfa53e36d31e54d9a4\": container with ID starting with e8e4ef8485f00b0fe3e9d34f3acdf5738e0cbfe250c212cfa53e36d31e54d9a4 not found: ID does not exist" containerID="e8e4ef8485f00b0fe3e9d34f3acdf5738e0cbfe250c212cfa53e36d31e54d9a4" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.315271 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8e4ef8485f00b0fe3e9d34f3acdf5738e0cbfe250c212cfa53e36d31e54d9a4"} err="failed to get container status \"e8e4ef8485f00b0fe3e9d34f3acdf5738e0cbfe250c212cfa53e36d31e54d9a4\": rpc error: code = NotFound desc = could not find container \"e8e4ef8485f00b0fe3e9d34f3acdf5738e0cbfe250c212cfa53e36d31e54d9a4\": container with ID starting with e8e4ef8485f00b0fe3e9d34f3acdf5738e0cbfe250c212cfa53e36d31e54d9a4 not found: ID does not exist" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.334565 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gfcf"] Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.338050 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-7gfcf"] Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347447 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-router-certs\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347517 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-user-template-login\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347548 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347579 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347606 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347668 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/24a9abfe-477b-4e31-9f27-7876dd371348-audit-dir\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347692 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-service-ca\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347739 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347764 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-session\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347792 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/24a9abfe-477b-4e31-9f27-7876dd371348-audit-policies\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347818 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-cliconfig\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347845 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-user-template-error\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347865 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms4q9\" (UniqueName: \"kubernetes.io/projected/24a9abfe-477b-4e31-9f27-7876dd371348-kube-api-access-ms4q9\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347891 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-serving-cert\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347942 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.347962 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.348125 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c69pf\" (UniqueName: \"kubernetes.io/projected/3e6497ca-0523-4658-8cbd-2340ed601e77-kube-api-access-c69pf\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.349365 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.349419 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/24a9abfe-477b-4e31-9f27-7876dd371348-audit-dir\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.349132 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-service-ca\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.349601 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.349628 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.349647 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.349688 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.349708 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.349727 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.349747 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.349764 4775 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.349781 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.349829 4775 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3e6497ca-0523-4658-8cbd-2340ed601e77-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.349848 4775 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3e6497ca-0523-4658-8cbd-2340ed601e77-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.350149 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/24a9abfe-477b-4e31-9f27-7876dd371348-audit-policies\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.352233 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-user-template-login\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.352493 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-cliconfig\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.352871 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.353113 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-user-template-error\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.353364 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.354244 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-serving-cert\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.354398 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-router-certs\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.356327 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-system-session\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.357129 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/24a9abfe-477b-4e31-9f27-7876dd371348-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.365377 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms4q9\" (UniqueName: \"kubernetes.io/projected/24a9abfe-477b-4e31-9f27-7876dd371348-kube-api-access-ms4q9\") pod \"oauth-openshift-748578cd96-4jndj\" (UID: \"24a9abfe-477b-4e31-9f27-7876dd371348\") " pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.501673 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:21 crc kubenswrapper[4775]: I1002 01:45:21.777941 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e6497ca-0523-4658-8cbd-2340ed601e77" path="/var/lib/kubelet/pods/3e6497ca-0523-4658-8cbd-2340ed601e77/volumes" Oct 02 01:45:22 crc kubenswrapper[4775]: I1002 01:45:22.004342 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-748578cd96-4jndj"] Oct 02 01:45:22 crc kubenswrapper[4775]: I1002 01:45:22.292743 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" event={"ID":"24a9abfe-477b-4e31-9f27-7876dd371348","Type":"ContainerStarted","Data":"c6a3bcea5f9cd3c2f525c31f3fc73d8f2f1d1cdb9d1a3d19ccb3d6e588ed26c5"} Oct 02 01:45:23 crc kubenswrapper[4775]: I1002 01:45:23.303301 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" event={"ID":"24a9abfe-477b-4e31-9f27-7876dd371348","Type":"ContainerStarted","Data":"9f53e88c9dcdf46963e1259d03f675d3abc57825f643c33abdf85162b2f35542"} Oct 02 01:45:23 crc kubenswrapper[4775]: I1002 01:45:23.303648 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:23 crc kubenswrapper[4775]: I1002 01:45:23.315413 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" Oct 02 01:45:23 crc kubenswrapper[4775]: I1002 01:45:23.337460 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-748578cd96-4jndj" podStartSLOduration=28.337433915 podStartE2EDuration="28.337433915s" podCreationTimestamp="2025-10-02 01:44:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:45:23.332495874 +0000 UTC m=+260.499239954" watchObservedRunningTime="2025-10-02 01:45:23.337433915 +0000 UTC m=+260.504177965" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.318157 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zzlw6"] Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.319392 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zzlw6" podUID="c13d75ef-3ed7-46ea-a2c6-879f1427ac41" containerName="registry-server" containerID="cri-o://84e601cc3683e54ef987352c8987aeb1a8e13742bfd22f4b70f473a5c26d8a4a" gracePeriod=30 Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.325254 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zj54d"] Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.325512 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zj54d" podUID="2d1808de-113f-4223-99eb-0f72d7a1b8de" containerName="registry-server" containerID="cri-o://3e174e58fc35e133b2319c734bdb722ea5821e22f2979a6b0e3fdc53b0c4b18b" gracePeriod=30 Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.340111 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d8qrs"] Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.341401 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" podUID="2e1b3210-4726-450d-8da5-766903e6d91b" containerName="marketplace-operator" containerID="cri-o://c335259f9d23a5a1f5199708b3dddfb4a2e5fcaa12f0bf36c4c41679845b2865" gracePeriod=30 Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.343922 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8mr77"] Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.344152 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8mr77" podUID="0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" containerName="registry-server" containerID="cri-o://b6581c6c4f123e1ef81dfec9083a405d3392019901bf2942712a4b738236fd9c" gracePeriod=30 Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.383022 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-drbkg"] Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.383383 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-drbkg" podUID="c8ccbabf-0c44-467d-812c-79477c19e012" containerName="registry-server" containerID="cri-o://868175f2791bd40663953c18e25df7b84e007a69401f15db9feeaf4a7a2f15fc" gracePeriod=30 Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.397108 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-n4prl"] Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.398505 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.405628 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-n4prl"] Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.464268 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6f9799cb-9770-4dd2-87a3-55604cad82d2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-n4prl\" (UID: \"6f9799cb-9770-4dd2-87a3-55604cad82d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.464337 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f9799cb-9770-4dd2-87a3-55604cad82d2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-n4prl\" (UID: \"6f9799cb-9770-4dd2-87a3-55604cad82d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.464364 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vjsf\" (UniqueName: \"kubernetes.io/projected/6f9799cb-9770-4dd2-87a3-55604cad82d2-kube-api-access-2vjsf\") pod \"marketplace-operator-79b997595-n4prl\" (UID: \"6f9799cb-9770-4dd2-87a3-55604cad82d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.565314 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6f9799cb-9770-4dd2-87a3-55604cad82d2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-n4prl\" (UID: \"6f9799cb-9770-4dd2-87a3-55604cad82d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.565406 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f9799cb-9770-4dd2-87a3-55604cad82d2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-n4prl\" (UID: \"6f9799cb-9770-4dd2-87a3-55604cad82d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.565436 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vjsf\" (UniqueName: \"kubernetes.io/projected/6f9799cb-9770-4dd2-87a3-55604cad82d2-kube-api-access-2vjsf\") pod \"marketplace-operator-79b997595-n4prl\" (UID: \"6f9799cb-9770-4dd2-87a3-55604cad82d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.567766 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f9799cb-9770-4dd2-87a3-55604cad82d2-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-n4prl\" (UID: \"6f9799cb-9770-4dd2-87a3-55604cad82d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.573775 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6f9799cb-9770-4dd2-87a3-55604cad82d2-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-n4prl\" (UID: \"6f9799cb-9770-4dd2-87a3-55604cad82d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.582607 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vjsf\" (UniqueName: \"kubernetes.io/projected/6f9799cb-9770-4dd2-87a3-55604cad82d2-kube-api-access-2vjsf\") pod \"marketplace-operator-79b997595-n4prl\" (UID: \"6f9799cb-9770-4dd2-87a3-55604cad82d2\") " pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.772888 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.784370 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.789163 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.806444 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.810018 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.840769 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.869166 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d1808de-113f-4223-99eb-0f72d7a1b8de-utilities\") pod \"2d1808de-113f-4223-99eb-0f72d7a1b8de\" (UID: \"2d1808de-113f-4223-99eb-0f72d7a1b8de\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.869216 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzgzq\" (UniqueName: \"kubernetes.io/projected/2e1b3210-4726-450d-8da5-766903e6d91b-kube-api-access-gzgzq\") pod \"2e1b3210-4726-450d-8da5-766903e6d91b\" (UID: \"2e1b3210-4726-450d-8da5-766903e6d91b\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.869244 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjs88\" (UniqueName: \"kubernetes.io/projected/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-kube-api-access-rjs88\") pod \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\" (UID: \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.869264 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-catalog-content\") pod \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\" (UID: \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.869308 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d1808de-113f-4223-99eb-0f72d7a1b8de-catalog-content\") pod \"2d1808de-113f-4223-99eb-0f72d7a1b8de\" (UID: \"2d1808de-113f-4223-99eb-0f72d7a1b8de\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.869323 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-catalog-content\") pod \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\" (UID: \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.869342 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e1b3210-4726-450d-8da5-766903e6d91b-marketplace-trusted-ca\") pod \"2e1b3210-4726-450d-8da5-766903e6d91b\" (UID: \"2e1b3210-4726-450d-8da5-766903e6d91b\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.869358 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-utilities\") pod \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\" (UID: \"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.869375 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2fxg\" (UniqueName: \"kubernetes.io/projected/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-kube-api-access-x2fxg\") pod \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\" (UID: \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.869415 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-utilities\") pod \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\" (UID: \"c13d75ef-3ed7-46ea-a2c6-879f1427ac41\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.869432 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2e1b3210-4726-450d-8da5-766903e6d91b-marketplace-operator-metrics\") pod \"2e1b3210-4726-450d-8da5-766903e6d91b\" (UID: \"2e1b3210-4726-450d-8da5-766903e6d91b\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.869450 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slg4j\" (UniqueName: \"kubernetes.io/projected/2d1808de-113f-4223-99eb-0f72d7a1b8de-kube-api-access-slg4j\") pod \"2d1808de-113f-4223-99eb-0f72d7a1b8de\" (UID: \"2d1808de-113f-4223-99eb-0f72d7a1b8de\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.871507 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-utilities" (OuterVolumeSpecName: "utilities") pod "0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" (UID: "0f666cea-2dd7-4e5e-b5a2-d3a2120535d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.872902 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d1808de-113f-4223-99eb-0f72d7a1b8de-utilities" (OuterVolumeSpecName: "utilities") pod "2d1808de-113f-4223-99eb-0f72d7a1b8de" (UID: "2d1808de-113f-4223-99eb-0f72d7a1b8de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.873800 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-kube-api-access-x2fxg" (OuterVolumeSpecName: "kube-api-access-x2fxg") pod "c13d75ef-3ed7-46ea-a2c6-879f1427ac41" (UID: "c13d75ef-3ed7-46ea-a2c6-879f1427ac41"). InnerVolumeSpecName "kube-api-access-x2fxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.873843 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-utilities" (OuterVolumeSpecName: "utilities") pod "c13d75ef-3ed7-46ea-a2c6-879f1427ac41" (UID: "c13d75ef-3ed7-46ea-a2c6-879f1427ac41"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.874349 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1b3210-4726-450d-8da5-766903e6d91b-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "2e1b3210-4726-450d-8da5-766903e6d91b" (UID: "2e1b3210-4726-450d-8da5-766903e6d91b"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.874769 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d1808de-113f-4223-99eb-0f72d7a1b8de-kube-api-access-slg4j" (OuterVolumeSpecName: "kube-api-access-slg4j") pod "2d1808de-113f-4223-99eb-0f72d7a1b8de" (UID: "2d1808de-113f-4223-99eb-0f72d7a1b8de"). InnerVolumeSpecName "kube-api-access-slg4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.875435 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e1b3210-4726-450d-8da5-766903e6d91b-kube-api-access-gzgzq" (OuterVolumeSpecName: "kube-api-access-gzgzq") pod "2e1b3210-4726-450d-8da5-766903e6d91b" (UID: "2e1b3210-4726-450d-8da5-766903e6d91b"). InnerVolumeSpecName "kube-api-access-gzgzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.876430 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-kube-api-access-rjs88" (OuterVolumeSpecName: "kube-api-access-rjs88") pod "0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" (UID: "0f666cea-2dd7-4e5e-b5a2-d3a2120535d0"). InnerVolumeSpecName "kube-api-access-rjs88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.884498 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e1b3210-4726-450d-8da5-766903e6d91b-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "2e1b3210-4726-450d-8da5-766903e6d91b" (UID: "2e1b3210-4726-450d-8da5-766903e6d91b"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.903373 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" (UID: "0f666cea-2dd7-4e5e-b5a2-d3a2120535d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.935184 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d1808de-113f-4223-99eb-0f72d7a1b8de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2d1808de-113f-4223-99eb-0f72d7a1b8de" (UID: "2d1808de-113f-4223-99eb-0f72d7a1b8de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.936894 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c13d75ef-3ed7-46ea-a2c6-879f1427ac41" (UID: "c13d75ef-3ed7-46ea-a2c6-879f1427ac41"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970211 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ccbabf-0c44-467d-812c-79477c19e012-utilities\") pod \"c8ccbabf-0c44-467d-812c-79477c19e012\" (UID: \"c8ccbabf-0c44-467d-812c-79477c19e012\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970332 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7gpc\" (UniqueName: \"kubernetes.io/projected/c8ccbabf-0c44-467d-812c-79477c19e012-kube-api-access-d7gpc\") pod \"c8ccbabf-0c44-467d-812c-79477c19e012\" (UID: \"c8ccbabf-0c44-467d-812c-79477c19e012\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970388 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ccbabf-0c44-467d-812c-79477c19e012-catalog-content\") pod \"c8ccbabf-0c44-467d-812c-79477c19e012\" (UID: \"c8ccbabf-0c44-467d-812c-79477c19e012\") " Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970574 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970594 4775 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2e1b3210-4726-450d-8da5-766903e6d91b-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970605 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970614 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2fxg\" (UniqueName: \"kubernetes.io/projected/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-kube-api-access-x2fxg\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970623 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c13d75ef-3ed7-46ea-a2c6-879f1427ac41-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970631 4775 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/2e1b3210-4726-450d-8da5-766903e6d91b-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970641 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slg4j\" (UniqueName: \"kubernetes.io/projected/2d1808de-113f-4223-99eb-0f72d7a1b8de-kube-api-access-slg4j\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970649 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d1808de-113f-4223-99eb-0f72d7a1b8de-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970657 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzgzq\" (UniqueName: \"kubernetes.io/projected/2e1b3210-4726-450d-8da5-766903e6d91b-kube-api-access-gzgzq\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970665 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjs88\" (UniqueName: \"kubernetes.io/projected/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-kube-api-access-rjs88\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970674 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.970682 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d1808de-113f-4223-99eb-0f72d7a1b8de-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.971267 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8ccbabf-0c44-467d-812c-79477c19e012-utilities" (OuterVolumeSpecName: "utilities") pod "c8ccbabf-0c44-467d-812c-79477c19e012" (UID: "c8ccbabf-0c44-467d-812c-79477c19e012"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:45:43 crc kubenswrapper[4775]: I1002 01:45:43.974008 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8ccbabf-0c44-467d-812c-79477c19e012-kube-api-access-d7gpc" (OuterVolumeSpecName: "kube-api-access-d7gpc") pod "c8ccbabf-0c44-467d-812c-79477c19e012" (UID: "c8ccbabf-0c44-467d-812c-79477c19e012"). InnerVolumeSpecName "kube-api-access-d7gpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.056803 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8ccbabf-0c44-467d-812c-79477c19e012-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8ccbabf-0c44-467d-812c-79477c19e012" (UID: "c8ccbabf-0c44-467d-812c-79477c19e012"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.071587 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7gpc\" (UniqueName: \"kubernetes.io/projected/c8ccbabf-0c44-467d-812c-79477c19e012-kube-api-access-d7gpc\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.071617 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ccbabf-0c44-467d-812c-79477c19e012-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.071626 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ccbabf-0c44-467d-812c-79477c19e012-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.215354 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-n4prl"] Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.437776 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzlw6" event={"ID":"c13d75ef-3ed7-46ea-a2c6-879f1427ac41","Type":"ContainerDied","Data":"84e601cc3683e54ef987352c8987aeb1a8e13742bfd22f4b70f473a5c26d8a4a"} Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.437810 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zzlw6" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.437827 4775 scope.go:117] "RemoveContainer" containerID="84e601cc3683e54ef987352c8987aeb1a8e13742bfd22f4b70f473a5c26d8a4a" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.437730 4775 generic.go:334] "Generic (PLEG): container finished" podID="c13d75ef-3ed7-46ea-a2c6-879f1427ac41" containerID="84e601cc3683e54ef987352c8987aeb1a8e13742bfd22f4b70f473a5c26d8a4a" exitCode=0 Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.438065 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zzlw6" event={"ID":"c13d75ef-3ed7-46ea-a2c6-879f1427ac41","Type":"ContainerDied","Data":"7fa53129fdb07b020ea7a8216ba9eb355d893cda011460e3cb8f02887aa6ac9d"} Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.441392 4775 generic.go:334] "Generic (PLEG): container finished" podID="c8ccbabf-0c44-467d-812c-79477c19e012" containerID="868175f2791bd40663953c18e25df7b84e007a69401f15db9feeaf4a7a2f15fc" exitCode=0 Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.441440 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drbkg" event={"ID":"c8ccbabf-0c44-467d-812c-79477c19e012","Type":"ContainerDied","Data":"868175f2791bd40663953c18e25df7b84e007a69401f15db9feeaf4a7a2f15fc"} Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.441461 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-drbkg" event={"ID":"c8ccbabf-0c44-467d-812c-79477c19e012","Type":"ContainerDied","Data":"05f88e0900a667b132ff7c22147aaf3b0f1352e6b1215a80ea522f0773577649"} Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.441520 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-drbkg" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.451760 4775 generic.go:334] "Generic (PLEG): container finished" podID="0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" containerID="b6581c6c4f123e1ef81dfec9083a405d3392019901bf2942712a4b738236fd9c" exitCode=0 Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.451825 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8mr77" event={"ID":"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0","Type":"ContainerDied","Data":"b6581c6c4f123e1ef81dfec9083a405d3392019901bf2942712a4b738236fd9c"} Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.451852 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8mr77" event={"ID":"0f666cea-2dd7-4e5e-b5a2-d3a2120535d0","Type":"ContainerDied","Data":"66ad110d117b33900253e43bbd91f2631ed5d7d0efb66a37f3f0943c8447edb4"} Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.451915 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8mr77" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.455181 4775 generic.go:334] "Generic (PLEG): container finished" podID="2e1b3210-4726-450d-8da5-766903e6d91b" containerID="c335259f9d23a5a1f5199708b3dddfb4a2e5fcaa12f0bf36c4c41679845b2865" exitCode=0 Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.455285 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" event={"ID":"2e1b3210-4726-450d-8da5-766903e6d91b","Type":"ContainerDied","Data":"c335259f9d23a5a1f5199708b3dddfb4a2e5fcaa12f0bf36c4c41679845b2865"} Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.455293 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.455313 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-d8qrs" event={"ID":"2e1b3210-4726-450d-8da5-766903e6d91b","Type":"ContainerDied","Data":"d47590691bd098c633778e9ae8a10ac266792bfb91e92b59da1881531200e7f6"} Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.457483 4775 generic.go:334] "Generic (PLEG): container finished" podID="2d1808de-113f-4223-99eb-0f72d7a1b8de" containerID="3e174e58fc35e133b2319c734bdb722ea5821e22f2979a6b0e3fdc53b0c4b18b" exitCode=0 Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.457566 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zj54d" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.457564 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zj54d" event={"ID":"2d1808de-113f-4223-99eb-0f72d7a1b8de","Type":"ContainerDied","Data":"3e174e58fc35e133b2319c734bdb722ea5821e22f2979a6b0e3fdc53b0c4b18b"} Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.457618 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zj54d" event={"ID":"2d1808de-113f-4223-99eb-0f72d7a1b8de","Type":"ContainerDied","Data":"b48f0bcfb4276d2988743e5cb63dcf53c455135c9a9678bee700d5e26272e796"} Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.459397 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" event={"ID":"6f9799cb-9770-4dd2-87a3-55604cad82d2","Type":"ContainerStarted","Data":"e1ac18f77b2cac7e31336d0498992e0599fbce055a2e4949e53aa8d27e49cf76"} Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.459435 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" event={"ID":"6f9799cb-9770-4dd2-87a3-55604cad82d2","Type":"ContainerStarted","Data":"e379570132caf56e160acc1d97fc617e9a8fca7fb7920b61d45b8e3bfde601e7"} Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.459647 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.461332 4775 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-n4prl container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" start-of-body= Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.461369 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" podUID="6f9799cb-9770-4dd2-87a3-55604cad82d2" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.473717 4775 scope.go:117] "RemoveContainer" containerID="ebe7fec64953fd11163fa5282af0a9cf3a6e1568a1469a787c9052026143b42f" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.479255 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" podStartSLOduration=1.479232832 podStartE2EDuration="1.479232832s" podCreationTimestamp="2025-10-02 01:45:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:45:44.472775931 +0000 UTC m=+281.639519971" watchObservedRunningTime="2025-10-02 01:45:44.479232832 +0000 UTC m=+281.645976872" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.487400 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-drbkg"] Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.491117 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-drbkg"] Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.501311 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zzlw6"] Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.503775 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zzlw6"] Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.511402 4775 scope.go:117] "RemoveContainer" containerID="ab4a78408ec0e7ab99fd20209e0a79003560537d24bee261cd65c21b986d4626" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.513756 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zj54d"] Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.516394 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zj54d"] Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.521238 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8mr77"] Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.525660 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8mr77"] Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.536358 4775 scope.go:117] "RemoveContainer" containerID="84e601cc3683e54ef987352c8987aeb1a8e13742bfd22f4b70f473a5c26d8a4a" Oct 02 01:45:44 crc kubenswrapper[4775]: E1002 01:45:44.537711 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84e601cc3683e54ef987352c8987aeb1a8e13742bfd22f4b70f473a5c26d8a4a\": container with ID starting with 84e601cc3683e54ef987352c8987aeb1a8e13742bfd22f4b70f473a5c26d8a4a not found: ID does not exist" containerID="84e601cc3683e54ef987352c8987aeb1a8e13742bfd22f4b70f473a5c26d8a4a" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.537791 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84e601cc3683e54ef987352c8987aeb1a8e13742bfd22f4b70f473a5c26d8a4a"} err="failed to get container status \"84e601cc3683e54ef987352c8987aeb1a8e13742bfd22f4b70f473a5c26d8a4a\": rpc error: code = NotFound desc = could not find container \"84e601cc3683e54ef987352c8987aeb1a8e13742bfd22f4b70f473a5c26d8a4a\": container with ID starting with 84e601cc3683e54ef987352c8987aeb1a8e13742bfd22f4b70f473a5c26d8a4a not found: ID does not exist" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.537824 4775 scope.go:117] "RemoveContainer" containerID="ebe7fec64953fd11163fa5282af0a9cf3a6e1568a1469a787c9052026143b42f" Oct 02 01:45:44 crc kubenswrapper[4775]: E1002 01:45:44.538370 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebe7fec64953fd11163fa5282af0a9cf3a6e1568a1469a787c9052026143b42f\": container with ID starting with ebe7fec64953fd11163fa5282af0a9cf3a6e1568a1469a787c9052026143b42f not found: ID does not exist" containerID="ebe7fec64953fd11163fa5282af0a9cf3a6e1568a1469a787c9052026143b42f" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.538408 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebe7fec64953fd11163fa5282af0a9cf3a6e1568a1469a787c9052026143b42f"} err="failed to get container status \"ebe7fec64953fd11163fa5282af0a9cf3a6e1568a1469a787c9052026143b42f\": rpc error: code = NotFound desc = could not find container \"ebe7fec64953fd11163fa5282af0a9cf3a6e1568a1469a787c9052026143b42f\": container with ID starting with ebe7fec64953fd11163fa5282af0a9cf3a6e1568a1469a787c9052026143b42f not found: ID does not exist" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.538428 4775 scope.go:117] "RemoveContainer" containerID="ab4a78408ec0e7ab99fd20209e0a79003560537d24bee261cd65c21b986d4626" Oct 02 01:45:44 crc kubenswrapper[4775]: E1002 01:45:44.538889 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab4a78408ec0e7ab99fd20209e0a79003560537d24bee261cd65c21b986d4626\": container with ID starting with ab4a78408ec0e7ab99fd20209e0a79003560537d24bee261cd65c21b986d4626 not found: ID does not exist" containerID="ab4a78408ec0e7ab99fd20209e0a79003560537d24bee261cd65c21b986d4626" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.538925 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab4a78408ec0e7ab99fd20209e0a79003560537d24bee261cd65c21b986d4626"} err="failed to get container status \"ab4a78408ec0e7ab99fd20209e0a79003560537d24bee261cd65c21b986d4626\": rpc error: code = NotFound desc = could not find container \"ab4a78408ec0e7ab99fd20209e0a79003560537d24bee261cd65c21b986d4626\": container with ID starting with ab4a78408ec0e7ab99fd20209e0a79003560537d24bee261cd65c21b986d4626 not found: ID does not exist" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.538943 4775 scope.go:117] "RemoveContainer" containerID="868175f2791bd40663953c18e25df7b84e007a69401f15db9feeaf4a7a2f15fc" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.549607 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d8qrs"] Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.552746 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d8qrs"] Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.555277 4775 scope.go:117] "RemoveContainer" containerID="2f11b0584a76dab4307e3624ed58106b8ed4e7f3c2ca57957276a319f0690ca5" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.568019 4775 scope.go:117] "RemoveContainer" containerID="cf6f56dabc2f8834a410370468a3854d57aafe3f2ad97c0a6e6b71904f0227c8" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.589256 4775 scope.go:117] "RemoveContainer" containerID="868175f2791bd40663953c18e25df7b84e007a69401f15db9feeaf4a7a2f15fc" Oct 02 01:45:44 crc kubenswrapper[4775]: E1002 01:45:44.591415 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"868175f2791bd40663953c18e25df7b84e007a69401f15db9feeaf4a7a2f15fc\": container with ID starting with 868175f2791bd40663953c18e25df7b84e007a69401f15db9feeaf4a7a2f15fc not found: ID does not exist" containerID="868175f2791bd40663953c18e25df7b84e007a69401f15db9feeaf4a7a2f15fc" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.591454 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"868175f2791bd40663953c18e25df7b84e007a69401f15db9feeaf4a7a2f15fc"} err="failed to get container status \"868175f2791bd40663953c18e25df7b84e007a69401f15db9feeaf4a7a2f15fc\": rpc error: code = NotFound desc = could not find container \"868175f2791bd40663953c18e25df7b84e007a69401f15db9feeaf4a7a2f15fc\": container with ID starting with 868175f2791bd40663953c18e25df7b84e007a69401f15db9feeaf4a7a2f15fc not found: ID does not exist" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.591480 4775 scope.go:117] "RemoveContainer" containerID="2f11b0584a76dab4307e3624ed58106b8ed4e7f3c2ca57957276a319f0690ca5" Oct 02 01:45:44 crc kubenswrapper[4775]: E1002 01:45:44.592177 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f11b0584a76dab4307e3624ed58106b8ed4e7f3c2ca57957276a319f0690ca5\": container with ID starting with 2f11b0584a76dab4307e3624ed58106b8ed4e7f3c2ca57957276a319f0690ca5 not found: ID does not exist" containerID="2f11b0584a76dab4307e3624ed58106b8ed4e7f3c2ca57957276a319f0690ca5" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.592239 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f11b0584a76dab4307e3624ed58106b8ed4e7f3c2ca57957276a319f0690ca5"} err="failed to get container status \"2f11b0584a76dab4307e3624ed58106b8ed4e7f3c2ca57957276a319f0690ca5\": rpc error: code = NotFound desc = could not find container \"2f11b0584a76dab4307e3624ed58106b8ed4e7f3c2ca57957276a319f0690ca5\": container with ID starting with 2f11b0584a76dab4307e3624ed58106b8ed4e7f3c2ca57957276a319f0690ca5 not found: ID does not exist" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.592267 4775 scope.go:117] "RemoveContainer" containerID="cf6f56dabc2f8834a410370468a3854d57aafe3f2ad97c0a6e6b71904f0227c8" Oct 02 01:45:44 crc kubenswrapper[4775]: E1002 01:45:44.592580 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf6f56dabc2f8834a410370468a3854d57aafe3f2ad97c0a6e6b71904f0227c8\": container with ID starting with cf6f56dabc2f8834a410370468a3854d57aafe3f2ad97c0a6e6b71904f0227c8 not found: ID does not exist" containerID="cf6f56dabc2f8834a410370468a3854d57aafe3f2ad97c0a6e6b71904f0227c8" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.592682 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf6f56dabc2f8834a410370468a3854d57aafe3f2ad97c0a6e6b71904f0227c8"} err="failed to get container status \"cf6f56dabc2f8834a410370468a3854d57aafe3f2ad97c0a6e6b71904f0227c8\": rpc error: code = NotFound desc = could not find container \"cf6f56dabc2f8834a410370468a3854d57aafe3f2ad97c0a6e6b71904f0227c8\": container with ID starting with cf6f56dabc2f8834a410370468a3854d57aafe3f2ad97c0a6e6b71904f0227c8 not found: ID does not exist" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.592871 4775 scope.go:117] "RemoveContainer" containerID="b6581c6c4f123e1ef81dfec9083a405d3392019901bf2942712a4b738236fd9c" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.607845 4775 scope.go:117] "RemoveContainer" containerID="ea2c0f7560d7d3ec7004ec2cbc0e88db0ff7eae8adf65345e209ac67816aff1d" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.626239 4775 scope.go:117] "RemoveContainer" containerID="9c975c6f0575fcbd4a37bfc8d8f5c239ae0354fec3f031c733423ad259ed4eaf" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.642654 4775 scope.go:117] "RemoveContainer" containerID="b6581c6c4f123e1ef81dfec9083a405d3392019901bf2942712a4b738236fd9c" Oct 02 01:45:44 crc kubenswrapper[4775]: E1002 01:45:44.643798 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6581c6c4f123e1ef81dfec9083a405d3392019901bf2942712a4b738236fd9c\": container with ID starting with b6581c6c4f123e1ef81dfec9083a405d3392019901bf2942712a4b738236fd9c not found: ID does not exist" containerID="b6581c6c4f123e1ef81dfec9083a405d3392019901bf2942712a4b738236fd9c" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.643841 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6581c6c4f123e1ef81dfec9083a405d3392019901bf2942712a4b738236fd9c"} err="failed to get container status \"b6581c6c4f123e1ef81dfec9083a405d3392019901bf2942712a4b738236fd9c\": rpc error: code = NotFound desc = could not find container \"b6581c6c4f123e1ef81dfec9083a405d3392019901bf2942712a4b738236fd9c\": container with ID starting with b6581c6c4f123e1ef81dfec9083a405d3392019901bf2942712a4b738236fd9c not found: ID does not exist" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.643869 4775 scope.go:117] "RemoveContainer" containerID="ea2c0f7560d7d3ec7004ec2cbc0e88db0ff7eae8adf65345e209ac67816aff1d" Oct 02 01:45:44 crc kubenswrapper[4775]: E1002 01:45:44.644411 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea2c0f7560d7d3ec7004ec2cbc0e88db0ff7eae8adf65345e209ac67816aff1d\": container with ID starting with ea2c0f7560d7d3ec7004ec2cbc0e88db0ff7eae8adf65345e209ac67816aff1d not found: ID does not exist" containerID="ea2c0f7560d7d3ec7004ec2cbc0e88db0ff7eae8adf65345e209ac67816aff1d" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.644456 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea2c0f7560d7d3ec7004ec2cbc0e88db0ff7eae8adf65345e209ac67816aff1d"} err="failed to get container status \"ea2c0f7560d7d3ec7004ec2cbc0e88db0ff7eae8adf65345e209ac67816aff1d\": rpc error: code = NotFound desc = could not find container \"ea2c0f7560d7d3ec7004ec2cbc0e88db0ff7eae8adf65345e209ac67816aff1d\": container with ID starting with ea2c0f7560d7d3ec7004ec2cbc0e88db0ff7eae8adf65345e209ac67816aff1d not found: ID does not exist" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.644489 4775 scope.go:117] "RemoveContainer" containerID="9c975c6f0575fcbd4a37bfc8d8f5c239ae0354fec3f031c733423ad259ed4eaf" Oct 02 01:45:44 crc kubenswrapper[4775]: E1002 01:45:44.644981 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c975c6f0575fcbd4a37bfc8d8f5c239ae0354fec3f031c733423ad259ed4eaf\": container with ID starting with 9c975c6f0575fcbd4a37bfc8d8f5c239ae0354fec3f031c733423ad259ed4eaf not found: ID does not exist" containerID="9c975c6f0575fcbd4a37bfc8d8f5c239ae0354fec3f031c733423ad259ed4eaf" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.645011 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c975c6f0575fcbd4a37bfc8d8f5c239ae0354fec3f031c733423ad259ed4eaf"} err="failed to get container status \"9c975c6f0575fcbd4a37bfc8d8f5c239ae0354fec3f031c733423ad259ed4eaf\": rpc error: code = NotFound desc = could not find container \"9c975c6f0575fcbd4a37bfc8d8f5c239ae0354fec3f031c733423ad259ed4eaf\": container with ID starting with 9c975c6f0575fcbd4a37bfc8d8f5c239ae0354fec3f031c733423ad259ed4eaf not found: ID does not exist" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.645032 4775 scope.go:117] "RemoveContainer" containerID="c335259f9d23a5a1f5199708b3dddfb4a2e5fcaa12f0bf36c4c41679845b2865" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.661692 4775 scope.go:117] "RemoveContainer" containerID="c335259f9d23a5a1f5199708b3dddfb4a2e5fcaa12f0bf36c4c41679845b2865" Oct 02 01:45:44 crc kubenswrapper[4775]: E1002 01:45:44.667161 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c335259f9d23a5a1f5199708b3dddfb4a2e5fcaa12f0bf36c4c41679845b2865\": container with ID starting with c335259f9d23a5a1f5199708b3dddfb4a2e5fcaa12f0bf36c4c41679845b2865 not found: ID does not exist" containerID="c335259f9d23a5a1f5199708b3dddfb4a2e5fcaa12f0bf36c4c41679845b2865" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.667192 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c335259f9d23a5a1f5199708b3dddfb4a2e5fcaa12f0bf36c4c41679845b2865"} err="failed to get container status \"c335259f9d23a5a1f5199708b3dddfb4a2e5fcaa12f0bf36c4c41679845b2865\": rpc error: code = NotFound desc = could not find container \"c335259f9d23a5a1f5199708b3dddfb4a2e5fcaa12f0bf36c4c41679845b2865\": container with ID starting with c335259f9d23a5a1f5199708b3dddfb4a2e5fcaa12f0bf36c4c41679845b2865 not found: ID does not exist" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.667229 4775 scope.go:117] "RemoveContainer" containerID="3e174e58fc35e133b2319c734bdb722ea5821e22f2979a6b0e3fdc53b0c4b18b" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.680429 4775 scope.go:117] "RemoveContainer" containerID="2ec623203c86b2ad1073eee1eab7a31ee2a78303144ca39b8628cdd524c9b269" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.695705 4775 scope.go:117] "RemoveContainer" containerID="16c3739cae5327a8336493d7ff12700ee47b92f838f449a7ccb697d50bac1fbc" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.709850 4775 scope.go:117] "RemoveContainer" containerID="3e174e58fc35e133b2319c734bdb722ea5821e22f2979a6b0e3fdc53b0c4b18b" Oct 02 01:45:44 crc kubenswrapper[4775]: E1002 01:45:44.710205 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e174e58fc35e133b2319c734bdb722ea5821e22f2979a6b0e3fdc53b0c4b18b\": container with ID starting with 3e174e58fc35e133b2319c734bdb722ea5821e22f2979a6b0e3fdc53b0c4b18b not found: ID does not exist" containerID="3e174e58fc35e133b2319c734bdb722ea5821e22f2979a6b0e3fdc53b0c4b18b" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.710240 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e174e58fc35e133b2319c734bdb722ea5821e22f2979a6b0e3fdc53b0c4b18b"} err="failed to get container status \"3e174e58fc35e133b2319c734bdb722ea5821e22f2979a6b0e3fdc53b0c4b18b\": rpc error: code = NotFound desc = could not find container \"3e174e58fc35e133b2319c734bdb722ea5821e22f2979a6b0e3fdc53b0c4b18b\": container with ID starting with 3e174e58fc35e133b2319c734bdb722ea5821e22f2979a6b0e3fdc53b0c4b18b not found: ID does not exist" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.710264 4775 scope.go:117] "RemoveContainer" containerID="2ec623203c86b2ad1073eee1eab7a31ee2a78303144ca39b8628cdd524c9b269" Oct 02 01:45:44 crc kubenswrapper[4775]: E1002 01:45:44.710686 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ec623203c86b2ad1073eee1eab7a31ee2a78303144ca39b8628cdd524c9b269\": container with ID starting with 2ec623203c86b2ad1073eee1eab7a31ee2a78303144ca39b8628cdd524c9b269 not found: ID does not exist" containerID="2ec623203c86b2ad1073eee1eab7a31ee2a78303144ca39b8628cdd524c9b269" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.710710 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ec623203c86b2ad1073eee1eab7a31ee2a78303144ca39b8628cdd524c9b269"} err="failed to get container status \"2ec623203c86b2ad1073eee1eab7a31ee2a78303144ca39b8628cdd524c9b269\": rpc error: code = NotFound desc = could not find container \"2ec623203c86b2ad1073eee1eab7a31ee2a78303144ca39b8628cdd524c9b269\": container with ID starting with 2ec623203c86b2ad1073eee1eab7a31ee2a78303144ca39b8628cdd524c9b269 not found: ID does not exist" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.710729 4775 scope.go:117] "RemoveContainer" containerID="16c3739cae5327a8336493d7ff12700ee47b92f838f449a7ccb697d50bac1fbc" Oct 02 01:45:44 crc kubenswrapper[4775]: E1002 01:45:44.710977 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16c3739cae5327a8336493d7ff12700ee47b92f838f449a7ccb697d50bac1fbc\": container with ID starting with 16c3739cae5327a8336493d7ff12700ee47b92f838f449a7ccb697d50bac1fbc not found: ID does not exist" containerID="16c3739cae5327a8336493d7ff12700ee47b92f838f449a7ccb697d50bac1fbc" Oct 02 01:45:44 crc kubenswrapper[4775]: I1002 01:45:44.711008 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16c3739cae5327a8336493d7ff12700ee47b92f838f449a7ccb697d50bac1fbc"} err="failed to get container status \"16c3739cae5327a8336493d7ff12700ee47b92f838f449a7ccb697d50bac1fbc\": rpc error: code = NotFound desc = could not find container \"16c3739cae5327a8336493d7ff12700ee47b92f838f449a7ccb697d50bac1fbc\": container with ID starting with 16c3739cae5327a8336493d7ff12700ee47b92f838f449a7ccb697d50bac1fbc not found: ID does not exist" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.474406 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-n4prl" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.553266 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mpj5m"] Oct 02 01:45:45 crc kubenswrapper[4775]: E1002 01:45:45.553456 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ccbabf-0c44-467d-812c-79477c19e012" containerName="extract-utilities" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.553471 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ccbabf-0c44-467d-812c-79477c19e012" containerName="extract-utilities" Oct 02 01:45:45 crc kubenswrapper[4775]: E1002 01:45:45.553482 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d1808de-113f-4223-99eb-0f72d7a1b8de" containerName="extract-content" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.553488 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d1808de-113f-4223-99eb-0f72d7a1b8de" containerName="extract-content" Oct 02 01:45:45 crc kubenswrapper[4775]: E1002 01:45:45.553498 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d1808de-113f-4223-99eb-0f72d7a1b8de" containerName="extract-utilities" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.553504 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d1808de-113f-4223-99eb-0f72d7a1b8de" containerName="extract-utilities" Oct 02 01:45:45 crc kubenswrapper[4775]: E1002 01:45:45.553512 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d1808de-113f-4223-99eb-0f72d7a1b8de" containerName="registry-server" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.553517 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d1808de-113f-4223-99eb-0f72d7a1b8de" containerName="registry-server" Oct 02 01:45:45 crc kubenswrapper[4775]: E1002 01:45:45.553527 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c13d75ef-3ed7-46ea-a2c6-879f1427ac41" containerName="extract-utilities" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.553532 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c13d75ef-3ed7-46ea-a2c6-879f1427ac41" containerName="extract-utilities" Oct 02 01:45:45 crc kubenswrapper[4775]: E1002 01:45:45.553540 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ccbabf-0c44-467d-812c-79477c19e012" containerName="registry-server" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.553547 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ccbabf-0c44-467d-812c-79477c19e012" containerName="registry-server" Oct 02 01:45:45 crc kubenswrapper[4775]: E1002 01:45:45.553557 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" containerName="extract-utilities" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.553563 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" containerName="extract-utilities" Oct 02 01:45:45 crc kubenswrapper[4775]: E1002 01:45:45.553571 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e1b3210-4726-450d-8da5-766903e6d91b" containerName="marketplace-operator" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.553577 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e1b3210-4726-450d-8da5-766903e6d91b" containerName="marketplace-operator" Oct 02 01:45:45 crc kubenswrapper[4775]: E1002 01:45:45.553585 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c13d75ef-3ed7-46ea-a2c6-879f1427ac41" containerName="registry-server" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.553593 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c13d75ef-3ed7-46ea-a2c6-879f1427ac41" containerName="registry-server" Oct 02 01:45:45 crc kubenswrapper[4775]: E1002 01:45:45.553601 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" containerName="extract-content" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.553607 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" containerName="extract-content" Oct 02 01:45:45 crc kubenswrapper[4775]: E1002 01:45:45.553618 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ccbabf-0c44-467d-812c-79477c19e012" containerName="extract-content" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.553624 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ccbabf-0c44-467d-812c-79477c19e012" containerName="extract-content" Oct 02 01:45:45 crc kubenswrapper[4775]: E1002 01:45:45.554631 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c13d75ef-3ed7-46ea-a2c6-879f1427ac41" containerName="extract-content" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.554644 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c13d75ef-3ed7-46ea-a2c6-879f1427ac41" containerName="extract-content" Oct 02 01:45:45 crc kubenswrapper[4775]: E1002 01:45:45.554651 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" containerName="registry-server" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.554657 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" containerName="registry-server" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.554738 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d1808de-113f-4223-99eb-0f72d7a1b8de" containerName="registry-server" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.554747 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ccbabf-0c44-467d-812c-79477c19e012" containerName="registry-server" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.554759 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e1b3210-4726-450d-8da5-766903e6d91b" containerName="marketplace-operator" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.554766 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c13d75ef-3ed7-46ea-a2c6-879f1427ac41" containerName="registry-server" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.554776 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" containerName="registry-server" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.555649 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.558945 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.581626 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpj5m"] Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.692059 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwwvk\" (UniqueName: \"kubernetes.io/projected/83400c9e-1f98-434e-a937-9acb45d8a61b-kube-api-access-xwwvk\") pod \"redhat-marketplace-mpj5m\" (UID: \"83400c9e-1f98-434e-a937-9acb45d8a61b\") " pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.692123 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83400c9e-1f98-434e-a937-9acb45d8a61b-utilities\") pod \"redhat-marketplace-mpj5m\" (UID: \"83400c9e-1f98-434e-a937-9acb45d8a61b\") " pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.692152 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83400c9e-1f98-434e-a937-9acb45d8a61b-catalog-content\") pod \"redhat-marketplace-mpj5m\" (UID: \"83400c9e-1f98-434e-a937-9acb45d8a61b\") " pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.734720 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ltrcf"] Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.735595 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.738235 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.754306 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ltrcf"] Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.777851 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f666cea-2dd7-4e5e-b5a2-d3a2120535d0" path="/var/lib/kubelet/pods/0f666cea-2dd7-4e5e-b5a2-d3a2120535d0/volumes" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.779177 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d1808de-113f-4223-99eb-0f72d7a1b8de" path="/var/lib/kubelet/pods/2d1808de-113f-4223-99eb-0f72d7a1b8de/volumes" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.780948 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e1b3210-4726-450d-8da5-766903e6d91b" path="/var/lib/kubelet/pods/2e1b3210-4726-450d-8da5-766903e6d91b/volumes" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.783411 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c13d75ef-3ed7-46ea-a2c6-879f1427ac41" path="/var/lib/kubelet/pods/c13d75ef-3ed7-46ea-a2c6-879f1427ac41/volumes" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.784554 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8ccbabf-0c44-467d-812c-79477c19e012" path="/var/lib/kubelet/pods/c8ccbabf-0c44-467d-812c-79477c19e012/volumes" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.792741 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83400c9e-1f98-434e-a937-9acb45d8a61b-utilities\") pod \"redhat-marketplace-mpj5m\" (UID: \"83400c9e-1f98-434e-a937-9acb45d8a61b\") " pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.792783 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83400c9e-1f98-434e-a937-9acb45d8a61b-catalog-content\") pod \"redhat-marketplace-mpj5m\" (UID: \"83400c9e-1f98-434e-a937-9acb45d8a61b\") " pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.792836 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a17388ea-bfd3-40b8-8b69-984e95eec731-utilities\") pod \"certified-operators-ltrcf\" (UID: \"a17388ea-bfd3-40b8-8b69-984e95eec731\") " pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.792861 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a17388ea-bfd3-40b8-8b69-984e95eec731-catalog-content\") pod \"certified-operators-ltrcf\" (UID: \"a17388ea-bfd3-40b8-8b69-984e95eec731\") " pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.792882 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrggs\" (UniqueName: \"kubernetes.io/projected/a17388ea-bfd3-40b8-8b69-984e95eec731-kube-api-access-mrggs\") pod \"certified-operators-ltrcf\" (UID: \"a17388ea-bfd3-40b8-8b69-984e95eec731\") " pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.792904 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwwvk\" (UniqueName: \"kubernetes.io/projected/83400c9e-1f98-434e-a937-9acb45d8a61b-kube-api-access-xwwvk\") pod \"redhat-marketplace-mpj5m\" (UID: \"83400c9e-1f98-434e-a937-9acb45d8a61b\") " pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.793251 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83400c9e-1f98-434e-a937-9acb45d8a61b-catalog-content\") pod \"redhat-marketplace-mpj5m\" (UID: \"83400c9e-1f98-434e-a937-9acb45d8a61b\") " pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.793247 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83400c9e-1f98-434e-a937-9acb45d8a61b-utilities\") pod \"redhat-marketplace-mpj5m\" (UID: \"83400c9e-1f98-434e-a937-9acb45d8a61b\") " pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.810411 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwwvk\" (UniqueName: \"kubernetes.io/projected/83400c9e-1f98-434e-a937-9acb45d8a61b-kube-api-access-xwwvk\") pod \"redhat-marketplace-mpj5m\" (UID: \"83400c9e-1f98-434e-a937-9acb45d8a61b\") " pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.885921 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.893862 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a17388ea-bfd3-40b8-8b69-984e95eec731-utilities\") pod \"certified-operators-ltrcf\" (UID: \"a17388ea-bfd3-40b8-8b69-984e95eec731\") " pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.893908 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a17388ea-bfd3-40b8-8b69-984e95eec731-catalog-content\") pod \"certified-operators-ltrcf\" (UID: \"a17388ea-bfd3-40b8-8b69-984e95eec731\") " pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.893936 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrggs\" (UniqueName: \"kubernetes.io/projected/a17388ea-bfd3-40b8-8b69-984e95eec731-kube-api-access-mrggs\") pod \"certified-operators-ltrcf\" (UID: \"a17388ea-bfd3-40b8-8b69-984e95eec731\") " pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.894640 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a17388ea-bfd3-40b8-8b69-984e95eec731-utilities\") pod \"certified-operators-ltrcf\" (UID: \"a17388ea-bfd3-40b8-8b69-984e95eec731\") " pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.894760 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a17388ea-bfd3-40b8-8b69-984e95eec731-catalog-content\") pod \"certified-operators-ltrcf\" (UID: \"a17388ea-bfd3-40b8-8b69-984e95eec731\") " pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:45 crc kubenswrapper[4775]: I1002 01:45:45.909531 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrggs\" (UniqueName: \"kubernetes.io/projected/a17388ea-bfd3-40b8-8b69-984e95eec731-kube-api-access-mrggs\") pod \"certified-operators-ltrcf\" (UID: \"a17388ea-bfd3-40b8-8b69-984e95eec731\") " pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:46 crc kubenswrapper[4775]: I1002 01:45:46.061638 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:46 crc kubenswrapper[4775]: I1002 01:45:46.077414 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mpj5m"] Oct 02 01:45:46 crc kubenswrapper[4775]: W1002 01:45:46.086356 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83400c9e_1f98_434e_a937_9acb45d8a61b.slice/crio-978b50f5c690fceda55f2632da7884e0127711cac22ca7b3b2084e1ed55f87ae WatchSource:0}: Error finding container 978b50f5c690fceda55f2632da7884e0127711cac22ca7b3b2084e1ed55f87ae: Status 404 returned error can't find the container with id 978b50f5c690fceda55f2632da7884e0127711cac22ca7b3b2084e1ed55f87ae Oct 02 01:45:46 crc kubenswrapper[4775]: I1002 01:45:46.251412 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ltrcf"] Oct 02 01:45:46 crc kubenswrapper[4775]: W1002 01:45:46.287478 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda17388ea_bfd3_40b8_8b69_984e95eec731.slice/crio-6f75c6e406b67b91afdbb474c5b626745867fd92b3ab6d61ac3563da18bf4925 WatchSource:0}: Error finding container 6f75c6e406b67b91afdbb474c5b626745867fd92b3ab6d61ac3563da18bf4925: Status 404 returned error can't find the container with id 6f75c6e406b67b91afdbb474c5b626745867fd92b3ab6d61ac3563da18bf4925 Oct 02 01:45:46 crc kubenswrapper[4775]: I1002 01:45:46.477669 4775 generic.go:334] "Generic (PLEG): container finished" podID="a17388ea-bfd3-40b8-8b69-984e95eec731" containerID="78f882cc647a65a6f7a5a405d34e16b7c793fe47f59ba5cc2d0c90831ea804fc" exitCode=0 Oct 02 01:45:46 crc kubenswrapper[4775]: I1002 01:45:46.477717 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltrcf" event={"ID":"a17388ea-bfd3-40b8-8b69-984e95eec731","Type":"ContainerDied","Data":"78f882cc647a65a6f7a5a405d34e16b7c793fe47f59ba5cc2d0c90831ea804fc"} Oct 02 01:45:46 crc kubenswrapper[4775]: I1002 01:45:46.477756 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltrcf" event={"ID":"a17388ea-bfd3-40b8-8b69-984e95eec731","Type":"ContainerStarted","Data":"6f75c6e406b67b91afdbb474c5b626745867fd92b3ab6d61ac3563da18bf4925"} Oct 02 01:45:46 crc kubenswrapper[4775]: I1002 01:45:46.479918 4775 generic.go:334] "Generic (PLEG): container finished" podID="83400c9e-1f98-434e-a937-9acb45d8a61b" containerID="b1a37a5f03ce89c850fb286092fa887d4030b972c43af134317a72ed79922f53" exitCode=0 Oct 02 01:45:46 crc kubenswrapper[4775]: I1002 01:45:46.480602 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpj5m" event={"ID":"83400c9e-1f98-434e-a937-9acb45d8a61b","Type":"ContainerDied","Data":"b1a37a5f03ce89c850fb286092fa887d4030b972c43af134317a72ed79922f53"} Oct 02 01:45:46 crc kubenswrapper[4775]: I1002 01:45:46.480617 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpj5m" event={"ID":"83400c9e-1f98-434e-a937-9acb45d8a61b","Type":"ContainerStarted","Data":"978b50f5c690fceda55f2632da7884e0127711cac22ca7b3b2084e1ed55f87ae"} Oct 02 01:45:47 crc kubenswrapper[4775]: I1002 01:45:47.493147 4775 generic.go:334] "Generic (PLEG): container finished" podID="a17388ea-bfd3-40b8-8b69-984e95eec731" containerID="d4f66dc1f3de7296dffc127b8d68f12477fe3dc5d472b4e670bc5ee76b356c90" exitCode=0 Oct 02 01:45:47 crc kubenswrapper[4775]: I1002 01:45:47.493220 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltrcf" event={"ID":"a17388ea-bfd3-40b8-8b69-984e95eec731","Type":"ContainerDied","Data":"d4f66dc1f3de7296dffc127b8d68f12477fe3dc5d472b4e670bc5ee76b356c90"} Oct 02 01:45:47 crc kubenswrapper[4775]: I1002 01:45:47.502366 4775 generic.go:334] "Generic (PLEG): container finished" podID="83400c9e-1f98-434e-a937-9acb45d8a61b" containerID="17c46c39cc20c892f4c79f8b8a655728fd63699d5eae875719d67e845165d0c4" exitCode=0 Oct 02 01:45:47 crc kubenswrapper[4775]: I1002 01:45:47.502424 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpj5m" event={"ID":"83400c9e-1f98-434e-a937-9acb45d8a61b","Type":"ContainerDied","Data":"17c46c39cc20c892f4c79f8b8a655728fd63699d5eae875719d67e845165d0c4"} Oct 02 01:45:47 crc kubenswrapper[4775]: I1002 01:45:47.936078 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-sdnvn"] Oct 02 01:45:47 crc kubenswrapper[4775]: I1002 01:45:47.937360 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:47 crc kubenswrapper[4775]: I1002 01:45:47.940217 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 02 01:45:47 crc kubenswrapper[4775]: I1002 01:45:47.950412 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdnvn"] Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.028713 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/763784f6-4d43-4db9-88cb-0ba8f213f78e-utilities\") pod \"redhat-operators-sdnvn\" (UID: \"763784f6-4d43-4db9-88cb-0ba8f213f78e\") " pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.028780 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjx5d\" (UniqueName: \"kubernetes.io/projected/763784f6-4d43-4db9-88cb-0ba8f213f78e-kube-api-access-mjx5d\") pod \"redhat-operators-sdnvn\" (UID: \"763784f6-4d43-4db9-88cb-0ba8f213f78e\") " pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.028801 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/763784f6-4d43-4db9-88cb-0ba8f213f78e-catalog-content\") pod \"redhat-operators-sdnvn\" (UID: \"763784f6-4d43-4db9-88cb-0ba8f213f78e\") " pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.129501 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-74hz2"] Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.130161 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/763784f6-4d43-4db9-88cb-0ba8f213f78e-utilities\") pod \"redhat-operators-sdnvn\" (UID: \"763784f6-4d43-4db9-88cb-0ba8f213f78e\") " pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.130274 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjx5d\" (UniqueName: \"kubernetes.io/projected/763784f6-4d43-4db9-88cb-0ba8f213f78e-kube-api-access-mjx5d\") pod \"redhat-operators-sdnvn\" (UID: \"763784f6-4d43-4db9-88cb-0ba8f213f78e\") " pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.130309 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/763784f6-4d43-4db9-88cb-0ba8f213f78e-catalog-content\") pod \"redhat-operators-sdnvn\" (UID: \"763784f6-4d43-4db9-88cb-0ba8f213f78e\") " pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.130407 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.131121 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/763784f6-4d43-4db9-88cb-0ba8f213f78e-catalog-content\") pod \"redhat-operators-sdnvn\" (UID: \"763784f6-4d43-4db9-88cb-0ba8f213f78e\") " pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.131397 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/763784f6-4d43-4db9-88cb-0ba8f213f78e-utilities\") pod \"redhat-operators-sdnvn\" (UID: \"763784f6-4d43-4db9-88cb-0ba8f213f78e\") " pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.133014 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.155063 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-74hz2"] Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.163686 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjx5d\" (UniqueName: \"kubernetes.io/projected/763784f6-4d43-4db9-88cb-0ba8f213f78e-kube-api-access-mjx5d\") pod \"redhat-operators-sdnvn\" (UID: \"763784f6-4d43-4db9-88cb-0ba8f213f78e\") " pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.230826 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02a708a6-f32e-45b7-8989-4983ee6ada6c-utilities\") pod \"community-operators-74hz2\" (UID: \"02a708a6-f32e-45b7-8989-4983ee6ada6c\") " pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.230888 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02a708a6-f32e-45b7-8989-4983ee6ada6c-catalog-content\") pod \"community-operators-74hz2\" (UID: \"02a708a6-f32e-45b7-8989-4983ee6ada6c\") " pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.230911 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pjd8\" (UniqueName: \"kubernetes.io/projected/02a708a6-f32e-45b7-8989-4983ee6ada6c-kube-api-access-9pjd8\") pod \"community-operators-74hz2\" (UID: \"02a708a6-f32e-45b7-8989-4983ee6ada6c\") " pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.332275 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02a708a6-f32e-45b7-8989-4983ee6ada6c-utilities\") pod \"community-operators-74hz2\" (UID: \"02a708a6-f32e-45b7-8989-4983ee6ada6c\") " pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.332343 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02a708a6-f32e-45b7-8989-4983ee6ada6c-catalog-content\") pod \"community-operators-74hz2\" (UID: \"02a708a6-f32e-45b7-8989-4983ee6ada6c\") " pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.332367 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pjd8\" (UniqueName: \"kubernetes.io/projected/02a708a6-f32e-45b7-8989-4983ee6ada6c-kube-api-access-9pjd8\") pod \"community-operators-74hz2\" (UID: \"02a708a6-f32e-45b7-8989-4983ee6ada6c\") " pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.332848 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/02a708a6-f32e-45b7-8989-4983ee6ada6c-utilities\") pod \"community-operators-74hz2\" (UID: \"02a708a6-f32e-45b7-8989-4983ee6ada6c\") " pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.333063 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/02a708a6-f32e-45b7-8989-4983ee6ada6c-catalog-content\") pod \"community-operators-74hz2\" (UID: \"02a708a6-f32e-45b7-8989-4983ee6ada6c\") " pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.344100 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.349807 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pjd8\" (UniqueName: \"kubernetes.io/projected/02a708a6-f32e-45b7-8989-4983ee6ada6c-kube-api-access-9pjd8\") pod \"community-operators-74hz2\" (UID: \"02a708a6-f32e-45b7-8989-4983ee6ada6c\") " pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.452848 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.509259 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ltrcf" event={"ID":"a17388ea-bfd3-40b8-8b69-984e95eec731","Type":"ContainerStarted","Data":"2a5eb4c35671d519396954ca3a845caaa8669889f3e8a646648a091c72a5dfc3"} Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.511533 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mpj5m" event={"ID":"83400c9e-1f98-434e-a937-9acb45d8a61b","Type":"ContainerStarted","Data":"4f858af565feb44fffd990b96249bb74397c692fef1c452e97d4525107994ca9"} Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.530649 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ltrcf" podStartSLOduration=2.1159004 podStartE2EDuration="3.530630398s" podCreationTimestamp="2025-10-02 01:45:45 +0000 UTC" firstStartedPulling="2025-10-02 01:45:46.479079284 +0000 UTC m=+283.645823324" lastFinishedPulling="2025-10-02 01:45:47.893809252 +0000 UTC m=+285.060553322" observedRunningTime="2025-10-02 01:45:48.527607148 +0000 UTC m=+285.694351198" watchObservedRunningTime="2025-10-02 01:45:48.530630398 +0000 UTC m=+285.697374438" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.545151 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mpj5m" podStartSLOduration=2.072557611 podStartE2EDuration="3.545136843s" podCreationTimestamp="2025-10-02 01:45:45 +0000 UTC" firstStartedPulling="2025-10-02 01:45:46.481233821 +0000 UTC m=+283.647977861" lastFinishedPulling="2025-10-02 01:45:47.953813053 +0000 UTC m=+285.120557093" observedRunningTime="2025-10-02 01:45:48.544378713 +0000 UTC m=+285.711122753" watchObservedRunningTime="2025-10-02 01:45:48.545136843 +0000 UTC m=+285.711880883" Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.663104 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-74hz2"] Oct 02 01:45:48 crc kubenswrapper[4775]: I1002 01:45:48.752208 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-sdnvn"] Oct 02 01:45:49 crc kubenswrapper[4775]: I1002 01:45:49.519862 4775 generic.go:334] "Generic (PLEG): container finished" podID="02a708a6-f32e-45b7-8989-4983ee6ada6c" containerID="959f71dc84a0beef6a402cf3ad9c445749411687545890a054a0787aaf179a1f" exitCode=0 Oct 02 01:45:49 crc kubenswrapper[4775]: I1002 01:45:49.520012 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-74hz2" event={"ID":"02a708a6-f32e-45b7-8989-4983ee6ada6c","Type":"ContainerDied","Data":"959f71dc84a0beef6a402cf3ad9c445749411687545890a054a0787aaf179a1f"} Oct 02 01:45:49 crc kubenswrapper[4775]: I1002 01:45:49.520367 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-74hz2" event={"ID":"02a708a6-f32e-45b7-8989-4983ee6ada6c","Type":"ContainerStarted","Data":"e5b04c5d6b2cb0b2cca1777d134b56e1e4de4d3af2645fc423b824803cd87e69"} Oct 02 01:45:49 crc kubenswrapper[4775]: I1002 01:45:49.522087 4775 generic.go:334] "Generic (PLEG): container finished" podID="763784f6-4d43-4db9-88cb-0ba8f213f78e" containerID="981caa6ee5fbd323fcb575e1d7cb1d04e2a0d623ed1a901c959491fd6ac0b640" exitCode=0 Oct 02 01:45:49 crc kubenswrapper[4775]: I1002 01:45:49.522214 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdnvn" event={"ID":"763784f6-4d43-4db9-88cb-0ba8f213f78e","Type":"ContainerDied","Data":"981caa6ee5fbd323fcb575e1d7cb1d04e2a0d623ed1a901c959491fd6ac0b640"} Oct 02 01:45:49 crc kubenswrapper[4775]: I1002 01:45:49.522262 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdnvn" event={"ID":"763784f6-4d43-4db9-88cb-0ba8f213f78e","Type":"ContainerStarted","Data":"dfb68b319ca37cedac9c57ab93c503c02081b75940e218227e49ba72498e4fc1"} Oct 02 01:45:50 crc kubenswrapper[4775]: I1002 01:45:50.530332 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdnvn" event={"ID":"763784f6-4d43-4db9-88cb-0ba8f213f78e","Type":"ContainerStarted","Data":"89df2e35a550473ad29e792293f0c4430779041f7c23071f6d5581ce5dd5e5bc"} Oct 02 01:45:50 crc kubenswrapper[4775]: I1002 01:45:50.533350 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-74hz2" event={"ID":"02a708a6-f32e-45b7-8989-4983ee6ada6c","Type":"ContainerStarted","Data":"f5f57c274e0d4830a88222451c9d536cdf35edfd69ca30438bba52fe1c6ed8b7"} Oct 02 01:45:51 crc kubenswrapper[4775]: I1002 01:45:51.540659 4775 generic.go:334] "Generic (PLEG): container finished" podID="763784f6-4d43-4db9-88cb-0ba8f213f78e" containerID="89df2e35a550473ad29e792293f0c4430779041f7c23071f6d5581ce5dd5e5bc" exitCode=0 Oct 02 01:45:51 crc kubenswrapper[4775]: I1002 01:45:51.540744 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdnvn" event={"ID":"763784f6-4d43-4db9-88cb-0ba8f213f78e","Type":"ContainerDied","Data":"89df2e35a550473ad29e792293f0c4430779041f7c23071f6d5581ce5dd5e5bc"} Oct 02 01:45:51 crc kubenswrapper[4775]: I1002 01:45:51.543870 4775 generic.go:334] "Generic (PLEG): container finished" podID="02a708a6-f32e-45b7-8989-4983ee6ada6c" containerID="f5f57c274e0d4830a88222451c9d536cdf35edfd69ca30438bba52fe1c6ed8b7" exitCode=0 Oct 02 01:45:51 crc kubenswrapper[4775]: I1002 01:45:51.543913 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-74hz2" event={"ID":"02a708a6-f32e-45b7-8989-4983ee6ada6c","Type":"ContainerDied","Data":"f5f57c274e0d4830a88222451c9d536cdf35edfd69ca30438bba52fe1c6ed8b7"} Oct 02 01:45:52 crc kubenswrapper[4775]: I1002 01:45:52.551332 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-74hz2" event={"ID":"02a708a6-f32e-45b7-8989-4983ee6ada6c","Type":"ContainerStarted","Data":"fb5056f9a8760a780f6610283ca216e14400be5568e8b268d46b41f9206ccc7b"} Oct 02 01:45:52 crc kubenswrapper[4775]: I1002 01:45:52.552767 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-sdnvn" event={"ID":"763784f6-4d43-4db9-88cb-0ba8f213f78e","Type":"ContainerStarted","Data":"5e4de33c583aad372edc284e8e720110498d311af06a287599260b4af218f65f"} Oct 02 01:45:52 crc kubenswrapper[4775]: I1002 01:45:52.587921 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-sdnvn" podStartSLOduration=2.808468626 podStartE2EDuration="5.587894382s" podCreationTimestamp="2025-10-02 01:45:47 +0000 UTC" firstStartedPulling="2025-10-02 01:45:49.523999532 +0000 UTC m=+286.690743572" lastFinishedPulling="2025-10-02 01:45:52.303425288 +0000 UTC m=+289.470169328" observedRunningTime="2025-10-02 01:45:52.586532946 +0000 UTC m=+289.753276996" watchObservedRunningTime="2025-10-02 01:45:52.587894382 +0000 UTC m=+289.754638462" Oct 02 01:45:52 crc kubenswrapper[4775]: I1002 01:45:52.589359 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-74hz2" podStartSLOduration=2.150318092 podStartE2EDuration="4.589350921s" podCreationTimestamp="2025-10-02 01:45:48 +0000 UTC" firstStartedPulling="2025-10-02 01:45:49.521811924 +0000 UTC m=+286.688556004" lastFinishedPulling="2025-10-02 01:45:51.960844793 +0000 UTC m=+289.127588833" observedRunningTime="2025-10-02 01:45:52.57008856 +0000 UTC m=+289.736832610" watchObservedRunningTime="2025-10-02 01:45:52.589350921 +0000 UTC m=+289.756095001" Oct 02 01:45:55 crc kubenswrapper[4775]: I1002 01:45:55.886807 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:55 crc kubenswrapper[4775]: I1002 01:45:55.887539 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:55 crc kubenswrapper[4775]: I1002 01:45:55.937152 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:56 crc kubenswrapper[4775]: I1002 01:45:56.062568 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:56 crc kubenswrapper[4775]: I1002 01:45:56.063196 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:56 crc kubenswrapper[4775]: I1002 01:45:56.112852 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:56 crc kubenswrapper[4775]: I1002 01:45:56.617261 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ltrcf" Oct 02 01:45:56 crc kubenswrapper[4775]: I1002 01:45:56.652280 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mpj5m" Oct 02 01:45:58 crc kubenswrapper[4775]: I1002 01:45:58.344527 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:58 crc kubenswrapper[4775]: I1002 01:45:58.344907 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:58 crc kubenswrapper[4775]: I1002 01:45:58.403735 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:58 crc kubenswrapper[4775]: I1002 01:45:58.453655 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:45:58 crc kubenswrapper[4775]: I1002 01:45:58.453763 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:45:58 crc kubenswrapper[4775]: I1002 01:45:58.503431 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:45:58 crc kubenswrapper[4775]: I1002 01:45:58.649674 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-sdnvn" Oct 02 01:45:58 crc kubenswrapper[4775]: I1002 01:45:58.651488 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-74hz2" Oct 02 01:47:07 crc kubenswrapper[4775]: I1002 01:47:07.233802 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:47:07 crc kubenswrapper[4775]: I1002 01:47:07.234422 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:47:37 crc kubenswrapper[4775]: I1002 01:47:37.233883 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:47:37 crc kubenswrapper[4775]: I1002 01:47:37.234554 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:48:07 crc kubenswrapper[4775]: I1002 01:48:07.233750 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:48:07 crc kubenswrapper[4775]: I1002 01:48:07.234491 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:48:07 crc kubenswrapper[4775]: I1002 01:48:07.234555 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:48:07 crc kubenswrapper[4775]: I1002 01:48:07.235352 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e819a457bcc1b0b56d9d1d4506dfc8bb5cf3853dae381f750b921e36982f1712"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 01:48:07 crc kubenswrapper[4775]: I1002 01:48:07.235445 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://e819a457bcc1b0b56d9d1d4506dfc8bb5cf3853dae381f750b921e36982f1712" gracePeriod=600 Oct 02 01:48:07 crc kubenswrapper[4775]: I1002 01:48:07.417030 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="e819a457bcc1b0b56d9d1d4506dfc8bb5cf3853dae381f750b921e36982f1712" exitCode=0 Oct 02 01:48:07 crc kubenswrapper[4775]: I1002 01:48:07.417124 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"e819a457bcc1b0b56d9d1d4506dfc8bb5cf3853dae381f750b921e36982f1712"} Oct 02 01:48:07 crc kubenswrapper[4775]: I1002 01:48:07.417227 4775 scope.go:117] "RemoveContainer" containerID="d3ca8a8142143bdb16991dad0967d11d058f583de4d9489cf2c7a0dd466b13cb" Oct 02 01:48:08 crc kubenswrapper[4775]: I1002 01:48:08.428105 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"2c48cbf3d08cb4a2f350c67fef84e9504d7f8966e661be9b0356695ca1016ee7"} Oct 02 01:48:32 crc kubenswrapper[4775]: I1002 01:48:32.878142 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-d9trj"] Oct 02 01:48:32 crc kubenswrapper[4775]: I1002 01:48:32.879358 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:32 crc kubenswrapper[4775]: I1002 01:48:32.894786 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-d9trj"] Oct 02 01:48:32 crc kubenswrapper[4775]: I1002 01:48:32.963388 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:32 crc kubenswrapper[4775]: I1002 01:48:32.963469 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:32 crc kubenswrapper[4775]: I1002 01:48:32.963530 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-bound-sa-token\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:32 crc kubenswrapper[4775]: I1002 01:48:32.963556 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-trusted-ca\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:32 crc kubenswrapper[4775]: I1002 01:48:32.963584 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-registry-tls\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:32 crc kubenswrapper[4775]: I1002 01:48:32.963607 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-registry-certificates\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:32 crc kubenswrapper[4775]: I1002 01:48:32.963634 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:32 crc kubenswrapper[4775]: I1002 01:48:32.963663 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czgjg\" (UniqueName: \"kubernetes.io/projected/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-kube-api-access-czgjg\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:32 crc kubenswrapper[4775]: I1002 01:48:32.993908 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.064487 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-bound-sa-token\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.064791 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-trusted-ca\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.064947 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-registry-tls\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.065121 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-registry-certificates\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.065244 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.065394 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czgjg\" (UniqueName: \"kubernetes.io/projected/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-kube-api-access-czgjg\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.065524 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.066145 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.067052 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-trusted-ca\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.067199 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-registry-certificates\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.072337 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-registry-tls\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.086353 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czgjg\" (UniqueName: \"kubernetes.io/projected/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-kube-api-access-czgjg\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.086981 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.092953 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/de29ac67-49dc-4d3b-86f2-9f6aba2f438c-bound-sa-token\") pod \"image-registry-66df7c8f76-d9trj\" (UID: \"de29ac67-49dc-4d3b-86f2-9f6aba2f438c\") " pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.247654 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.521515 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-d9trj"] Oct 02 01:48:33 crc kubenswrapper[4775]: I1002 01:48:33.621498 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" event={"ID":"de29ac67-49dc-4d3b-86f2-9f6aba2f438c","Type":"ContainerStarted","Data":"036e3150206bd2a73305697abd3be5aa2bf3a26837314993528e0207212f8302"} Oct 02 01:48:34 crc kubenswrapper[4775]: I1002 01:48:34.631841 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" event={"ID":"de29ac67-49dc-4d3b-86f2-9f6aba2f438c","Type":"ContainerStarted","Data":"29944d2b9d4f602895e0925b9f991ca48c4de8bc320b2752fe5c76fe77898bda"} Oct 02 01:48:34 crc kubenswrapper[4775]: I1002 01:48:34.632329 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:34 crc kubenswrapper[4775]: I1002 01:48:34.660544 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" podStartSLOduration=2.660513411 podStartE2EDuration="2.660513411s" podCreationTimestamp="2025-10-02 01:48:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:48:34.658268863 +0000 UTC m=+451.825012943" watchObservedRunningTime="2025-10-02 01:48:34.660513411 +0000 UTC m=+451.827258491" Oct 02 01:48:53 crc kubenswrapper[4775]: I1002 01:48:53.258598 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-d9trj" Oct 02 01:48:53 crc kubenswrapper[4775]: I1002 01:48:53.353721 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nv4tc"] Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.408298 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" podUID="adc13fc6-a2d8-44fd-ab25-075857884cad" containerName="registry" containerID="cri-o://edf431bc8bde9ffadbbc3588ca4cad38ae75ab62a4b62381893a15d736d41671" gracePeriod=30 Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.759150 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.864520 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"adc13fc6-a2d8-44fd-ab25-075857884cad\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.864616 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/adc13fc6-a2d8-44fd-ab25-075857884cad-registry-certificates\") pod \"adc13fc6-a2d8-44fd-ab25-075857884cad\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.864672 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-bound-sa-token\") pod \"adc13fc6-a2d8-44fd-ab25-075857884cad\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.864711 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-registry-tls\") pod \"adc13fc6-a2d8-44fd-ab25-075857884cad\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.864748 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8g6d\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-kube-api-access-d8g6d\") pod \"adc13fc6-a2d8-44fd-ab25-075857884cad\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.864806 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/adc13fc6-a2d8-44fd-ab25-075857884cad-ca-trust-extracted\") pod \"adc13fc6-a2d8-44fd-ab25-075857884cad\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.864847 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adc13fc6-a2d8-44fd-ab25-075857884cad-trusted-ca\") pod \"adc13fc6-a2d8-44fd-ab25-075857884cad\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.864911 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/adc13fc6-a2d8-44fd-ab25-075857884cad-installation-pull-secrets\") pod \"adc13fc6-a2d8-44fd-ab25-075857884cad\" (UID: \"adc13fc6-a2d8-44fd-ab25-075857884cad\") " Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.867163 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adc13fc6-a2d8-44fd-ab25-075857884cad-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "adc13fc6-a2d8-44fd-ab25-075857884cad" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.870409 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adc13fc6-a2d8-44fd-ab25-075857884cad-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "adc13fc6-a2d8-44fd-ab25-075857884cad" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.875087 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adc13fc6-a2d8-44fd-ab25-075857884cad-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "adc13fc6-a2d8-44fd-ab25-075857884cad" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.876017 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "adc13fc6-a2d8-44fd-ab25-075857884cad" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.878265 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-kube-api-access-d8g6d" (OuterVolumeSpecName: "kube-api-access-d8g6d") pod "adc13fc6-a2d8-44fd-ab25-075857884cad" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad"). InnerVolumeSpecName "kube-api-access-d8g6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.879128 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "adc13fc6-a2d8-44fd-ab25-075857884cad" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.882480 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "adc13fc6-a2d8-44fd-ab25-075857884cad" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.902507 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adc13fc6-a2d8-44fd-ab25-075857884cad-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "adc13fc6-a2d8-44fd-ab25-075857884cad" (UID: "adc13fc6-a2d8-44fd-ab25-075857884cad"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.953016 4775 generic.go:334] "Generic (PLEG): container finished" podID="adc13fc6-a2d8-44fd-ab25-075857884cad" containerID="edf431bc8bde9ffadbbc3588ca4cad38ae75ab62a4b62381893a15d736d41671" exitCode=0 Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.953079 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" event={"ID":"adc13fc6-a2d8-44fd-ab25-075857884cad","Type":"ContainerDied","Data":"edf431bc8bde9ffadbbc3588ca4cad38ae75ab62a4b62381893a15d736d41671"} Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.953133 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.953161 4775 scope.go:117] "RemoveContainer" containerID="edf431bc8bde9ffadbbc3588ca4cad38ae75ab62a4b62381893a15d736d41671" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.953138 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nv4tc" event={"ID":"adc13fc6-a2d8-44fd-ab25-075857884cad","Type":"ContainerDied","Data":"edee121d93e0bedbc1dbe048021eed82daee6746e68ee586e5bc12cfbecb3c08"} Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.967304 4775 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.967365 4775 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.967390 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8g6d\" (UniqueName: \"kubernetes.io/projected/adc13fc6-a2d8-44fd-ab25-075857884cad-kube-api-access-d8g6d\") on node \"crc\" DevicePath \"\"" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.967409 4775 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/adc13fc6-a2d8-44fd-ab25-075857884cad-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.967427 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adc13fc6-a2d8-44fd-ab25-075857884cad-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.967444 4775 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/adc13fc6-a2d8-44fd-ab25-075857884cad-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.967460 4775 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/adc13fc6-a2d8-44fd-ab25-075857884cad-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.982278 4775 scope.go:117] "RemoveContainer" containerID="edf431bc8bde9ffadbbc3588ca4cad38ae75ab62a4b62381893a15d736d41671" Oct 02 01:49:18 crc kubenswrapper[4775]: E1002 01:49:18.983234 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edf431bc8bde9ffadbbc3588ca4cad38ae75ab62a4b62381893a15d736d41671\": container with ID starting with edf431bc8bde9ffadbbc3588ca4cad38ae75ab62a4b62381893a15d736d41671 not found: ID does not exist" containerID="edf431bc8bde9ffadbbc3588ca4cad38ae75ab62a4b62381893a15d736d41671" Oct 02 01:49:18 crc kubenswrapper[4775]: I1002 01:49:18.983294 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edf431bc8bde9ffadbbc3588ca4cad38ae75ab62a4b62381893a15d736d41671"} err="failed to get container status \"edf431bc8bde9ffadbbc3588ca4cad38ae75ab62a4b62381893a15d736d41671\": rpc error: code = NotFound desc = could not find container \"edf431bc8bde9ffadbbc3588ca4cad38ae75ab62a4b62381893a15d736d41671\": container with ID starting with edf431bc8bde9ffadbbc3588ca4cad38ae75ab62a4b62381893a15d736d41671 not found: ID does not exist" Oct 02 01:49:19 crc kubenswrapper[4775]: I1002 01:49:19.010591 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nv4tc"] Oct 02 01:49:19 crc kubenswrapper[4775]: I1002 01:49:19.015842 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nv4tc"] Oct 02 01:49:19 crc kubenswrapper[4775]: I1002 01:49:19.775689 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adc13fc6-a2d8-44fd-ab25-075857884cad" path="/var/lib/kubelet/pods/adc13fc6-a2d8-44fd-ab25-075857884cad/volumes" Oct 02 01:50:07 crc kubenswrapper[4775]: I1002 01:50:07.233762 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:50:07 crc kubenswrapper[4775]: I1002 01:50:07.235146 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:50:37 crc kubenswrapper[4775]: I1002 01:50:37.234097 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:50:37 crc kubenswrapper[4775]: I1002 01:50:37.234884 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:51:07 crc kubenswrapper[4775]: I1002 01:51:07.234331 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:51:07 crc kubenswrapper[4775]: I1002 01:51:07.235174 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:51:07 crc kubenswrapper[4775]: I1002 01:51:07.235238 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:51:07 crc kubenswrapper[4775]: I1002 01:51:07.236107 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2c48cbf3d08cb4a2f350c67fef84e9504d7f8966e661be9b0356695ca1016ee7"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 01:51:07 crc kubenswrapper[4775]: I1002 01:51:07.236205 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://2c48cbf3d08cb4a2f350c67fef84e9504d7f8966e661be9b0356695ca1016ee7" gracePeriod=600 Oct 02 01:51:07 crc kubenswrapper[4775]: I1002 01:51:07.714607 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="2c48cbf3d08cb4a2f350c67fef84e9504d7f8966e661be9b0356695ca1016ee7" exitCode=0 Oct 02 01:51:07 crc kubenswrapper[4775]: I1002 01:51:07.714856 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"2c48cbf3d08cb4a2f350c67fef84e9504d7f8966e661be9b0356695ca1016ee7"} Oct 02 01:51:07 crc kubenswrapper[4775]: I1002 01:51:07.715204 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"0af4ec8f881f09471035912f64ffe17497158c6365d4ee31b994b4425f60c6bb"} Oct 02 01:51:07 crc kubenswrapper[4775]: I1002 01:51:07.715249 4775 scope.go:117] "RemoveContainer" containerID="e819a457bcc1b0b56d9d1d4506dfc8bb5cf3853dae381f750b921e36982f1712" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.410450 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zbgms"] Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.412356 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovn-controller" containerID="cri-o://0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf" gracePeriod=30 Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.412435 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="nbdb" containerID="cri-o://309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b" gracePeriod=30 Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.412435 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780" gracePeriod=30 Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.412596 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="sbdb" containerID="cri-o://241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698" gracePeriod=30 Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.412853 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="northd" containerID="cri-o://c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352" gracePeriod=30 Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.412917 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovn-acl-logging" containerID="cri-o://f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654" gracePeriod=30 Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.413033 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="kube-rbac-proxy-node" containerID="cri-o://1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc" gracePeriod=30 Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.484755 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" containerID="cri-o://ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca" gracePeriod=30 Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.766829 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/3.log" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.768646 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovn-acl-logging/0.log" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.769014 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovn-controller/0.log" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.769348 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824385 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9946c"] Oct 02 01:52:26 crc kubenswrapper[4775]: E1002 01:52:26.824566 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="sbdb" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824579 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="sbdb" Oct 02 01:52:26 crc kubenswrapper[4775]: E1002 01:52:26.824590 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="northd" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824598 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="northd" Oct 02 01:52:26 crc kubenswrapper[4775]: E1002 01:52:26.824607 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="nbdb" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824615 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="nbdb" Oct 02 01:52:26 crc kubenswrapper[4775]: E1002 01:52:26.824623 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824629 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: E1002 01:52:26.824635 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="kubecfg-setup" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824641 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="kubecfg-setup" Oct 02 01:52:26 crc kubenswrapper[4775]: E1002 01:52:26.824648 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovn-acl-logging" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824654 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovn-acl-logging" Oct 02 01:52:26 crc kubenswrapper[4775]: E1002 01:52:26.824662 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824669 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: E1002 01:52:26.824678 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc13fc6-a2d8-44fd-ab25-075857884cad" containerName="registry" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824684 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc13fc6-a2d8-44fd-ab25-075857884cad" containerName="registry" Oct 02 01:52:26 crc kubenswrapper[4775]: E1002 01:52:26.824692 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824697 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: E1002 01:52:26.824704 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="kube-rbac-proxy-node" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824710 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="kube-rbac-proxy-node" Oct 02 01:52:26 crc kubenswrapper[4775]: E1002 01:52:26.824723 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovn-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824728 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovn-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: E1002 01:52:26.824738 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824744 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824828 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="nbdb" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824839 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824847 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="adc13fc6-a2d8-44fd-ab25-075857884cad" containerName="registry" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824855 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="kube-rbac-proxy-ovn-metrics" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824861 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="northd" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824868 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovn-acl-logging" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824876 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824881 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovn-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824890 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824896 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="sbdb" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824905 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="kube-rbac-proxy-node" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.824911 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: E1002 01:52:26.825011 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.825019 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: E1002 01:52:26.825029 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.825034 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.825112 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerName="ovnkube-controller" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.826549 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882617 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27cwc\" (UniqueName: \"kubernetes.io/projected/7178a2dd-9182-400a-959d-c0c4181b6f18-kube-api-access-27cwc\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882663 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-log-socket\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882684 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-ovnkube-config\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882704 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-node-log\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882719 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-cni-bin\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882736 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-etc-openvswitch\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882753 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-openvswitch\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882781 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-systemd-units\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882825 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-var-lib-openvswitch\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882851 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-var-lib-cni-networks-ovn-kubernetes\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882867 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-systemd\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882880 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-cni-netd\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882898 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-ovnkube-script-lib\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882918 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-env-overrides\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882923 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.882943 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-kubelet\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883017 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883049 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883050 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7178a2dd-9182-400a-959d-c0c4181b6f18-ovn-node-metrics-cert\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883067 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883085 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883089 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-slash\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883105 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883112 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-run-netns\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883154 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-ovn\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883179 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-run-ovn-kubernetes\") pod \"7178a2dd-9182-400a-959d-c0c4181b6f18\" (UID: \"7178a2dd-9182-400a-959d-c0c4181b6f18\") " Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883250 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-log-socket" (OuterVolumeSpecName: "log-socket") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883390 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-slash\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883436 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-node-log\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883457 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-log-socket\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883479 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-var-lib-openvswitch\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883522 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-run-ovn\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883541 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-etc-openvswitch\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883561 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1996395c-671a-4c74-b3ec-dad1dbeeb856-ovn-node-metrics-cert\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883598 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-run-systemd\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883621 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-run-ovn-kubernetes\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883640 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-run-netns\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883676 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1996395c-671a-4c74-b3ec-dad1dbeeb856-ovnkube-config\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883716 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1996395c-671a-4c74-b3ec-dad1dbeeb856-ovnkube-script-lib\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883772 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-cni-netd\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883750 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883788 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-run-openvswitch\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883902 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-kubelet\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883996 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cjd4\" (UniqueName: \"kubernetes.io/projected/1996395c-671a-4c74-b3ec-dad1dbeeb856-kube-api-access-7cjd4\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.884037 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1996395c-671a-4c74-b3ec-dad1dbeeb856-env-overrides\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.884074 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.884159 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-systemd-units\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.884210 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-cni-bin\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.884330 4775 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.884368 4775 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883822 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-slash" (OuterVolumeSpecName: "host-slash") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883839 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883854 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.883868 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.884032 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-node-log" (OuterVolumeSpecName: "node-log") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.884586 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.884728 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.884855 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.884745 4775 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.885174 4775 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-log-socket\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.885213 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.885237 4775 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.885256 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.885288 4775 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.885320 4775 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.889681 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7178a2dd-9182-400a-959d-c0c4181b6f18-kube-api-access-27cwc" (OuterVolumeSpecName: "kube-api-access-27cwc") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "kube-api-access-27cwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.890753 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7178a2dd-9182-400a-959d-c0c4181b6f18-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.899326 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "7178a2dd-9182-400a-959d-c0c4181b6f18" (UID: "7178a2dd-9182-400a-959d-c0c4181b6f18"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.986488 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-cni-netd\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.986538 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-run-openvswitch\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.986567 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-kubelet\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.986589 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cjd4\" (UniqueName: \"kubernetes.io/projected/1996395c-671a-4c74-b3ec-dad1dbeeb856-kube-api-access-7cjd4\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.986609 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1996395c-671a-4c74-b3ec-dad1dbeeb856-env-overrides\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.986632 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.986663 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-systemd-units\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.986643 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-cni-netd\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.986696 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-run-openvswitch\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.986752 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-kubelet\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.986740 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.986768 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-cni-bin\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.986764 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-systemd-units\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.986690 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-cni-bin\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987066 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-slash\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987132 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-node-log\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987152 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-slash\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987168 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-var-lib-openvswitch\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987191 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-node-log\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987207 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-log-socket\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987253 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-log-socket\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987267 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-run-ovn\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987278 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-var-lib-openvswitch\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987304 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-run-ovn\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987326 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-etc-openvswitch\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987378 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1996395c-671a-4c74-b3ec-dad1dbeeb856-ovn-node-metrics-cert\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987382 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-etc-openvswitch\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987471 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-run-systemd\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987520 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-run-netns\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987591 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-run-systemd\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987600 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-run-ovn-kubernetes\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987652 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-run-netns\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987714 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1996395c-671a-4c74-b3ec-dad1dbeeb856-ovnkube-config\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987751 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1996395c-671a-4c74-b3ec-dad1dbeeb856-host-run-ovn-kubernetes\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1996395c-671a-4c74-b3ec-dad1dbeeb856-ovnkube-script-lib\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.987950 4775 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.988047 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27cwc\" (UniqueName: \"kubernetes.io/projected/7178a2dd-9182-400a-959d-c0c4181b6f18-kube-api-access-27cwc\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.988076 4775 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-node-log\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.988050 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1996395c-671a-4c74-b3ec-dad1dbeeb856-env-overrides\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.988100 4775 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.988125 4775 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.988148 4775 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.988171 4775 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.988193 4775 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7178a2dd-9182-400a-959d-c0c4181b6f18-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.988217 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7178a2dd-9182-400a-959d-c0c4181b6f18-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.988239 4775 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-slash\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.988261 4775 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.988286 4775 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7178a2dd-9182-400a-959d-c0c4181b6f18-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.988848 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1996395c-671a-4c74-b3ec-dad1dbeeb856-ovnkube-config\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.991256 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1996395c-671a-4c74-b3ec-dad1dbeeb856-ovnkube-script-lib\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:26 crc kubenswrapper[4775]: I1002 01:52:26.992519 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1996395c-671a-4c74-b3ec-dad1dbeeb856-ovn-node-metrics-cert\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.020715 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cjd4\" (UniqueName: \"kubernetes.io/projected/1996395c-671a-4c74-b3ec-dad1dbeeb856-kube-api-access-7cjd4\") pod \"ovnkube-node-9946c\" (UID: \"1996395c-671a-4c74-b3ec-dad1dbeeb856\") " pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.148538 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.244310 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2xv98_ab7ffbf4-19df-4fff-aacc-344eab1d1089/kube-multus/2.log" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.245185 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2xv98_ab7ffbf4-19df-4fff-aacc-344eab1d1089/kube-multus/1.log" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.245256 4775 generic.go:334] "Generic (PLEG): container finished" podID="ab7ffbf4-19df-4fff-aacc-344eab1d1089" containerID="e6e3e3d42f733f6c72de0970398567ac16e6e561b792060ddeaa55cb5db1de67" exitCode=2 Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.245354 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2xv98" event={"ID":"ab7ffbf4-19df-4fff-aacc-344eab1d1089","Type":"ContainerDied","Data":"e6e3e3d42f733f6c72de0970398567ac16e6e561b792060ddeaa55cb5db1de67"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.245403 4775 scope.go:117] "RemoveContainer" containerID="816181b47894f3934c2d223e7a376a1adf06cc23e5d3529fc4fbe0c5f029503e" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.246526 4775 scope.go:117] "RemoveContainer" containerID="e6e3e3d42f733f6c72de0970398567ac16e6e561b792060ddeaa55cb5db1de67" Oct 02 01:52:27 crc kubenswrapper[4775]: E1002 01:52:27.247023 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2xv98_openshift-multus(ab7ffbf4-19df-4fff-aacc-344eab1d1089)\"" pod="openshift-multus/multus-2xv98" podUID="ab7ffbf4-19df-4fff-aacc-344eab1d1089" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.253503 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovnkube-controller/3.log" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.257557 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovn-acl-logging/0.log" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.258422 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-zbgms_7178a2dd-9182-400a-959d-c0c4181b6f18/ovn-controller/0.log" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.258830 4775 generic.go:334] "Generic (PLEG): container finished" podID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerID="ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca" exitCode=0 Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.258864 4775 generic.go:334] "Generic (PLEG): container finished" podID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerID="241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698" exitCode=0 Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.258876 4775 generic.go:334] "Generic (PLEG): container finished" podID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerID="309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b" exitCode=0 Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.258887 4775 generic.go:334] "Generic (PLEG): container finished" podID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerID="c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352" exitCode=0 Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.258897 4775 generic.go:334] "Generic (PLEG): container finished" podID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerID="070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780" exitCode=0 Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.258905 4775 generic.go:334] "Generic (PLEG): container finished" podID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerID="1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc" exitCode=0 Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.258914 4775 generic.go:334] "Generic (PLEG): container finished" podID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerID="f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654" exitCode=143 Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.258924 4775 generic.go:334] "Generic (PLEG): container finished" podID="7178a2dd-9182-400a-959d-c0c4181b6f18" containerID="0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf" exitCode=143 Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259002 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerDied","Data":"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259035 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerDied","Data":"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259052 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerDied","Data":"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259068 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerDied","Data":"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259081 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerDied","Data":"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259096 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerDied","Data":"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259110 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259124 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259133 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259142 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259150 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259158 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259166 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259175 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259183 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259191 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259201 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerDied","Data":"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259214 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259223 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259231 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259238 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259246 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259253 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259260 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259268 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259275 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259282 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259291 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerDied","Data":"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259302 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259314 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259321 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259329 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259340 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259349 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259359 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259369 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259379 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259388 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259401 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" event={"ID":"7178a2dd-9182-400a-959d-c0c4181b6f18","Type":"ContainerDied","Data":"96ee9704f7a21536812abcf8be7381ee439a206a18be456596d53a9a8bde4dee"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259417 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259429 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259437 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259444 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259451 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259458 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259465 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259472 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259479 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259486 4775 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.259163 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-zbgms" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.260385 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" event={"ID":"1996395c-671a-4c74-b3ec-dad1dbeeb856","Type":"ContainerStarted","Data":"74b56e2aa3670ea42fa4431d02047d89bd2542d1a7310bc53fb5f1278bda90c1"} Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.308730 4775 scope.go:117] "RemoveContainer" containerID="ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.323994 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zbgms"] Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.329925 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-zbgms"] Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.340080 4775 scope.go:117] "RemoveContainer" containerID="12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.370195 4775 scope.go:117] "RemoveContainer" containerID="241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.443664 4775 scope.go:117] "RemoveContainer" containerID="309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.465447 4775 scope.go:117] "RemoveContainer" containerID="c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.485991 4775 scope.go:117] "RemoveContainer" containerID="070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.508681 4775 scope.go:117] "RemoveContainer" containerID="1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.529214 4775 scope.go:117] "RemoveContainer" containerID="f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.548275 4775 scope.go:117] "RemoveContainer" containerID="0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.601590 4775 scope.go:117] "RemoveContainer" containerID="a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.622514 4775 scope.go:117] "RemoveContainer" containerID="ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca" Oct 02 01:52:27 crc kubenswrapper[4775]: E1002 01:52:27.623151 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca\": container with ID starting with ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca not found: ID does not exist" containerID="ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.623218 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca"} err="failed to get container status \"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca\": rpc error: code = NotFound desc = could not find container \"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca\": container with ID starting with ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.623263 4775 scope.go:117] "RemoveContainer" containerID="12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3" Oct 02 01:52:27 crc kubenswrapper[4775]: E1002 01:52:27.624073 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\": container with ID starting with 12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3 not found: ID does not exist" containerID="12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.624125 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3"} err="failed to get container status \"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\": rpc error: code = NotFound desc = could not find container \"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\": container with ID starting with 12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.624161 4775 scope.go:117] "RemoveContainer" containerID="241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698" Oct 02 01:52:27 crc kubenswrapper[4775]: E1002 01:52:27.624945 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\": container with ID starting with 241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698 not found: ID does not exist" containerID="241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.625020 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698"} err="failed to get container status \"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\": rpc error: code = NotFound desc = could not find container \"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\": container with ID starting with 241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.625047 4775 scope.go:117] "RemoveContainer" containerID="309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b" Oct 02 01:52:27 crc kubenswrapper[4775]: E1002 01:52:27.625543 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\": container with ID starting with 309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b not found: ID does not exist" containerID="309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.625611 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b"} err="failed to get container status \"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\": rpc error: code = NotFound desc = could not find container \"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\": container with ID starting with 309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.625654 4775 scope.go:117] "RemoveContainer" containerID="c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352" Oct 02 01:52:27 crc kubenswrapper[4775]: E1002 01:52:27.626191 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\": container with ID starting with c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352 not found: ID does not exist" containerID="c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.626275 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352"} err="failed to get container status \"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\": rpc error: code = NotFound desc = could not find container \"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\": container with ID starting with c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.626317 4775 scope.go:117] "RemoveContainer" containerID="070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780" Oct 02 01:52:27 crc kubenswrapper[4775]: E1002 01:52:27.626727 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\": container with ID starting with 070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780 not found: ID does not exist" containerID="070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.626764 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780"} err="failed to get container status \"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\": rpc error: code = NotFound desc = could not find container \"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\": container with ID starting with 070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.626791 4775 scope.go:117] "RemoveContainer" containerID="1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc" Oct 02 01:52:27 crc kubenswrapper[4775]: E1002 01:52:27.627419 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\": container with ID starting with 1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc not found: ID does not exist" containerID="1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.627466 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc"} err="failed to get container status \"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\": rpc error: code = NotFound desc = could not find container \"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\": container with ID starting with 1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.627495 4775 scope.go:117] "RemoveContainer" containerID="f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654" Oct 02 01:52:27 crc kubenswrapper[4775]: E1002 01:52:27.628213 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\": container with ID starting with f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654 not found: ID does not exist" containerID="f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.628249 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654"} err="failed to get container status \"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\": rpc error: code = NotFound desc = could not find container \"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\": container with ID starting with f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.628281 4775 scope.go:117] "RemoveContainer" containerID="0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf" Oct 02 01:52:27 crc kubenswrapper[4775]: E1002 01:52:27.628880 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\": container with ID starting with 0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf not found: ID does not exist" containerID="0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.628921 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf"} err="failed to get container status \"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\": rpc error: code = NotFound desc = could not find container \"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\": container with ID starting with 0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.629014 4775 scope.go:117] "RemoveContainer" containerID="a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352" Oct 02 01:52:27 crc kubenswrapper[4775]: E1002 01:52:27.629645 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\": container with ID starting with a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352 not found: ID does not exist" containerID="a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.629688 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352"} err="failed to get container status \"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\": rpc error: code = NotFound desc = could not find container \"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\": container with ID starting with a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.629715 4775 scope.go:117] "RemoveContainer" containerID="ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.630524 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca"} err="failed to get container status \"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca\": rpc error: code = NotFound desc = could not find container \"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca\": container with ID starting with ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.630565 4775 scope.go:117] "RemoveContainer" containerID="12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.631154 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3"} err="failed to get container status \"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\": rpc error: code = NotFound desc = could not find container \"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\": container with ID starting with 12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.631190 4775 scope.go:117] "RemoveContainer" containerID="241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.632061 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698"} err="failed to get container status \"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\": rpc error: code = NotFound desc = could not find container \"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\": container with ID starting with 241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.632097 4775 scope.go:117] "RemoveContainer" containerID="309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.632529 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b"} err="failed to get container status \"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\": rpc error: code = NotFound desc = could not find container \"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\": container with ID starting with 309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.632566 4775 scope.go:117] "RemoveContainer" containerID="c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.632947 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352"} err="failed to get container status \"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\": rpc error: code = NotFound desc = could not find container \"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\": container with ID starting with c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.633012 4775 scope.go:117] "RemoveContainer" containerID="070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.633769 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780"} err="failed to get container status \"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\": rpc error: code = NotFound desc = could not find container \"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\": container with ID starting with 070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.633824 4775 scope.go:117] "RemoveContainer" containerID="1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.634358 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc"} err="failed to get container status \"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\": rpc error: code = NotFound desc = could not find container \"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\": container with ID starting with 1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.634399 4775 scope.go:117] "RemoveContainer" containerID="f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.634873 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654"} err="failed to get container status \"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\": rpc error: code = NotFound desc = could not find container \"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\": container with ID starting with f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.634913 4775 scope.go:117] "RemoveContainer" containerID="0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.635657 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf"} err="failed to get container status \"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\": rpc error: code = NotFound desc = could not find container \"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\": container with ID starting with 0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.635707 4775 scope.go:117] "RemoveContainer" containerID="a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.636090 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352"} err="failed to get container status \"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\": rpc error: code = NotFound desc = could not find container \"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\": container with ID starting with a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.636126 4775 scope.go:117] "RemoveContainer" containerID="ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.636780 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca"} err="failed to get container status \"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca\": rpc error: code = NotFound desc = could not find container \"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca\": container with ID starting with ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.636829 4775 scope.go:117] "RemoveContainer" containerID="12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.637338 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3"} err="failed to get container status \"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\": rpc error: code = NotFound desc = could not find container \"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\": container with ID starting with 12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.637375 4775 scope.go:117] "RemoveContainer" containerID="241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.637836 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698"} err="failed to get container status \"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\": rpc error: code = NotFound desc = could not find container \"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\": container with ID starting with 241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.637874 4775 scope.go:117] "RemoveContainer" containerID="309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.638308 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b"} err="failed to get container status \"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\": rpc error: code = NotFound desc = could not find container \"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\": container with ID starting with 309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.638366 4775 scope.go:117] "RemoveContainer" containerID="c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.638805 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352"} err="failed to get container status \"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\": rpc error: code = NotFound desc = could not find container \"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\": container with ID starting with c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.638849 4775 scope.go:117] "RemoveContainer" containerID="070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.639679 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780"} err="failed to get container status \"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\": rpc error: code = NotFound desc = could not find container \"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\": container with ID starting with 070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.639727 4775 scope.go:117] "RemoveContainer" containerID="1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.640650 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc"} err="failed to get container status \"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\": rpc error: code = NotFound desc = could not find container \"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\": container with ID starting with 1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.640699 4775 scope.go:117] "RemoveContainer" containerID="f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.641368 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654"} err="failed to get container status \"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\": rpc error: code = NotFound desc = could not find container \"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\": container with ID starting with f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.641406 4775 scope.go:117] "RemoveContainer" containerID="0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.642072 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf"} err="failed to get container status \"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\": rpc error: code = NotFound desc = could not find container \"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\": container with ID starting with 0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.642123 4775 scope.go:117] "RemoveContainer" containerID="a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.642536 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352"} err="failed to get container status \"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\": rpc error: code = NotFound desc = could not find container \"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\": container with ID starting with a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.642575 4775 scope.go:117] "RemoveContainer" containerID="ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.643340 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca"} err="failed to get container status \"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca\": rpc error: code = NotFound desc = could not find container \"ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca\": container with ID starting with ac6fa1d0f770e80a3145cf2a86a9b731f1ecbd000931646ee4e1f18d612d9dca not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.643387 4775 scope.go:117] "RemoveContainer" containerID="12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.644086 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3"} err="failed to get container status \"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\": rpc error: code = NotFound desc = could not find container \"12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3\": container with ID starting with 12e870377b188d59bae376e44b9446b710275954b2704d71c4ce22a051739dc3 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.644135 4775 scope.go:117] "RemoveContainer" containerID="241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.644586 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698"} err="failed to get container status \"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\": rpc error: code = NotFound desc = could not find container \"241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698\": container with ID starting with 241fed824d73d1a22a9ba5670be56d0e66d50a510f1c0f29c7f49c60a491e698 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.644628 4775 scope.go:117] "RemoveContainer" containerID="309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.645256 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b"} err="failed to get container status \"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\": rpc error: code = NotFound desc = could not find container \"309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b\": container with ID starting with 309fa89549604fa35b8a6a83045137c7f333cce814d16ac9b864408883ca605b not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.645293 4775 scope.go:117] "RemoveContainer" containerID="c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.645768 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352"} err="failed to get container status \"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\": rpc error: code = NotFound desc = could not find container \"c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352\": container with ID starting with c0bcd59557640df8c1f8c5a2a81389e7d34da71db705c28a1a167159995ce352 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.645806 4775 scope.go:117] "RemoveContainer" containerID="070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.646242 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780"} err="failed to get container status \"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\": rpc error: code = NotFound desc = could not find container \"070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780\": container with ID starting with 070fe37253b6199cb8581af765d94b83a9fd37c2409044c24cef219d8454a780 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.646278 4775 scope.go:117] "RemoveContainer" containerID="1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.646644 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc"} err="failed to get container status \"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\": rpc error: code = NotFound desc = could not find container \"1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc\": container with ID starting with 1d441aba881c39a7b9df0e9278044c9673fffcfaa4281c37a395548fc7e2b8cc not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.646683 4775 scope.go:117] "RemoveContainer" containerID="f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.647011 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654"} err="failed to get container status \"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\": rpc error: code = NotFound desc = could not find container \"f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654\": container with ID starting with f55016ea92158ba0c27b6f9898f3156dd73d6feebf35c3e96dbf06d826908654 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.647046 4775 scope.go:117] "RemoveContainer" containerID="0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.647493 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf"} err="failed to get container status \"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\": rpc error: code = NotFound desc = could not find container \"0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf\": container with ID starting with 0b7776e18365c9445a06854bb4657f63f7cdc63b8817de33ab012906f7c916bf not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.647533 4775 scope.go:117] "RemoveContainer" containerID="a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.648044 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352"} err="failed to get container status \"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\": rpc error: code = NotFound desc = could not find container \"a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352\": container with ID starting with a4a83f91df650239140a2ed5db7bf1296069a2ce90de930d22b7a8b187e4d352 not found: ID does not exist" Oct 02 01:52:27 crc kubenswrapper[4775]: I1002 01:52:27.777633 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7178a2dd-9182-400a-959d-c0c4181b6f18" path="/var/lib/kubelet/pods/7178a2dd-9182-400a-959d-c0c4181b6f18/volumes" Oct 02 01:52:28 crc kubenswrapper[4775]: I1002 01:52:28.271302 4775 generic.go:334] "Generic (PLEG): container finished" podID="1996395c-671a-4c74-b3ec-dad1dbeeb856" containerID="ff65e3cb015597011ab0fa73568c4cbd63c256d6bbe608c1553c0b44185085f7" exitCode=0 Oct 02 01:52:28 crc kubenswrapper[4775]: I1002 01:52:28.271406 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" event={"ID":"1996395c-671a-4c74-b3ec-dad1dbeeb856","Type":"ContainerDied","Data":"ff65e3cb015597011ab0fa73568c4cbd63c256d6bbe608c1553c0b44185085f7"} Oct 02 01:52:28 crc kubenswrapper[4775]: I1002 01:52:28.274333 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2xv98_ab7ffbf4-19df-4fff-aacc-344eab1d1089/kube-multus/2.log" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.255597 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-w7vtm"] Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.257354 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.260437 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.262028 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.262420 4775 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-6dh6b" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.263405 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.287344 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" event={"ID":"1996395c-671a-4c74-b3ec-dad1dbeeb856","Type":"ContainerStarted","Data":"978fd836ea5b91cd1149f4f4ac476c3b72b7d6331893d3aae7c5336ac0512bab"} Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.287395 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" event={"ID":"1996395c-671a-4c74-b3ec-dad1dbeeb856","Type":"ContainerStarted","Data":"fd0e6595cc557b1f89b643bd7fc3286f7b89466774f98879a075ce31fe8b464b"} Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.287410 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" event={"ID":"1996395c-671a-4c74-b3ec-dad1dbeeb856","Type":"ContainerStarted","Data":"5f32c4a25c5171e3c755b535083b699dc77498c9ae3db2ccb102dbc64c097b47"} Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.287422 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" event={"ID":"1996395c-671a-4c74-b3ec-dad1dbeeb856","Type":"ContainerStarted","Data":"12b1fbed9ce7a5b9b925233a1d8ec8d34b7a18266cc8808604038d2160d18264"} Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.287433 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" event={"ID":"1996395c-671a-4c74-b3ec-dad1dbeeb856","Type":"ContainerStarted","Data":"0a6df3951ace0d6cc9186c317d22e0f31acda3816c83a2ca323ee9b9c3c15507"} Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.287446 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" event={"ID":"1996395c-671a-4c74-b3ec-dad1dbeeb856","Type":"ContainerStarted","Data":"64203efa6ec73229a1cc790efdc4beed6b122f64a940f24058774195504cc163"} Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.323104 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-node-mnt\") pod \"crc-storage-crc-w7vtm\" (UID: \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\") " pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.323574 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blnmh\" (UniqueName: \"kubernetes.io/projected/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-kube-api-access-blnmh\") pod \"crc-storage-crc-w7vtm\" (UID: \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\") " pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.323810 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-crc-storage\") pod \"crc-storage-crc-w7vtm\" (UID: \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\") " pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.432284 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blnmh\" (UniqueName: \"kubernetes.io/projected/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-kube-api-access-blnmh\") pod \"crc-storage-crc-w7vtm\" (UID: \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\") " pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.432456 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-crc-storage\") pod \"crc-storage-crc-w7vtm\" (UID: \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\") " pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.432621 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-node-mnt\") pod \"crc-storage-crc-w7vtm\" (UID: \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\") " pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.433195 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-node-mnt\") pod \"crc-storage-crc-w7vtm\" (UID: \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\") " pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.440342 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-crc-storage\") pod \"crc-storage-crc-w7vtm\" (UID: \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\") " pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.481176 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blnmh\" (UniqueName: \"kubernetes.io/projected/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-kube-api-access-blnmh\") pod \"crc-storage-crc-w7vtm\" (UID: \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\") " pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:29 crc kubenswrapper[4775]: I1002 01:52:29.580618 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:29 crc kubenswrapper[4775]: E1002 01:52:29.612537 4775 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-w7vtm_crc-storage_fa5168f2-58f8-4af5-85b3-c8e3b9531ffd_0(ea47550c6e2d436dd65d84802f66431598a33f3f7b5d3ff2a4b12b43e957fea5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 01:52:29 crc kubenswrapper[4775]: E1002 01:52:29.612638 4775 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-w7vtm_crc-storage_fa5168f2-58f8-4af5-85b3-c8e3b9531ffd_0(ea47550c6e2d436dd65d84802f66431598a33f3f7b5d3ff2a4b12b43e957fea5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:29 crc kubenswrapper[4775]: E1002 01:52:29.612674 4775 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-w7vtm_crc-storage_fa5168f2-58f8-4af5-85b3-c8e3b9531ffd_0(ea47550c6e2d436dd65d84802f66431598a33f3f7b5d3ff2a4b12b43e957fea5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:29 crc kubenswrapper[4775]: E1002 01:52:29.612768 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-w7vtm_crc-storage(fa5168f2-58f8-4af5-85b3-c8e3b9531ffd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-w7vtm_crc-storage(fa5168f2-58f8-4af5-85b3-c8e3b9531ffd)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-w7vtm_crc-storage_fa5168f2-58f8-4af5-85b3-c8e3b9531ffd_0(ea47550c6e2d436dd65d84802f66431598a33f3f7b5d3ff2a4b12b43e957fea5): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-w7vtm" podUID="fa5168f2-58f8-4af5-85b3-c8e3b9531ffd" Oct 02 01:52:32 crc kubenswrapper[4775]: I1002 01:52:32.319637 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" event={"ID":"1996395c-671a-4c74-b3ec-dad1dbeeb856","Type":"ContainerStarted","Data":"463d66e52fd9211b0304199aa98a6ccdbab34a3ad0b41b5d56a41d9d84d7981e"} Oct 02 01:52:34 crc kubenswrapper[4775]: I1002 01:52:34.334826 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" event={"ID":"1996395c-671a-4c74-b3ec-dad1dbeeb856","Type":"ContainerStarted","Data":"eecf040e60114939048c8b883441b769471e080c9209b262b95994ca0f887f98"} Oct 02 01:52:34 crc kubenswrapper[4775]: I1002 01:52:34.335228 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:34 crc kubenswrapper[4775]: I1002 01:52:34.335258 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:34 crc kubenswrapper[4775]: I1002 01:52:34.335271 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:34 crc kubenswrapper[4775]: I1002 01:52:34.363610 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:34 crc kubenswrapper[4775]: I1002 01:52:34.366827 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:52:34 crc kubenswrapper[4775]: I1002 01:52:34.378142 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" podStartSLOduration=8.378122518 podStartE2EDuration="8.378122518s" podCreationTimestamp="2025-10-02 01:52:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:52:34.376646431 +0000 UTC m=+691.543390471" watchObservedRunningTime="2025-10-02 01:52:34.378122518 +0000 UTC m=+691.544866588" Oct 02 01:52:34 crc kubenswrapper[4775]: I1002 01:52:34.593545 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-w7vtm"] Oct 02 01:52:34 crc kubenswrapper[4775]: I1002 01:52:34.593692 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:34 crc kubenswrapper[4775]: I1002 01:52:34.594174 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:34 crc kubenswrapper[4775]: E1002 01:52:34.629130 4775 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-w7vtm_crc-storage_fa5168f2-58f8-4af5-85b3-c8e3b9531ffd_0(29cd052b6c5caefb4e8c690c88e63e628e9a708ee4399001a4ccff0ab3793f63): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 01:52:34 crc kubenswrapper[4775]: E1002 01:52:34.629246 4775 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-w7vtm_crc-storage_fa5168f2-58f8-4af5-85b3-c8e3b9531ffd_0(29cd052b6c5caefb4e8c690c88e63e628e9a708ee4399001a4ccff0ab3793f63): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:34 crc kubenswrapper[4775]: E1002 01:52:34.629286 4775 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-w7vtm_crc-storage_fa5168f2-58f8-4af5-85b3-c8e3b9531ffd_0(29cd052b6c5caefb4e8c690c88e63e628e9a708ee4399001a4ccff0ab3793f63): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:34 crc kubenswrapper[4775]: E1002 01:52:34.629367 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-w7vtm_crc-storage(fa5168f2-58f8-4af5-85b3-c8e3b9531ffd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-w7vtm_crc-storage(fa5168f2-58f8-4af5-85b3-c8e3b9531ffd)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-w7vtm_crc-storage_fa5168f2-58f8-4af5-85b3-c8e3b9531ffd_0(29cd052b6c5caefb4e8c690c88e63e628e9a708ee4399001a4ccff0ab3793f63): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-w7vtm" podUID="fa5168f2-58f8-4af5-85b3-c8e3b9531ffd" Oct 02 01:52:39 crc kubenswrapper[4775]: I1002 01:52:39.766120 4775 scope.go:117] "RemoveContainer" containerID="e6e3e3d42f733f6c72de0970398567ac16e6e561b792060ddeaa55cb5db1de67" Oct 02 01:52:39 crc kubenswrapper[4775]: E1002 01:52:39.766705 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-2xv98_openshift-multus(ab7ffbf4-19df-4fff-aacc-344eab1d1089)\"" pod="openshift-multus/multus-2xv98" podUID="ab7ffbf4-19df-4fff-aacc-344eab1d1089" Oct 02 01:52:48 crc kubenswrapper[4775]: I1002 01:52:48.764776 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:48 crc kubenswrapper[4775]: I1002 01:52:48.767898 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:48 crc kubenswrapper[4775]: E1002 01:52:48.811241 4775 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-w7vtm_crc-storage_fa5168f2-58f8-4af5-85b3-c8e3b9531ffd_0(07d684955b606f2cafbcb74b12290e3e8d05ce32913b4b224eec94963fb17638): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 02 01:52:48 crc kubenswrapper[4775]: E1002 01:52:48.811342 4775 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-w7vtm_crc-storage_fa5168f2-58f8-4af5-85b3-c8e3b9531ffd_0(07d684955b606f2cafbcb74b12290e3e8d05ce32913b4b224eec94963fb17638): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:48 crc kubenswrapper[4775]: E1002 01:52:48.811382 4775 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-w7vtm_crc-storage_fa5168f2-58f8-4af5-85b3-c8e3b9531ffd_0(07d684955b606f2cafbcb74b12290e3e8d05ce32913b4b224eec94963fb17638): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:52:48 crc kubenswrapper[4775]: E1002 01:52:48.811477 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-w7vtm_crc-storage(fa5168f2-58f8-4af5-85b3-c8e3b9531ffd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-w7vtm_crc-storage(fa5168f2-58f8-4af5-85b3-c8e3b9531ffd)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-w7vtm_crc-storage_fa5168f2-58f8-4af5-85b3-c8e3b9531ffd_0(07d684955b606f2cafbcb74b12290e3e8d05ce32913b4b224eec94963fb17638): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-w7vtm" podUID="fa5168f2-58f8-4af5-85b3-c8e3b9531ffd" Oct 02 01:52:51 crc kubenswrapper[4775]: I1002 01:52:51.765413 4775 scope.go:117] "RemoveContainer" containerID="e6e3e3d42f733f6c72de0970398567ac16e6e561b792060ddeaa55cb5db1de67" Oct 02 01:52:52 crc kubenswrapper[4775]: I1002 01:52:52.458676 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-2xv98_ab7ffbf4-19df-4fff-aacc-344eab1d1089/kube-multus/2.log" Oct 02 01:52:52 crc kubenswrapper[4775]: I1002 01:52:52.458777 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-2xv98" event={"ID":"ab7ffbf4-19df-4fff-aacc-344eab1d1089","Type":"ContainerStarted","Data":"299f1a9948f34948b81854e26963bbc90fbcf85ca780dd703a164b934d9eb9ab"} Oct 02 01:52:57 crc kubenswrapper[4775]: I1002 01:52:57.188827 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9946c" Oct 02 01:53:01 crc kubenswrapper[4775]: I1002 01:53:01.765344 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:53:01 crc kubenswrapper[4775]: I1002 01:53:01.766473 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:53:01 crc kubenswrapper[4775]: I1002 01:53:01.997308 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-w7vtm"] Oct 02 01:53:02 crc kubenswrapper[4775]: I1002 01:53:02.007550 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 01:53:02 crc kubenswrapper[4775]: I1002 01:53:02.525026 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-w7vtm" event={"ID":"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd","Type":"ContainerStarted","Data":"af52793b8e6caa4190dd50f2a7ecdb41399e4ec119cc2966363ab392ac2405a9"} Oct 02 01:53:04 crc kubenswrapper[4775]: I1002 01:53:04.537893 4775 generic.go:334] "Generic (PLEG): container finished" podID="fa5168f2-58f8-4af5-85b3-c8e3b9531ffd" containerID="d7aa23990581129b590ef8697a4348d53eb75c9327128fa2f71f21e590de372a" exitCode=0 Oct 02 01:53:04 crc kubenswrapper[4775]: I1002 01:53:04.538008 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-w7vtm" event={"ID":"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd","Type":"ContainerDied","Data":"d7aa23990581129b590ef8697a4348d53eb75c9327128fa2f71f21e590de372a"} Oct 02 01:53:05 crc kubenswrapper[4775]: I1002 01:53:05.848305 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:53:05 crc kubenswrapper[4775]: I1002 01:53:05.944661 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-crc-storage\") pod \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\" (UID: \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\") " Oct 02 01:53:05 crc kubenswrapper[4775]: I1002 01:53:05.944740 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blnmh\" (UniqueName: \"kubernetes.io/projected/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-kube-api-access-blnmh\") pod \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\" (UID: \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\") " Oct 02 01:53:05 crc kubenswrapper[4775]: I1002 01:53:05.944825 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-node-mnt\") pod \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\" (UID: \"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd\") " Oct 02 01:53:05 crc kubenswrapper[4775]: I1002 01:53:05.945065 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "fa5168f2-58f8-4af5-85b3-c8e3b9531ffd" (UID: "fa5168f2-58f8-4af5-85b3-c8e3b9531ffd"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:53:05 crc kubenswrapper[4775]: I1002 01:53:05.951336 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-kube-api-access-blnmh" (OuterVolumeSpecName: "kube-api-access-blnmh") pod "fa5168f2-58f8-4af5-85b3-c8e3b9531ffd" (UID: "fa5168f2-58f8-4af5-85b3-c8e3b9531ffd"). InnerVolumeSpecName "kube-api-access-blnmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:53:05 crc kubenswrapper[4775]: I1002 01:53:05.964694 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "fa5168f2-58f8-4af5-85b3-c8e3b9531ffd" (UID: "fa5168f2-58f8-4af5-85b3-c8e3b9531ffd"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:53:06 crc kubenswrapper[4775]: I1002 01:53:06.046185 4775 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:06 crc kubenswrapper[4775]: I1002 01:53:06.046220 4775 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:06 crc kubenswrapper[4775]: I1002 01:53:06.046232 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blnmh\" (UniqueName: \"kubernetes.io/projected/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd-kube-api-access-blnmh\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:06 crc kubenswrapper[4775]: I1002 01:53:06.553815 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-w7vtm" event={"ID":"fa5168f2-58f8-4af5-85b3-c8e3b9531ffd","Type":"ContainerDied","Data":"af52793b8e6caa4190dd50f2a7ecdb41399e4ec119cc2966363ab392ac2405a9"} Oct 02 01:53:06 crc kubenswrapper[4775]: I1002 01:53:06.553883 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af52793b8e6caa4190dd50f2a7ecdb41399e4ec119cc2966363ab392ac2405a9" Oct 02 01:53:06 crc kubenswrapper[4775]: I1002 01:53:06.553893 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-w7vtm" Oct 02 01:53:07 crc kubenswrapper[4775]: I1002 01:53:07.234269 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:53:07 crc kubenswrapper[4775]: I1002 01:53:07.234526 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.184285 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp"] Oct 02 01:53:14 crc kubenswrapper[4775]: E1002 01:53:14.184740 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa5168f2-58f8-4af5-85b3-c8e3b9531ffd" containerName="storage" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.184754 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa5168f2-58f8-4af5-85b3-c8e3b9531ffd" containerName="storage" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.184847 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa5168f2-58f8-4af5-85b3-c8e3b9531ffd" containerName="storage" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.185525 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.190185 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.207648 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp"] Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.364109 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e933c35-a14e-4439-ae32-20f95df5eb7b-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp\" (UID: \"7e933c35-a14e-4439-ae32-20f95df5eb7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.364193 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtp24\" (UniqueName: \"kubernetes.io/projected/7e933c35-a14e-4439-ae32-20f95df5eb7b-kube-api-access-gtp24\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp\" (UID: \"7e933c35-a14e-4439-ae32-20f95df5eb7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.364311 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e933c35-a14e-4439-ae32-20f95df5eb7b-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp\" (UID: \"7e933c35-a14e-4439-ae32-20f95df5eb7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.465442 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e933c35-a14e-4439-ae32-20f95df5eb7b-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp\" (UID: \"7e933c35-a14e-4439-ae32-20f95df5eb7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.465508 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtp24\" (UniqueName: \"kubernetes.io/projected/7e933c35-a14e-4439-ae32-20f95df5eb7b-kube-api-access-gtp24\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp\" (UID: \"7e933c35-a14e-4439-ae32-20f95df5eb7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.465588 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e933c35-a14e-4439-ae32-20f95df5eb7b-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp\" (UID: \"7e933c35-a14e-4439-ae32-20f95df5eb7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.466152 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e933c35-a14e-4439-ae32-20f95df5eb7b-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp\" (UID: \"7e933c35-a14e-4439-ae32-20f95df5eb7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.466250 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e933c35-a14e-4439-ae32-20f95df5eb7b-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp\" (UID: \"7e933c35-a14e-4439-ae32-20f95df5eb7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.504375 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtp24\" (UniqueName: \"kubernetes.io/projected/7e933c35-a14e-4439-ae32-20f95df5eb7b-kube-api-access-gtp24\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp\" (UID: \"7e933c35-a14e-4439-ae32-20f95df5eb7b\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.510994 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" Oct 02 01:53:14 crc kubenswrapper[4775]: I1002 01:53:14.785140 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp"] Oct 02 01:53:15 crc kubenswrapper[4775]: I1002 01:53:15.607069 4775 generic.go:334] "Generic (PLEG): container finished" podID="7e933c35-a14e-4439-ae32-20f95df5eb7b" containerID="74ee703bca5fa191ca3f41f6ea4b2fdc3bce051724a5a6511ea717a7c3ce6df6" exitCode=0 Oct 02 01:53:15 crc kubenswrapper[4775]: I1002 01:53:15.607168 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" event={"ID":"7e933c35-a14e-4439-ae32-20f95df5eb7b","Type":"ContainerDied","Data":"74ee703bca5fa191ca3f41f6ea4b2fdc3bce051724a5a6511ea717a7c3ce6df6"} Oct 02 01:53:15 crc kubenswrapper[4775]: I1002 01:53:15.607404 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" event={"ID":"7e933c35-a14e-4439-ae32-20f95df5eb7b","Type":"ContainerStarted","Data":"cf55d241bd2ed7368eba03bb3c706e61912f3d063f5f36da966c89b0a1b4c7f5"} Oct 02 01:53:17 crc kubenswrapper[4775]: I1002 01:53:17.619969 4775 generic.go:334] "Generic (PLEG): container finished" podID="7e933c35-a14e-4439-ae32-20f95df5eb7b" containerID="63682d68cf80e484caccc3425f5de3e67fa7bc70000d6639701d6d69cb65d943" exitCode=0 Oct 02 01:53:17 crc kubenswrapper[4775]: I1002 01:53:17.620027 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" event={"ID":"7e933c35-a14e-4439-ae32-20f95df5eb7b","Type":"ContainerDied","Data":"63682d68cf80e484caccc3425f5de3e67fa7bc70000d6639701d6d69cb65d943"} Oct 02 01:53:18 crc kubenswrapper[4775]: I1002 01:53:18.631012 4775 generic.go:334] "Generic (PLEG): container finished" podID="7e933c35-a14e-4439-ae32-20f95df5eb7b" containerID="e5b2867393f4d1ded02469dee4105e9ba8ed2fff518aa747399de3596c4166c8" exitCode=0 Oct 02 01:53:18 crc kubenswrapper[4775]: I1002 01:53:18.631085 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" event={"ID":"7e933c35-a14e-4439-ae32-20f95df5eb7b","Type":"ContainerDied","Data":"e5b2867393f4d1ded02469dee4105e9ba8ed2fff518aa747399de3596c4166c8"} Oct 02 01:53:19 crc kubenswrapper[4775]: I1002 01:53:19.922910 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" Oct 02 01:53:20 crc kubenswrapper[4775]: I1002 01:53:20.050618 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e933c35-a14e-4439-ae32-20f95df5eb7b-bundle\") pod \"7e933c35-a14e-4439-ae32-20f95df5eb7b\" (UID: \"7e933c35-a14e-4439-ae32-20f95df5eb7b\") " Oct 02 01:53:20 crc kubenswrapper[4775]: I1002 01:53:20.050878 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtp24\" (UniqueName: \"kubernetes.io/projected/7e933c35-a14e-4439-ae32-20f95df5eb7b-kube-api-access-gtp24\") pod \"7e933c35-a14e-4439-ae32-20f95df5eb7b\" (UID: \"7e933c35-a14e-4439-ae32-20f95df5eb7b\") " Oct 02 01:53:20 crc kubenswrapper[4775]: I1002 01:53:20.051809 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e933c35-a14e-4439-ae32-20f95df5eb7b-bundle" (OuterVolumeSpecName: "bundle") pod "7e933c35-a14e-4439-ae32-20f95df5eb7b" (UID: "7e933c35-a14e-4439-ae32-20f95df5eb7b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:53:20 crc kubenswrapper[4775]: I1002 01:53:20.052202 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e933c35-a14e-4439-ae32-20f95df5eb7b-util\") pod \"7e933c35-a14e-4439-ae32-20f95df5eb7b\" (UID: \"7e933c35-a14e-4439-ae32-20f95df5eb7b\") " Oct 02 01:53:20 crc kubenswrapper[4775]: I1002 01:53:20.052512 4775 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7e933c35-a14e-4439-ae32-20f95df5eb7b-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:20 crc kubenswrapper[4775]: I1002 01:53:20.061823 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e933c35-a14e-4439-ae32-20f95df5eb7b-kube-api-access-gtp24" (OuterVolumeSpecName: "kube-api-access-gtp24") pod "7e933c35-a14e-4439-ae32-20f95df5eb7b" (UID: "7e933c35-a14e-4439-ae32-20f95df5eb7b"). InnerVolumeSpecName "kube-api-access-gtp24". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:53:20 crc kubenswrapper[4775]: I1002 01:53:20.085702 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e933c35-a14e-4439-ae32-20f95df5eb7b-util" (OuterVolumeSpecName: "util") pod "7e933c35-a14e-4439-ae32-20f95df5eb7b" (UID: "7e933c35-a14e-4439-ae32-20f95df5eb7b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:53:20 crc kubenswrapper[4775]: I1002 01:53:20.153784 4775 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7e933c35-a14e-4439-ae32-20f95df5eb7b-util\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:20 crc kubenswrapper[4775]: I1002 01:53:20.154117 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtp24\" (UniqueName: \"kubernetes.io/projected/7e933c35-a14e-4439-ae32-20f95df5eb7b-kube-api-access-gtp24\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:20 crc kubenswrapper[4775]: I1002 01:53:20.648058 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" event={"ID":"7e933c35-a14e-4439-ae32-20f95df5eb7b","Type":"ContainerDied","Data":"cf55d241bd2ed7368eba03bb3c706e61912f3d063f5f36da966c89b0a1b4c7f5"} Oct 02 01:53:20 crc kubenswrapper[4775]: I1002 01:53:20.648115 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cf55d241bd2ed7368eba03bb3c706e61912f3d063f5f36da966c89b0a1b4c7f5" Oct 02 01:53:20 crc kubenswrapper[4775]: I1002 01:53:20.648483 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp" Oct 02 01:53:22 crc kubenswrapper[4775]: I1002 01:53:22.814849 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-t84m4"] Oct 02 01:53:22 crc kubenswrapper[4775]: E1002 01:53:22.815046 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e933c35-a14e-4439-ae32-20f95df5eb7b" containerName="util" Oct 02 01:53:22 crc kubenswrapper[4775]: I1002 01:53:22.815058 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e933c35-a14e-4439-ae32-20f95df5eb7b" containerName="util" Oct 02 01:53:22 crc kubenswrapper[4775]: E1002 01:53:22.815072 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e933c35-a14e-4439-ae32-20f95df5eb7b" containerName="extract" Oct 02 01:53:22 crc kubenswrapper[4775]: I1002 01:53:22.815078 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e933c35-a14e-4439-ae32-20f95df5eb7b" containerName="extract" Oct 02 01:53:22 crc kubenswrapper[4775]: E1002 01:53:22.815087 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e933c35-a14e-4439-ae32-20f95df5eb7b" containerName="pull" Oct 02 01:53:22 crc kubenswrapper[4775]: I1002 01:53:22.815093 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e933c35-a14e-4439-ae32-20f95df5eb7b" containerName="pull" Oct 02 01:53:22 crc kubenswrapper[4775]: I1002 01:53:22.815371 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e933c35-a14e-4439-ae32-20f95df5eb7b" containerName="extract" Oct 02 01:53:22 crc kubenswrapper[4775]: I1002 01:53:22.815697 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-t84m4" Oct 02 01:53:22 crc kubenswrapper[4775]: I1002 01:53:22.821641 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 02 01:53:22 crc kubenswrapper[4775]: I1002 01:53:22.822108 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-lwrz9" Oct 02 01:53:22 crc kubenswrapper[4775]: I1002 01:53:22.824646 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 02 01:53:22 crc kubenswrapper[4775]: I1002 01:53:22.836166 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-t84m4"] Oct 02 01:53:22 crc kubenswrapper[4775]: I1002 01:53:22.992309 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq4l5\" (UniqueName: \"kubernetes.io/projected/bc7f4fa5-9d4f-4c61-9cd1-b91f57af5ca0-kube-api-access-lq4l5\") pod \"nmstate-operator-858ddd8f98-t84m4\" (UID: \"bc7f4fa5-9d4f-4c61-9cd1-b91f57af5ca0\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-t84m4" Oct 02 01:53:23 crc kubenswrapper[4775]: I1002 01:53:23.092925 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq4l5\" (UniqueName: \"kubernetes.io/projected/bc7f4fa5-9d4f-4c61-9cd1-b91f57af5ca0-kube-api-access-lq4l5\") pod \"nmstate-operator-858ddd8f98-t84m4\" (UID: \"bc7f4fa5-9d4f-4c61-9cd1-b91f57af5ca0\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-t84m4" Oct 02 01:53:23 crc kubenswrapper[4775]: I1002 01:53:23.123979 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq4l5\" (UniqueName: \"kubernetes.io/projected/bc7f4fa5-9d4f-4c61-9cd1-b91f57af5ca0-kube-api-access-lq4l5\") pod \"nmstate-operator-858ddd8f98-t84m4\" (UID: \"bc7f4fa5-9d4f-4c61-9cd1-b91f57af5ca0\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-t84m4" Oct 02 01:53:23 crc kubenswrapper[4775]: I1002 01:53:23.128427 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-t84m4" Oct 02 01:53:23 crc kubenswrapper[4775]: I1002 01:53:23.371855 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-t84m4"] Oct 02 01:53:23 crc kubenswrapper[4775]: I1002 01:53:23.667846 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-t84m4" event={"ID":"bc7f4fa5-9d4f-4c61-9cd1-b91f57af5ca0","Type":"ContainerStarted","Data":"04d3f51e809908d352b681ab92b23c99ee9ac35e5dc34baea1858ac45a968ac0"} Oct 02 01:53:26 crc kubenswrapper[4775]: I1002 01:53:26.694412 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-t84m4" event={"ID":"bc7f4fa5-9d4f-4c61-9cd1-b91f57af5ca0","Type":"ContainerStarted","Data":"f1ccdf08bea357d246a2bd4c2e24d92a9b51ae16e6fb26abf61ad6a6d7a2a541"} Oct 02 01:53:26 crc kubenswrapper[4775]: I1002 01:53:26.727090 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-t84m4" podStartSLOduration=2.463948196 podStartE2EDuration="4.727063544s" podCreationTimestamp="2025-10-02 01:53:22 +0000 UTC" firstStartedPulling="2025-10-02 01:53:23.381267253 +0000 UTC m=+740.548011323" lastFinishedPulling="2025-10-02 01:53:25.644382591 +0000 UTC m=+742.811126671" observedRunningTime="2025-10-02 01:53:26.722200379 +0000 UTC m=+743.888944479" watchObservedRunningTime="2025-10-02 01:53:26.727063544 +0000 UTC m=+743.893807624" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.730476 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-xfm7h"] Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.732928 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xfm7h" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.737441 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-sqj7t" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.743780 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-xfm7h"] Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.752918 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj"] Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.753968 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.765948 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.793406 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-qf7xb"] Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.801170 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj"] Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.801292 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.861820 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv"] Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.862459 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.864071 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.864286 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-dxp27" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.872832 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-g5pnj\" (UID: \"b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.872911 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pk7z\" (UniqueName: \"kubernetes.io/projected/9922a2fe-fb9f-4622-ba18-518eae27300f-kube-api-access-8pk7z\") pod \"nmstate-metrics-fdff9cb8d-xfm7h\" (UID: \"9922a2fe-fb9f-4622-ba18-518eae27300f\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xfm7h" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.872967 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m5sm\" (UniqueName: \"kubernetes.io/projected/b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50-kube-api-access-7m5sm\") pod \"nmstate-webhook-6cdbc54649-g5pnj\" (UID: \"b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.873807 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.883156 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv"] Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.973923 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m5sm\" (UniqueName: \"kubernetes.io/projected/b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50-kube-api-access-7m5sm\") pod \"nmstate-webhook-6cdbc54649-g5pnj\" (UID: \"b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.974000 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/336c609a-f0e8-42c1-8cd8-bafb09510123-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-q9jhv\" (UID: \"336c609a-f0e8-42c1-8cd8-bafb09510123\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.974052 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm5pg\" (UniqueName: \"kubernetes.io/projected/336c609a-f0e8-42c1-8cd8-bafb09510123-kube-api-access-rm5pg\") pod \"nmstate-console-plugin-6b874cbd85-q9jhv\" (UID: \"336c609a-f0e8-42c1-8cd8-bafb09510123\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.974073 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/69f99de2-12a0-4dbb-9d40-c44352e5859b-nmstate-lock\") pod \"nmstate-handler-qf7xb\" (UID: \"69f99de2-12a0-4dbb-9d40-c44352e5859b\") " pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.974095 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m697z\" (UniqueName: \"kubernetes.io/projected/69f99de2-12a0-4dbb-9d40-c44352e5859b-kube-api-access-m697z\") pod \"nmstate-handler-qf7xb\" (UID: \"69f99de2-12a0-4dbb-9d40-c44352e5859b\") " pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.974113 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/69f99de2-12a0-4dbb-9d40-c44352e5859b-dbus-socket\") pod \"nmstate-handler-qf7xb\" (UID: \"69f99de2-12a0-4dbb-9d40-c44352e5859b\") " pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.974144 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-g5pnj\" (UID: \"b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.974215 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/336c609a-f0e8-42c1-8cd8-bafb09510123-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-q9jhv\" (UID: \"336c609a-f0e8-42c1-8cd8-bafb09510123\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.974267 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/69f99de2-12a0-4dbb-9d40-c44352e5859b-ovs-socket\") pod \"nmstate-handler-qf7xb\" (UID: \"69f99de2-12a0-4dbb-9d40-c44352e5859b\") " pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.974298 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pk7z\" (UniqueName: \"kubernetes.io/projected/9922a2fe-fb9f-4622-ba18-518eae27300f-kube-api-access-8pk7z\") pod \"nmstate-metrics-fdff9cb8d-xfm7h\" (UID: \"9922a2fe-fb9f-4622-ba18-518eae27300f\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xfm7h" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.979742 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-g5pnj\" (UID: \"b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj" Oct 02 01:53:27 crc kubenswrapper[4775]: I1002 01:53:27.997222 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m5sm\" (UniqueName: \"kubernetes.io/projected/b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50-kube-api-access-7m5sm\") pod \"nmstate-webhook-6cdbc54649-g5pnj\" (UID: \"b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.014103 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pk7z\" (UniqueName: \"kubernetes.io/projected/9922a2fe-fb9f-4622-ba18-518eae27300f-kube-api-access-8pk7z\") pod \"nmstate-metrics-fdff9cb8d-xfm7h\" (UID: \"9922a2fe-fb9f-4622-ba18-518eae27300f\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xfm7h" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.049880 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xfm7h" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.074695 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-59cc9ddbfc-529ct"] Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.075137 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/336c609a-f0e8-42c1-8cd8-bafb09510123-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-q9jhv\" (UID: \"336c609a-f0e8-42c1-8cd8-bafb09510123\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.076632 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/69f99de2-12a0-4dbb-9d40-c44352e5859b-ovs-socket\") pod \"nmstate-handler-qf7xb\" (UID: \"69f99de2-12a0-4dbb-9d40-c44352e5859b\") " pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.076983 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/336c609a-f0e8-42c1-8cd8-bafb09510123-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-q9jhv\" (UID: \"336c609a-f0e8-42c1-8cd8-bafb09510123\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.080649 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm5pg\" (UniqueName: \"kubernetes.io/projected/336c609a-f0e8-42c1-8cd8-bafb09510123-kube-api-access-rm5pg\") pod \"nmstate-console-plugin-6b874cbd85-q9jhv\" (UID: \"336c609a-f0e8-42c1-8cd8-bafb09510123\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.080713 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/69f99de2-12a0-4dbb-9d40-c44352e5859b-nmstate-lock\") pod \"nmstate-handler-qf7xb\" (UID: \"69f99de2-12a0-4dbb-9d40-c44352e5859b\") " pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.080771 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m697z\" (UniqueName: \"kubernetes.io/projected/69f99de2-12a0-4dbb-9d40-c44352e5859b-kube-api-access-m697z\") pod \"nmstate-handler-qf7xb\" (UID: \"69f99de2-12a0-4dbb-9d40-c44352e5859b\") " pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.080800 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/69f99de2-12a0-4dbb-9d40-c44352e5859b-dbus-socket\") pod \"nmstate-handler-qf7xb\" (UID: \"69f99de2-12a0-4dbb-9d40-c44352e5859b\") " pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.081176 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/69f99de2-12a0-4dbb-9d40-c44352e5859b-dbus-socket\") pod \"nmstate-handler-qf7xb\" (UID: \"69f99de2-12a0-4dbb-9d40-c44352e5859b\") " pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.081595 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/69f99de2-12a0-4dbb-9d40-c44352e5859b-nmstate-lock\") pod \"nmstate-handler-qf7xb\" (UID: \"69f99de2-12a0-4dbb-9d40-c44352e5859b\") " pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.082215 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/336c609a-f0e8-42c1-8cd8-bafb09510123-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-q9jhv\" (UID: \"336c609a-f0e8-42c1-8cd8-bafb09510123\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.082337 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/69f99de2-12a0-4dbb-9d40-c44352e5859b-ovs-socket\") pod \"nmstate-handler-qf7xb\" (UID: \"69f99de2-12a0-4dbb-9d40-c44352e5859b\") " pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.083539 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.084750 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.090130 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/336c609a-f0e8-42c1-8cd8-bafb09510123-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-q9jhv\" (UID: \"336c609a-f0e8-42c1-8cd8-bafb09510123\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.107575 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m697z\" (UniqueName: \"kubernetes.io/projected/69f99de2-12a0-4dbb-9d40-c44352e5859b-kube-api-access-m697z\") pod \"nmstate-handler-qf7xb\" (UID: \"69f99de2-12a0-4dbb-9d40-c44352e5859b\") " pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.110541 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-59cc9ddbfc-529ct"] Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.113015 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm5pg\" (UniqueName: \"kubernetes.io/projected/336c609a-f0e8-42c1-8cd8-bafb09510123-kube-api-access-rm5pg\") pod \"nmstate-console-plugin-6b874cbd85-q9jhv\" (UID: \"336c609a-f0e8-42c1-8cd8-bafb09510123\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.117555 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:28 crc kubenswrapper[4775]: W1002 01:53:28.140768 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod69f99de2_12a0_4dbb_9d40_c44352e5859b.slice/crio-0574e7c27e3a600883b995c443436cea18234a6f9f4f05831ba2bca32a38e1ab WatchSource:0}: Error finding container 0574e7c27e3a600883b995c443436cea18234a6f9f4f05831ba2bca32a38e1ab: Status 404 returned error can't find the container with id 0574e7c27e3a600883b995c443436cea18234a6f9f4f05831ba2bca32a38e1ab Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.180975 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.183797 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f6658705-1cc5-492a-a2a4-327446c4bcad-console-oauth-config\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.183832 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f6658705-1cc5-492a-a2a4-327446c4bcad-console-config\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.183859 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f6658705-1cc5-492a-a2a4-327446c4bcad-service-ca\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.183877 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klvxd\" (UniqueName: \"kubernetes.io/projected/f6658705-1cc5-492a-a2a4-327446c4bcad-kube-api-access-klvxd\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.183919 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f6658705-1cc5-492a-a2a4-327446c4bcad-console-serving-cert\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.183936 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f6658705-1cc5-492a-a2a4-327446c4bcad-oauth-serving-cert\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.183983 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6658705-1cc5-492a-a2a4-327446c4bcad-trusted-ca-bundle\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.271543 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-xfm7h"] Oct 02 01:53:28 crc kubenswrapper[4775]: W1002 01:53:28.277737 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9922a2fe_fb9f_4622_ba18_518eae27300f.slice/crio-b2597c9d398510e988f9e83416b80608d7a6095be571c79ace3b087f2b5f41ec WatchSource:0}: Error finding container b2597c9d398510e988f9e83416b80608d7a6095be571c79ace3b087f2b5f41ec: Status 404 returned error can't find the container with id b2597c9d398510e988f9e83416b80608d7a6095be571c79ace3b087f2b5f41ec Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.284414 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f6658705-1cc5-492a-a2a4-327446c4bcad-service-ca\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.284446 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klvxd\" (UniqueName: \"kubernetes.io/projected/f6658705-1cc5-492a-a2a4-327446c4bcad-kube-api-access-klvxd\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.284481 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f6658705-1cc5-492a-a2a4-327446c4bcad-console-serving-cert\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.284497 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f6658705-1cc5-492a-a2a4-327446c4bcad-oauth-serving-cert\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.284521 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6658705-1cc5-492a-a2a4-327446c4bcad-trusted-ca-bundle\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.284557 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f6658705-1cc5-492a-a2a4-327446c4bcad-console-oauth-config\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.284579 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f6658705-1cc5-492a-a2a4-327446c4bcad-console-config\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.285197 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f6658705-1cc5-492a-a2a4-327446c4bcad-console-config\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.285197 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f6658705-1cc5-492a-a2a4-327446c4bcad-service-ca\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.285849 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f6658705-1cc5-492a-a2a4-327446c4bcad-oauth-serving-cert\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.286094 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6658705-1cc5-492a-a2a4-327446c4bcad-trusted-ca-bundle\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.288738 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f6658705-1cc5-492a-a2a4-327446c4bcad-console-serving-cert\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.289545 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f6658705-1cc5-492a-a2a4-327446c4bcad-console-oauth-config\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.302637 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klvxd\" (UniqueName: \"kubernetes.io/projected/f6658705-1cc5-492a-a2a4-327446c4bcad-kube-api-access-klvxd\") pod \"console-59cc9ddbfc-529ct\" (UID: \"f6658705-1cc5-492a-a2a4-327446c4bcad\") " pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.321550 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj"] Oct 02 01:53:28 crc kubenswrapper[4775]: W1002 01:53:28.338161 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4c70ff6_dec5_4d9e_a7ac_b3f047e99c50.slice/crio-5215b82334c77cd362040f27e0792cdf1e6b2abb4f3584c7fab80a82f55acb18 WatchSource:0}: Error finding container 5215b82334c77cd362040f27e0792cdf1e6b2abb4f3584c7fab80a82f55acb18: Status 404 returned error can't find the container with id 5215b82334c77cd362040f27e0792cdf1e6b2abb4f3584c7fab80a82f55acb18 Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.397317 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv"] Oct 02 01:53:28 crc kubenswrapper[4775]: W1002 01:53:28.408289 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod336c609a_f0e8_42c1_8cd8_bafb09510123.slice/crio-3167eb7933e440f936c34f84bac2e29d606cd097b15976e4901cbc68ba5a3346 WatchSource:0}: Error finding container 3167eb7933e440f936c34f84bac2e29d606cd097b15976e4901cbc68ba5a3346: Status 404 returned error can't find the container with id 3167eb7933e440f936c34f84bac2e29d606cd097b15976e4901cbc68ba5a3346 Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.426210 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.648293 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-59cc9ddbfc-529ct"] Oct 02 01:53:28 crc kubenswrapper[4775]: W1002 01:53:28.653194 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6658705_1cc5_492a_a2a4_327446c4bcad.slice/crio-f1cc16b02a65985d5c7dabd186456723e0bd1eefff63d12a38f71955cfa0783d WatchSource:0}: Error finding container f1cc16b02a65985d5c7dabd186456723e0bd1eefff63d12a38f71955cfa0783d: Status 404 returned error can't find the container with id f1cc16b02a65985d5c7dabd186456723e0bd1eefff63d12a38f71955cfa0783d Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.709981 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj" event={"ID":"b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50","Type":"ContainerStarted","Data":"5215b82334c77cd362040f27e0792cdf1e6b2abb4f3584c7fab80a82f55acb18"} Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.711616 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-59cc9ddbfc-529ct" event={"ID":"f6658705-1cc5-492a-a2a4-327446c4bcad","Type":"ContainerStarted","Data":"f1cc16b02a65985d5c7dabd186456723e0bd1eefff63d12a38f71955cfa0783d"} Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.713237 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv" event={"ID":"336c609a-f0e8-42c1-8cd8-bafb09510123","Type":"ContainerStarted","Data":"3167eb7933e440f936c34f84bac2e29d606cd097b15976e4901cbc68ba5a3346"} Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.715042 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-qf7xb" event={"ID":"69f99de2-12a0-4dbb-9d40-c44352e5859b","Type":"ContainerStarted","Data":"0574e7c27e3a600883b995c443436cea18234a6f9f4f05831ba2bca32a38e1ab"} Oct 02 01:53:28 crc kubenswrapper[4775]: I1002 01:53:28.716431 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xfm7h" event={"ID":"9922a2fe-fb9f-4622-ba18-518eae27300f","Type":"ContainerStarted","Data":"b2597c9d398510e988f9e83416b80608d7a6095be571c79ace3b087f2b5f41ec"} Oct 02 01:53:29 crc kubenswrapper[4775]: I1002 01:53:29.722763 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-59cc9ddbfc-529ct" event={"ID":"f6658705-1cc5-492a-a2a4-327446c4bcad","Type":"ContainerStarted","Data":"a9c6c40f409c764a8d1b85a05e11b812549753956e27c0997214c1c2bcc121fe"} Oct 02 01:53:29 crc kubenswrapper[4775]: I1002 01:53:29.760547 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-59cc9ddbfc-529ct" podStartSLOduration=1.760523141 podStartE2EDuration="1.760523141s" podCreationTimestamp="2025-10-02 01:53:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:53:29.756418225 +0000 UTC m=+746.923162275" watchObservedRunningTime="2025-10-02 01:53:29.760523141 +0000 UTC m=+746.927267191" Oct 02 01:53:30 crc kubenswrapper[4775]: I1002 01:53:30.209752 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rkpgm"] Oct 02 01:53:30 crc kubenswrapper[4775]: I1002 01:53:30.210010 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" podUID="f5887047-3348-478d-a1d5-694762ac0dec" containerName="controller-manager" containerID="cri-o://42121b862c9a4f5b5dfd180b378dfe19328d9c496506c9ce1daee560238b75d2" gracePeriod=30 Oct 02 01:53:30 crc kubenswrapper[4775]: I1002 01:53:30.324832 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n"] Oct 02 01:53:30 crc kubenswrapper[4775]: I1002 01:53:30.325046 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" podUID="b746ad9f-7c63-4945-adce-aa2e0311ba30" containerName="route-controller-manager" containerID="cri-o://273908bae7ad3e2d8c3d077b9b42606c9b7cef5cf0db3ba510318fbe159fc240" gracePeriod=30 Oct 02 01:53:30 crc kubenswrapper[4775]: I1002 01:53:30.743628 4775 generic.go:334] "Generic (PLEG): container finished" podID="b746ad9f-7c63-4945-adce-aa2e0311ba30" containerID="273908bae7ad3e2d8c3d077b9b42606c9b7cef5cf0db3ba510318fbe159fc240" exitCode=0 Oct 02 01:53:30 crc kubenswrapper[4775]: I1002 01:53:30.743717 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" event={"ID":"b746ad9f-7c63-4945-adce-aa2e0311ba30","Type":"ContainerDied","Data":"273908bae7ad3e2d8c3d077b9b42606c9b7cef5cf0db3ba510318fbe159fc240"} Oct 02 01:53:30 crc kubenswrapper[4775]: I1002 01:53:30.745749 4775 generic.go:334] "Generic (PLEG): container finished" podID="f5887047-3348-478d-a1d5-694762ac0dec" containerID="42121b862c9a4f5b5dfd180b378dfe19328d9c496506c9ce1daee560238b75d2" exitCode=0 Oct 02 01:53:30 crc kubenswrapper[4775]: I1002 01:53:30.745778 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" event={"ID":"f5887047-3348-478d-a1d5-694762ac0dec","Type":"ContainerDied","Data":"42121b862c9a4f5b5dfd180b378dfe19328d9c496506c9ce1daee560238b75d2"} Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.657220 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.662297 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.689553 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq"] Oct 02 01:53:31 crc kubenswrapper[4775]: E1002 01:53:31.689770 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b746ad9f-7c63-4945-adce-aa2e0311ba30" containerName="route-controller-manager" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.689789 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b746ad9f-7c63-4945-adce-aa2e0311ba30" containerName="route-controller-manager" Oct 02 01:53:31 crc kubenswrapper[4775]: E1002 01:53:31.689803 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5887047-3348-478d-a1d5-694762ac0dec" containerName="controller-manager" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.689812 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5887047-3348-478d-a1d5-694762ac0dec" containerName="controller-manager" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.689947 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5887047-3348-478d-a1d5-694762ac0dec" containerName="controller-manager" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.689979 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b746ad9f-7c63-4945-adce-aa2e0311ba30" containerName="route-controller-manager" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.690359 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.703161 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq"] Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.762594 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.762738 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-rkpgm" event={"ID":"f5887047-3348-478d-a1d5-694762ac0dec","Type":"ContainerDied","Data":"4fa8a584d1c579889578ba5f50b0476374d34aef04d0ba436b5387285bb6ae2e"} Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.762778 4775 scope.go:117] "RemoveContainer" containerID="42121b862c9a4f5b5dfd180b378dfe19328d9c496506c9ce1daee560238b75d2" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.768392 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.800695 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n" event={"ID":"b746ad9f-7c63-4945-adce-aa2e0311ba30","Type":"ContainerDied","Data":"3980ebabae2123f7cf0036e5ef6635edf5377a716ebaa09ff829c7f2d1128938"} Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.802196 4775 scope.go:117] "RemoveContainer" containerID="273908bae7ad3e2d8c3d077b9b42606c9b7cef5cf0db3ba510318fbe159fc240" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830031 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b746ad9f-7c63-4945-adce-aa2e0311ba30-config\") pod \"b746ad9f-7c63-4945-adce-aa2e0311ba30\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830067 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh5np\" (UniqueName: \"kubernetes.io/projected/f5887047-3348-478d-a1d5-694762ac0dec-kube-api-access-xh5np\") pod \"f5887047-3348-478d-a1d5-694762ac0dec\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830099 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-client-ca\") pod \"f5887047-3348-478d-a1d5-694762ac0dec\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830119 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b746ad9f-7c63-4945-adce-aa2e0311ba30-client-ca\") pod \"b746ad9f-7c63-4945-adce-aa2e0311ba30\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830136 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5887047-3348-478d-a1d5-694762ac0dec-serving-cert\") pod \"f5887047-3348-478d-a1d5-694762ac0dec\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830160 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rlrmc\" (UniqueName: \"kubernetes.io/projected/b746ad9f-7c63-4945-adce-aa2e0311ba30-kube-api-access-rlrmc\") pod \"b746ad9f-7c63-4945-adce-aa2e0311ba30\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830220 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b746ad9f-7c63-4945-adce-aa2e0311ba30-serving-cert\") pod \"b746ad9f-7c63-4945-adce-aa2e0311ba30\" (UID: \"b746ad9f-7c63-4945-adce-aa2e0311ba30\") " Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830241 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-proxy-ca-bundles\") pod \"f5887047-3348-478d-a1d5-694762ac0dec\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830265 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-config\") pod \"f5887047-3348-478d-a1d5-694762ac0dec\" (UID: \"f5887047-3348-478d-a1d5-694762ac0dec\") " Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830422 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10c15beb-69f5-46da-83d3-5f693f2fbd53-client-ca\") pod \"route-controller-manager-5c8b7848c8-59ktq\" (UID: \"10c15beb-69f5-46da-83d3-5f693f2fbd53\") " pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830464 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-654xr\" (UniqueName: \"kubernetes.io/projected/10c15beb-69f5-46da-83d3-5f693f2fbd53-kube-api-access-654xr\") pod \"route-controller-manager-5c8b7848c8-59ktq\" (UID: \"10c15beb-69f5-46da-83d3-5f693f2fbd53\") " pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830484 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10c15beb-69f5-46da-83d3-5f693f2fbd53-serving-cert\") pod \"route-controller-manager-5c8b7848c8-59ktq\" (UID: \"10c15beb-69f5-46da-83d3-5f693f2fbd53\") " pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830505 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10c15beb-69f5-46da-83d3-5f693f2fbd53-config\") pod \"route-controller-manager-5c8b7848c8-59ktq\" (UID: \"10c15beb-69f5-46da-83d3-5f693f2fbd53\") " pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830711 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-client-ca" (OuterVolumeSpecName: "client-ca") pod "f5887047-3348-478d-a1d5-694762ac0dec" (UID: "f5887047-3348-478d-a1d5-694762ac0dec"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.830850 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b746ad9f-7c63-4945-adce-aa2e0311ba30-client-ca" (OuterVolumeSpecName: "client-ca") pod "b746ad9f-7c63-4945-adce-aa2e0311ba30" (UID: "b746ad9f-7c63-4945-adce-aa2e0311ba30"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.831207 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-config" (OuterVolumeSpecName: "config") pod "f5887047-3348-478d-a1d5-694762ac0dec" (UID: "f5887047-3348-478d-a1d5-694762ac0dec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.831292 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b746ad9f-7c63-4945-adce-aa2e0311ba30-config" (OuterVolumeSpecName: "config") pod "b746ad9f-7c63-4945-adce-aa2e0311ba30" (UID: "b746ad9f-7c63-4945-adce-aa2e0311ba30"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.831302 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f5887047-3348-478d-a1d5-694762ac0dec" (UID: "f5887047-3348-478d-a1d5-694762ac0dec"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.834019 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5887047-3348-478d-a1d5-694762ac0dec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f5887047-3348-478d-a1d5-694762ac0dec" (UID: "f5887047-3348-478d-a1d5-694762ac0dec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.834290 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b746ad9f-7c63-4945-adce-aa2e0311ba30-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b746ad9f-7c63-4945-adce-aa2e0311ba30" (UID: "b746ad9f-7c63-4945-adce-aa2e0311ba30"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.834414 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5887047-3348-478d-a1d5-694762ac0dec-kube-api-access-xh5np" (OuterVolumeSpecName: "kube-api-access-xh5np") pod "f5887047-3348-478d-a1d5-694762ac0dec" (UID: "f5887047-3348-478d-a1d5-694762ac0dec"). InnerVolumeSpecName "kube-api-access-xh5np". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.834538 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b746ad9f-7c63-4945-adce-aa2e0311ba30-kube-api-access-rlrmc" (OuterVolumeSpecName: "kube-api-access-rlrmc") pod "b746ad9f-7c63-4945-adce-aa2e0311ba30" (UID: "b746ad9f-7c63-4945-adce-aa2e0311ba30"). InnerVolumeSpecName "kube-api-access-rlrmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.931409 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-654xr\" (UniqueName: \"kubernetes.io/projected/10c15beb-69f5-46da-83d3-5f693f2fbd53-kube-api-access-654xr\") pod \"route-controller-manager-5c8b7848c8-59ktq\" (UID: \"10c15beb-69f5-46da-83d3-5f693f2fbd53\") " pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.931453 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10c15beb-69f5-46da-83d3-5f693f2fbd53-serving-cert\") pod \"route-controller-manager-5c8b7848c8-59ktq\" (UID: \"10c15beb-69f5-46da-83d3-5f693f2fbd53\") " pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.931479 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10c15beb-69f5-46da-83d3-5f693f2fbd53-config\") pod \"route-controller-manager-5c8b7848c8-59ktq\" (UID: \"10c15beb-69f5-46da-83d3-5f693f2fbd53\") " pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.932590 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/10c15beb-69f5-46da-83d3-5f693f2fbd53-config\") pod \"route-controller-manager-5c8b7848c8-59ktq\" (UID: \"10c15beb-69f5-46da-83d3-5f693f2fbd53\") " pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.932654 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10c15beb-69f5-46da-83d3-5f693f2fbd53-client-ca\") pod \"route-controller-manager-5c8b7848c8-59ktq\" (UID: \"10c15beb-69f5-46da-83d3-5f693f2fbd53\") " pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.932703 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.932714 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5887047-3348-478d-a1d5-694762ac0dec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.932723 4775 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b746ad9f-7c63-4945-adce-aa2e0311ba30-client-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.932731 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rlrmc\" (UniqueName: \"kubernetes.io/projected/b746ad9f-7c63-4945-adce-aa2e0311ba30-kube-api-access-rlrmc\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.932741 4775 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b746ad9f-7c63-4945-adce-aa2e0311ba30-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.932749 4775 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.932756 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5887047-3348-478d-a1d5-694762ac0dec-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.932764 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b746ad9f-7c63-4945-adce-aa2e0311ba30-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.932771 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh5np\" (UniqueName: \"kubernetes.io/projected/f5887047-3348-478d-a1d5-694762ac0dec-kube-api-access-xh5np\") on node \"crc\" DevicePath \"\"" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.933342 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/10c15beb-69f5-46da-83d3-5f693f2fbd53-client-ca\") pod \"route-controller-manager-5c8b7848c8-59ktq\" (UID: \"10c15beb-69f5-46da-83d3-5f693f2fbd53\") " pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.944499 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/10c15beb-69f5-46da-83d3-5f693f2fbd53-serving-cert\") pod \"route-controller-manager-5c8b7848c8-59ktq\" (UID: \"10c15beb-69f5-46da-83d3-5f693f2fbd53\") " pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:31 crc kubenswrapper[4775]: I1002 01:53:31.954222 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-654xr\" (UniqueName: \"kubernetes.io/projected/10c15beb-69f5-46da-83d3-5f693f2fbd53-kube-api-access-654xr\") pod \"route-controller-manager-5c8b7848c8-59ktq\" (UID: \"10c15beb-69f5-46da-83d3-5f693f2fbd53\") " pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.013235 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.102636 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n"] Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.108269 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-whx2n"] Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.111624 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rkpgm"] Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.112641 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-rkpgm"] Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.225994 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq"] Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.777393 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj" event={"ID":"b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50","Type":"ContainerStarted","Data":"6b6271e3a5c753949cf51230c3a75ac46dd5ab742d5a6afc2250597bca165fb5"} Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.779028 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj" Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.779364 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" event={"ID":"10c15beb-69f5-46da-83d3-5f693f2fbd53","Type":"ContainerStarted","Data":"32e89a58994e537b86439c41593158ddce00ce0dc77ca8af5a01ed76fb64da39"} Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.779392 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" event={"ID":"10c15beb-69f5-46da-83d3-5f693f2fbd53","Type":"ContainerStarted","Data":"786f49d1818c63146e41865d6df99eeb40e2dbf854a61c6a24f857e4ae6c2689"} Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.781129 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.784403 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv" event={"ID":"336c609a-f0e8-42c1-8cd8-bafb09510123","Type":"ContainerStarted","Data":"bf3ae48fc243ffc4fe77fa08182e730d35bf203403bdf95675e74442941ea293"} Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.785920 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-qf7xb" event={"ID":"69f99de2-12a0-4dbb-9d40-c44352e5859b","Type":"ContainerStarted","Data":"2145648d01effd23782b9078cf41e99cae160a025b648f348a3369889ad2cab4"} Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.786157 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.788223 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xfm7h" event={"ID":"9922a2fe-fb9f-4622-ba18-518eae27300f","Type":"ContainerStarted","Data":"85da4d69c79cea71621200540bc79c8b6ac1fc7c23f48d4341ca28195bdde0ee"} Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.805027 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj" podStartSLOduration=2.463578532 podStartE2EDuration="5.80499037s" podCreationTimestamp="2025-10-02 01:53:27 +0000 UTC" firstStartedPulling="2025-10-02 01:53:28.341060961 +0000 UTC m=+745.507805001" lastFinishedPulling="2025-10-02 01:53:31.682472789 +0000 UTC m=+748.849216839" observedRunningTime="2025-10-02 01:53:32.803248905 +0000 UTC m=+749.969992945" watchObservedRunningTime="2025-10-02 01:53:32.80499037 +0000 UTC m=+749.971734410" Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.852504 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" podStartSLOduration=2.852481955 podStartE2EDuration="2.852481955s" podCreationTimestamp="2025-10-02 01:53:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:53:32.848281267 +0000 UTC m=+750.015025317" watchObservedRunningTime="2025-10-02 01:53:32.852481955 +0000 UTC m=+750.019225995" Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.867771 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-q9jhv" podStartSLOduration=2.624861972 podStartE2EDuration="5.867751979s" podCreationTimestamp="2025-10-02 01:53:27 +0000 UTC" firstStartedPulling="2025-10-02 01:53:28.410151903 +0000 UTC m=+745.576895953" lastFinishedPulling="2025-10-02 01:53:31.65304192 +0000 UTC m=+748.819785960" observedRunningTime="2025-10-02 01:53:32.823794085 +0000 UTC m=+749.990538125" watchObservedRunningTime="2025-10-02 01:53:32.867751979 +0000 UTC m=+750.034496019" Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.872882 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-qf7xb" podStartSLOduration=2.328656763 podStartE2EDuration="5.872865061s" podCreationTimestamp="2025-10-02 01:53:27 +0000 UTC" firstStartedPulling="2025-10-02 01:53:28.143961468 +0000 UTC m=+745.310705508" lastFinishedPulling="2025-10-02 01:53:31.688169756 +0000 UTC m=+748.854913806" observedRunningTime="2025-10-02 01:53:32.871488895 +0000 UTC m=+750.038232945" watchObservedRunningTime="2025-10-02 01:53:32.872865061 +0000 UTC m=+750.039609111" Oct 02 01:53:32 crc kubenswrapper[4775]: I1002 01:53:32.879672 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5c8b7848c8-59ktq" Oct 02 01:53:33 crc kubenswrapper[4775]: I1002 01:53:33.775229 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b746ad9f-7c63-4945-adce-aa2e0311ba30" path="/var/lib/kubelet/pods/b746ad9f-7c63-4945-adce-aa2e0311ba30/volumes" Oct 02 01:53:33 crc kubenswrapper[4775]: I1002 01:53:33.776078 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5887047-3348-478d-a1d5-694762ac0dec" path="/var/lib/kubelet/pods/f5887047-3348-478d-a1d5-694762ac0dec/volumes" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.323897 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5d8cf9f578-5864l"] Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.324918 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.327734 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.327927 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.328160 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.328332 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.328660 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.329240 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.339618 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d8cf9f578-5864l"] Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.342483 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.488275 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-config\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.488467 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-serving-cert\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.488703 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-client-ca\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.488764 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g28j\" (UniqueName: \"kubernetes.io/projected/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-kube-api-access-6g28j\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.488818 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-proxy-ca-bundles\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.589750 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g28j\" (UniqueName: \"kubernetes.io/projected/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-kube-api-access-6g28j\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.589847 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-proxy-ca-bundles\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.589922 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-config\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.589999 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-serving-cert\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.591350 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-client-ca\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.591427 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-proxy-ca-bundles\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.592411 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-config\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.592992 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-client-ca\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.598983 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-serving-cert\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.613383 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g28j\" (UniqueName: \"kubernetes.io/projected/cf9bd982-d4f9-4763-99ad-202d4dc4bac2-kube-api-access-6g28j\") pod \"controller-manager-5d8cf9f578-5864l\" (UID: \"cf9bd982-d4f9-4763-99ad-202d4dc4bac2\") " pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:34 crc kubenswrapper[4775]: I1002 01:53:34.655386 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:35 crc kubenswrapper[4775]: I1002 01:53:35.274243 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d8cf9f578-5864l"] Oct 02 01:53:35 crc kubenswrapper[4775]: W1002 01:53:35.275664 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf9bd982_d4f9_4763_99ad_202d4dc4bac2.slice/crio-cbfff151a70c1fb432129005c1f95a9c760c8011bbe1dba09bf90563fd9dbdbc WatchSource:0}: Error finding container cbfff151a70c1fb432129005c1f95a9c760c8011bbe1dba09bf90563fd9dbdbc: Status 404 returned error can't find the container with id cbfff151a70c1fb432129005c1f95a9c760c8011bbe1dba09bf90563fd9dbdbc Oct 02 01:53:35 crc kubenswrapper[4775]: I1002 01:53:35.808153 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" event={"ID":"cf9bd982-d4f9-4763-99ad-202d4dc4bac2","Type":"ContainerStarted","Data":"d489981041e06bc541804259365bb41821cb1f830d534832d4b902511b55b6b4"} Oct 02 01:53:35 crc kubenswrapper[4775]: I1002 01:53:35.808192 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" event={"ID":"cf9bd982-d4f9-4763-99ad-202d4dc4bac2","Type":"ContainerStarted","Data":"cbfff151a70c1fb432129005c1f95a9c760c8011bbe1dba09bf90563fd9dbdbc"} Oct 02 01:53:35 crc kubenswrapper[4775]: I1002 01:53:35.808326 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:35 crc kubenswrapper[4775]: I1002 01:53:35.809989 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xfm7h" event={"ID":"9922a2fe-fb9f-4622-ba18-518eae27300f","Type":"ContainerStarted","Data":"ecc5f0cdbce261c2414be7a7d5b45e370245e07133328dde427f21ac259cf7b5"} Oct 02 01:53:35 crc kubenswrapper[4775]: I1002 01:53:35.813897 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" Oct 02 01:53:35 crc kubenswrapper[4775]: I1002 01:53:35.827263 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5d8cf9f578-5864l" podStartSLOduration=5.827242107 podStartE2EDuration="5.827242107s" podCreationTimestamp="2025-10-02 01:53:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:53:35.822045283 +0000 UTC m=+752.988789343" watchObservedRunningTime="2025-10-02 01:53:35.827242107 +0000 UTC m=+752.993986187" Oct 02 01:53:35 crc kubenswrapper[4775]: I1002 01:53:35.861184 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xfm7h" podStartSLOduration=2.205791373 podStartE2EDuration="8.861168682s" podCreationTimestamp="2025-10-02 01:53:27 +0000 UTC" firstStartedPulling="2025-10-02 01:53:28.284842281 +0000 UTC m=+745.451586321" lastFinishedPulling="2025-10-02 01:53:34.94021958 +0000 UTC m=+752.106963630" observedRunningTime="2025-10-02 01:53:35.859070408 +0000 UTC m=+753.025814478" watchObservedRunningTime="2025-10-02 01:53:35.861168682 +0000 UTC m=+753.027912722" Oct 02 01:53:37 crc kubenswrapper[4775]: I1002 01:53:37.233771 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:53:37 crc kubenswrapper[4775]: I1002 01:53:37.234224 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:53:38 crc kubenswrapper[4775]: I1002 01:53:38.155458 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-qf7xb" Oct 02 01:53:38 crc kubenswrapper[4775]: I1002 01:53:38.426694 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:38 crc kubenswrapper[4775]: I1002 01:53:38.426772 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:38 crc kubenswrapper[4775]: I1002 01:53:38.434333 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:38 crc kubenswrapper[4775]: I1002 01:53:38.836436 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-59cc9ddbfc-529ct" Oct 02 01:53:38 crc kubenswrapper[4775]: I1002 01:53:38.901183 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-mgrh5"] Oct 02 01:53:39 crc kubenswrapper[4775]: I1002 01:53:39.132470 4775 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 02 01:53:48 crc kubenswrapper[4775]: I1002 01:53:48.099529 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-g5pnj" Oct 02 01:54:03 crc kubenswrapper[4775]: I1002 01:54:03.979702 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-mgrh5" podUID="c1a3714d-b7af-49ce-9ed7-64b75dd94ce8" containerName="console" containerID="cri-o://68d0c1e399050ecaed8b9a92889f547ea07e05b90af67879292d4c0f53ba71a1" gracePeriod=15 Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.509706 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-mgrh5_c1a3714d-b7af-49ce-9ed7-64b75dd94ce8/console/0.log" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.509982 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.638348 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-oauth-config\") pod \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.638733 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-trusted-ca-bundle\") pod \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.639297 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "c1a3714d-b7af-49ce-9ed7-64b75dd94ce8" (UID: "c1a3714d-b7af-49ce-9ed7-64b75dd94ce8"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.639425 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-serving-cert\") pod \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.639842 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-oauth-serving-cert\") pod \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.639889 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4xck\" (UniqueName: \"kubernetes.io/projected/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-kube-api-access-g4xck\") pod \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.639938 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-config\") pod \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.639979 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-service-ca\") pod \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\" (UID: \"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8\") " Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.640283 4775 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.641217 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-config" (OuterVolumeSpecName: "console-config") pod "c1a3714d-b7af-49ce-9ed7-64b75dd94ce8" (UID: "c1a3714d-b7af-49ce-9ed7-64b75dd94ce8"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.641225 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "c1a3714d-b7af-49ce-9ed7-64b75dd94ce8" (UID: "c1a3714d-b7af-49ce-9ed7-64b75dd94ce8"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.641239 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-service-ca" (OuterVolumeSpecName: "service-ca") pod "c1a3714d-b7af-49ce-9ed7-64b75dd94ce8" (UID: "c1a3714d-b7af-49ce-9ed7-64b75dd94ce8"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.644574 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "c1a3714d-b7af-49ce-9ed7-64b75dd94ce8" (UID: "c1a3714d-b7af-49ce-9ed7-64b75dd94ce8"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.645213 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "c1a3714d-b7af-49ce-9ed7-64b75dd94ce8" (UID: "c1a3714d-b7af-49ce-9ed7-64b75dd94ce8"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.645570 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-kube-api-access-g4xck" (OuterVolumeSpecName: "kube-api-access-g4xck") pod "c1a3714d-b7af-49ce-9ed7-64b75dd94ce8" (UID: "c1a3714d-b7af-49ce-9ed7-64b75dd94ce8"). InnerVolumeSpecName "kube-api-access-g4xck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.741601 4775 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.741632 4775 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.741644 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4xck\" (UniqueName: \"kubernetes.io/projected/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-kube-api-access-g4xck\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.741659 4775 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.741672 4775 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-service-ca\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:04 crc kubenswrapper[4775]: I1002 01:54:04.741684 4775 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.048376 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-mgrh5_c1a3714d-b7af-49ce-9ed7-64b75dd94ce8/console/0.log" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.048450 4775 generic.go:334] "Generic (PLEG): container finished" podID="c1a3714d-b7af-49ce-9ed7-64b75dd94ce8" containerID="68d0c1e399050ecaed8b9a92889f547ea07e05b90af67879292d4c0f53ba71a1" exitCode=2 Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.048491 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mgrh5" event={"ID":"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8","Type":"ContainerDied","Data":"68d0c1e399050ecaed8b9a92889f547ea07e05b90af67879292d4c0f53ba71a1"} Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.048527 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mgrh5" event={"ID":"c1a3714d-b7af-49ce-9ed7-64b75dd94ce8","Type":"ContainerDied","Data":"44074d7cde6ac69e469c678e6d3e861ab6256f1461bbb9b6ed65bb1447519f8e"} Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.048554 4775 scope.go:117] "RemoveContainer" containerID="68d0c1e399050ecaed8b9a92889f547ea07e05b90af67879292d4c0f53ba71a1" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.048711 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mgrh5" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.067375 4775 scope.go:117] "RemoveContainer" containerID="68d0c1e399050ecaed8b9a92889f547ea07e05b90af67879292d4c0f53ba71a1" Oct 02 01:54:05 crc kubenswrapper[4775]: E1002 01:54:05.067750 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68d0c1e399050ecaed8b9a92889f547ea07e05b90af67879292d4c0f53ba71a1\": container with ID starting with 68d0c1e399050ecaed8b9a92889f547ea07e05b90af67879292d4c0f53ba71a1 not found: ID does not exist" containerID="68d0c1e399050ecaed8b9a92889f547ea07e05b90af67879292d4c0f53ba71a1" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.067786 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68d0c1e399050ecaed8b9a92889f547ea07e05b90af67879292d4c0f53ba71a1"} err="failed to get container status \"68d0c1e399050ecaed8b9a92889f547ea07e05b90af67879292d4c0f53ba71a1\": rpc error: code = NotFound desc = could not find container \"68d0c1e399050ecaed8b9a92889f547ea07e05b90af67879292d4c0f53ba71a1\": container with ID starting with 68d0c1e399050ecaed8b9a92889f547ea07e05b90af67879292d4c0f53ba71a1 not found: ID does not exist" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.091188 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-mgrh5"] Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.098706 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-mgrh5"] Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.153873 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rvpz5"] Oct 02 01:54:05 crc kubenswrapper[4775]: E1002 01:54:05.154148 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1a3714d-b7af-49ce-9ed7-64b75dd94ce8" containerName="console" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.154164 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1a3714d-b7af-49ce-9ed7-64b75dd94ce8" containerName="console" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.154307 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1a3714d-b7af-49ce-9ed7-64b75dd94ce8" containerName="console" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.155252 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.165836 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rvpz5"] Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.247872 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b212cd7-598e-43b2-b2c0-de3e89358cde-utilities\") pod \"certified-operators-rvpz5\" (UID: \"6b212cd7-598e-43b2-b2c0-de3e89358cde\") " pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.247979 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsrlb\" (UniqueName: \"kubernetes.io/projected/6b212cd7-598e-43b2-b2c0-de3e89358cde-kube-api-access-jsrlb\") pod \"certified-operators-rvpz5\" (UID: \"6b212cd7-598e-43b2-b2c0-de3e89358cde\") " pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.248022 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b212cd7-598e-43b2-b2c0-de3e89358cde-catalog-content\") pod \"certified-operators-rvpz5\" (UID: \"6b212cd7-598e-43b2-b2c0-de3e89358cde\") " pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.349641 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsrlb\" (UniqueName: \"kubernetes.io/projected/6b212cd7-598e-43b2-b2c0-de3e89358cde-kube-api-access-jsrlb\") pod \"certified-operators-rvpz5\" (UID: \"6b212cd7-598e-43b2-b2c0-de3e89358cde\") " pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.349734 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b212cd7-598e-43b2-b2c0-de3e89358cde-catalog-content\") pod \"certified-operators-rvpz5\" (UID: \"6b212cd7-598e-43b2-b2c0-de3e89358cde\") " pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.349772 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b212cd7-598e-43b2-b2c0-de3e89358cde-utilities\") pod \"certified-operators-rvpz5\" (UID: \"6b212cd7-598e-43b2-b2c0-de3e89358cde\") " pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.350376 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b212cd7-598e-43b2-b2c0-de3e89358cde-utilities\") pod \"certified-operators-rvpz5\" (UID: \"6b212cd7-598e-43b2-b2c0-de3e89358cde\") " pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.351037 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b212cd7-598e-43b2-b2c0-de3e89358cde-catalog-content\") pod \"certified-operators-rvpz5\" (UID: \"6b212cd7-598e-43b2-b2c0-de3e89358cde\") " pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.378511 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsrlb\" (UniqueName: \"kubernetes.io/projected/6b212cd7-598e-43b2-b2c0-de3e89358cde-kube-api-access-jsrlb\") pod \"certified-operators-rvpz5\" (UID: \"6b212cd7-598e-43b2-b2c0-de3e89358cde\") " pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.480420 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.771183 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1a3714d-b7af-49ce-9ed7-64b75dd94ce8" path="/var/lib/kubelet/pods/c1a3714d-b7af-49ce-9ed7-64b75dd94ce8/volumes" Oct 02 01:54:05 crc kubenswrapper[4775]: I1002 01:54:05.944089 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rvpz5"] Oct 02 01:54:05 crc kubenswrapper[4775]: W1002 01:54:05.949512 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b212cd7_598e_43b2_b2c0_de3e89358cde.slice/crio-92ad2cc4be7ab97cd5d81a7287ce77d1cfc32a85c98ff75dc8b481339cb0613c WatchSource:0}: Error finding container 92ad2cc4be7ab97cd5d81a7287ce77d1cfc32a85c98ff75dc8b481339cb0613c: Status 404 returned error can't find the container with id 92ad2cc4be7ab97cd5d81a7287ce77d1cfc32a85c98ff75dc8b481339cb0613c Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.056792 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rvpz5" event={"ID":"6b212cd7-598e-43b2-b2c0-de3e89358cde","Type":"ContainerStarted","Data":"92ad2cc4be7ab97cd5d81a7287ce77d1cfc32a85c98ff75dc8b481339cb0613c"} Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.169994 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l"] Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.171207 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.177651 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.194798 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l"] Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.269792 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kd77q\" (UniqueName: \"kubernetes.io/projected/bb58aabc-fee6-480d-8d28-d7ebf0506266-kube-api-access-kd77q\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l\" (UID: \"bb58aabc-fee6-480d-8d28-d7ebf0506266\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.269883 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb58aabc-fee6-480d-8d28-d7ebf0506266-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l\" (UID: \"bb58aabc-fee6-480d-8d28-d7ebf0506266\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.270149 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb58aabc-fee6-480d-8d28-d7ebf0506266-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l\" (UID: \"bb58aabc-fee6-480d-8d28-d7ebf0506266\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.371942 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb58aabc-fee6-480d-8d28-d7ebf0506266-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l\" (UID: \"bb58aabc-fee6-480d-8d28-d7ebf0506266\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.372085 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kd77q\" (UniqueName: \"kubernetes.io/projected/bb58aabc-fee6-480d-8d28-d7ebf0506266-kube-api-access-kd77q\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l\" (UID: \"bb58aabc-fee6-480d-8d28-d7ebf0506266\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.372143 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb58aabc-fee6-480d-8d28-d7ebf0506266-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l\" (UID: \"bb58aabc-fee6-480d-8d28-d7ebf0506266\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.372793 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb58aabc-fee6-480d-8d28-d7ebf0506266-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l\" (UID: \"bb58aabc-fee6-480d-8d28-d7ebf0506266\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.372892 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb58aabc-fee6-480d-8d28-d7ebf0506266-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l\" (UID: \"bb58aabc-fee6-480d-8d28-d7ebf0506266\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.406286 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kd77q\" (UniqueName: \"kubernetes.io/projected/bb58aabc-fee6-480d-8d28-d7ebf0506266-kube-api-access-kd77q\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l\" (UID: \"bb58aabc-fee6-480d-8d28-d7ebf0506266\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" Oct 02 01:54:06 crc kubenswrapper[4775]: I1002 01:54:06.570550 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" Oct 02 01:54:07 crc kubenswrapper[4775]: I1002 01:54:07.082187 4775 generic.go:334] "Generic (PLEG): container finished" podID="6b212cd7-598e-43b2-b2c0-de3e89358cde" containerID="64ab50e35efaef6dd952d0de82b6860917df599072f66bea86f9ffcbd65b8b55" exitCode=0 Oct 02 01:54:07 crc kubenswrapper[4775]: I1002 01:54:07.082432 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rvpz5" event={"ID":"6b212cd7-598e-43b2-b2c0-de3e89358cde","Type":"ContainerDied","Data":"64ab50e35efaef6dd952d0de82b6860917df599072f66bea86f9ffcbd65b8b55"} Oct 02 01:54:07 crc kubenswrapper[4775]: I1002 01:54:07.085738 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l"] Oct 02 01:54:07 crc kubenswrapper[4775]: W1002 01:54:07.099711 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb58aabc_fee6_480d_8d28_d7ebf0506266.slice/crio-845cb221490123b99b159682636169e3cdd7008aee010c4c04b8e2e3a94783b2 WatchSource:0}: Error finding container 845cb221490123b99b159682636169e3cdd7008aee010c4c04b8e2e3a94783b2: Status 404 returned error can't find the container with id 845cb221490123b99b159682636169e3cdd7008aee010c4c04b8e2e3a94783b2 Oct 02 01:54:07 crc kubenswrapper[4775]: I1002 01:54:07.233669 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:54:07 crc kubenswrapper[4775]: I1002 01:54:07.233734 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:54:07 crc kubenswrapper[4775]: I1002 01:54:07.233780 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:54:07 crc kubenswrapper[4775]: I1002 01:54:07.234552 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0af4ec8f881f09471035912f64ffe17497158c6365d4ee31b994b4425f60c6bb"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 01:54:07 crc kubenswrapper[4775]: I1002 01:54:07.234638 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://0af4ec8f881f09471035912f64ffe17497158c6365d4ee31b994b4425f60c6bb" gracePeriod=600 Oct 02 01:54:08 crc kubenswrapper[4775]: I1002 01:54:08.098987 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="0af4ec8f881f09471035912f64ffe17497158c6365d4ee31b994b4425f60c6bb" exitCode=0 Oct 02 01:54:08 crc kubenswrapper[4775]: I1002 01:54:08.099529 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"0af4ec8f881f09471035912f64ffe17497158c6365d4ee31b994b4425f60c6bb"} Oct 02 01:54:08 crc kubenswrapper[4775]: I1002 01:54:08.099717 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"8d8df09487b33326d5ae12f2261c6dc85c52809ba9bfd51eac55d13c8fc891e6"} Oct 02 01:54:08 crc kubenswrapper[4775]: I1002 01:54:08.099755 4775 scope.go:117] "RemoveContainer" containerID="2c48cbf3d08cb4a2f350c67fef84e9504d7f8966e661be9b0356695ca1016ee7" Oct 02 01:54:08 crc kubenswrapper[4775]: I1002 01:54:08.112191 4775 generic.go:334] "Generic (PLEG): container finished" podID="bb58aabc-fee6-480d-8d28-d7ebf0506266" containerID="05524d8bb159fe9d29a333319c547fcc622166fe1c8ba9bc39e15e8c0a88720a" exitCode=0 Oct 02 01:54:08 crc kubenswrapper[4775]: I1002 01:54:08.112253 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" event={"ID":"bb58aabc-fee6-480d-8d28-d7ebf0506266","Type":"ContainerDied","Data":"05524d8bb159fe9d29a333319c547fcc622166fe1c8ba9bc39e15e8c0a88720a"} Oct 02 01:54:08 crc kubenswrapper[4775]: I1002 01:54:08.112299 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" event={"ID":"bb58aabc-fee6-480d-8d28-d7ebf0506266","Type":"ContainerStarted","Data":"845cb221490123b99b159682636169e3cdd7008aee010c4c04b8e2e3a94783b2"} Oct 02 01:54:09 crc kubenswrapper[4775]: I1002 01:54:09.124161 4775 generic.go:334] "Generic (PLEG): container finished" podID="6b212cd7-598e-43b2-b2c0-de3e89358cde" containerID="fa0b4c826756674926bb52677e6b8069c60a1ccdcba60fdc04da84badb9425b0" exitCode=0 Oct 02 01:54:09 crc kubenswrapper[4775]: I1002 01:54:09.124265 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rvpz5" event={"ID":"6b212cd7-598e-43b2-b2c0-de3e89358cde","Type":"ContainerDied","Data":"fa0b4c826756674926bb52677e6b8069c60a1ccdcba60fdc04da84badb9425b0"} Oct 02 01:54:09 crc kubenswrapper[4775]: I1002 01:54:09.699115 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-b6749"] Oct 02 01:54:09 crc kubenswrapper[4775]: I1002 01:54:09.700535 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:09 crc kubenswrapper[4775]: I1002 01:54:09.721100 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b6749"] Oct 02 01:54:09 crc kubenswrapper[4775]: I1002 01:54:09.820174 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9cch\" (UniqueName: \"kubernetes.io/projected/d520a768-2d91-45f0-a1d5-825b2c03853f-kube-api-access-s9cch\") pod \"redhat-operators-b6749\" (UID: \"d520a768-2d91-45f0-a1d5-825b2c03853f\") " pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:09 crc kubenswrapper[4775]: I1002 01:54:09.820755 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d520a768-2d91-45f0-a1d5-825b2c03853f-utilities\") pod \"redhat-operators-b6749\" (UID: \"d520a768-2d91-45f0-a1d5-825b2c03853f\") " pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:09 crc kubenswrapper[4775]: I1002 01:54:09.820814 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d520a768-2d91-45f0-a1d5-825b2c03853f-catalog-content\") pod \"redhat-operators-b6749\" (UID: \"d520a768-2d91-45f0-a1d5-825b2c03853f\") " pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:09 crc kubenswrapper[4775]: I1002 01:54:09.922390 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d520a768-2d91-45f0-a1d5-825b2c03853f-catalog-content\") pod \"redhat-operators-b6749\" (UID: \"d520a768-2d91-45f0-a1d5-825b2c03853f\") " pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:09 crc kubenswrapper[4775]: I1002 01:54:09.922460 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9cch\" (UniqueName: \"kubernetes.io/projected/d520a768-2d91-45f0-a1d5-825b2c03853f-kube-api-access-s9cch\") pod \"redhat-operators-b6749\" (UID: \"d520a768-2d91-45f0-a1d5-825b2c03853f\") " pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:09 crc kubenswrapper[4775]: I1002 01:54:09.922548 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d520a768-2d91-45f0-a1d5-825b2c03853f-utilities\") pod \"redhat-operators-b6749\" (UID: \"d520a768-2d91-45f0-a1d5-825b2c03853f\") " pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:09 crc kubenswrapper[4775]: I1002 01:54:09.922885 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d520a768-2d91-45f0-a1d5-825b2c03853f-catalog-content\") pod \"redhat-operators-b6749\" (UID: \"d520a768-2d91-45f0-a1d5-825b2c03853f\") " pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:09 crc kubenswrapper[4775]: I1002 01:54:09.922900 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d520a768-2d91-45f0-a1d5-825b2c03853f-utilities\") pod \"redhat-operators-b6749\" (UID: \"d520a768-2d91-45f0-a1d5-825b2c03853f\") " pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:09 crc kubenswrapper[4775]: I1002 01:54:09.943144 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9cch\" (UniqueName: \"kubernetes.io/projected/d520a768-2d91-45f0-a1d5-825b2c03853f-kube-api-access-s9cch\") pod \"redhat-operators-b6749\" (UID: \"d520a768-2d91-45f0-a1d5-825b2c03853f\") " pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:10 crc kubenswrapper[4775]: I1002 01:54:10.019492 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:10 crc kubenswrapper[4775]: I1002 01:54:10.149030 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rvpz5" event={"ID":"6b212cd7-598e-43b2-b2c0-de3e89358cde","Type":"ContainerStarted","Data":"2f6dec9987abc1444e3ad9103693e8dc153e7d7437fd8f8b26d5d7b4741f1f63"} Oct 02 01:54:10 crc kubenswrapper[4775]: I1002 01:54:10.176643 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rvpz5" podStartSLOduration=2.706674156 podStartE2EDuration="5.176627701s" podCreationTimestamp="2025-10-02 01:54:05 +0000 UTC" firstStartedPulling="2025-10-02 01:54:07.085572887 +0000 UTC m=+784.252316967" lastFinishedPulling="2025-10-02 01:54:09.555526462 +0000 UTC m=+786.722270512" observedRunningTime="2025-10-02 01:54:10.17410072 +0000 UTC m=+787.340844760" watchObservedRunningTime="2025-10-02 01:54:10.176627701 +0000 UTC m=+787.343371741" Oct 02 01:54:10 crc kubenswrapper[4775]: I1002 01:54:10.435175 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-b6749"] Oct 02 01:54:11 crc kubenswrapper[4775]: I1002 01:54:11.155521 4775 generic.go:334] "Generic (PLEG): container finished" podID="d520a768-2d91-45f0-a1d5-825b2c03853f" containerID="4205b4f4a2be9423a16ea8258f9e206d92f602e28407b6a9d88f2d773d13e1c5" exitCode=0 Oct 02 01:54:11 crc kubenswrapper[4775]: I1002 01:54:11.155624 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6749" event={"ID":"d520a768-2d91-45f0-a1d5-825b2c03853f","Type":"ContainerDied","Data":"4205b4f4a2be9423a16ea8258f9e206d92f602e28407b6a9d88f2d773d13e1c5"} Oct 02 01:54:11 crc kubenswrapper[4775]: I1002 01:54:11.156226 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6749" event={"ID":"d520a768-2d91-45f0-a1d5-825b2c03853f","Type":"ContainerStarted","Data":"eece375a8320b0b9bef235489168df836433fe74f008b208a9f55dd383238b80"} Oct 02 01:54:11 crc kubenswrapper[4775]: I1002 01:54:11.159702 4775 generic.go:334] "Generic (PLEG): container finished" podID="bb58aabc-fee6-480d-8d28-d7ebf0506266" containerID="344570e63f3a1bb08860a67f02fb7ba9ece41820171508f27e571d46ba13d992" exitCode=0 Oct 02 01:54:11 crc kubenswrapper[4775]: I1002 01:54:11.159810 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" event={"ID":"bb58aabc-fee6-480d-8d28-d7ebf0506266","Type":"ContainerDied","Data":"344570e63f3a1bb08860a67f02fb7ba9ece41820171508f27e571d46ba13d992"} Oct 02 01:54:12 crc kubenswrapper[4775]: I1002 01:54:12.168566 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6749" event={"ID":"d520a768-2d91-45f0-a1d5-825b2c03853f","Type":"ContainerStarted","Data":"f68c1eb2f8157eab2be8ad54a5121b589f40e622aeca3fa4e588dd8ff476ccef"} Oct 02 01:54:12 crc kubenswrapper[4775]: I1002 01:54:12.173356 4775 generic.go:334] "Generic (PLEG): container finished" podID="bb58aabc-fee6-480d-8d28-d7ebf0506266" containerID="d6b6683cebb8e8dd065285c08165bb64051dfcfd93aff4606030c517b9cd098d" exitCode=0 Oct 02 01:54:12 crc kubenswrapper[4775]: I1002 01:54:12.173435 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" event={"ID":"bb58aabc-fee6-480d-8d28-d7ebf0506266","Type":"ContainerDied","Data":"d6b6683cebb8e8dd065285c08165bb64051dfcfd93aff4606030c517b9cd098d"} Oct 02 01:54:13 crc kubenswrapper[4775]: I1002 01:54:13.185749 4775 generic.go:334] "Generic (PLEG): container finished" podID="d520a768-2d91-45f0-a1d5-825b2c03853f" containerID="f68c1eb2f8157eab2be8ad54a5121b589f40e622aeca3fa4e588dd8ff476ccef" exitCode=0 Oct 02 01:54:13 crc kubenswrapper[4775]: I1002 01:54:13.185886 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6749" event={"ID":"d520a768-2d91-45f0-a1d5-825b2c03853f","Type":"ContainerDied","Data":"f68c1eb2f8157eab2be8ad54a5121b589f40e622aeca3fa4e588dd8ff476ccef"} Oct 02 01:54:13 crc kubenswrapper[4775]: I1002 01:54:13.565038 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" Oct 02 01:54:13 crc kubenswrapper[4775]: I1002 01:54:13.679285 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb58aabc-fee6-480d-8d28-d7ebf0506266-bundle\") pod \"bb58aabc-fee6-480d-8d28-d7ebf0506266\" (UID: \"bb58aabc-fee6-480d-8d28-d7ebf0506266\") " Oct 02 01:54:13 crc kubenswrapper[4775]: I1002 01:54:13.679406 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb58aabc-fee6-480d-8d28-d7ebf0506266-util\") pod \"bb58aabc-fee6-480d-8d28-d7ebf0506266\" (UID: \"bb58aabc-fee6-480d-8d28-d7ebf0506266\") " Oct 02 01:54:13 crc kubenswrapper[4775]: I1002 01:54:13.679630 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kd77q\" (UniqueName: \"kubernetes.io/projected/bb58aabc-fee6-480d-8d28-d7ebf0506266-kube-api-access-kd77q\") pod \"bb58aabc-fee6-480d-8d28-d7ebf0506266\" (UID: \"bb58aabc-fee6-480d-8d28-d7ebf0506266\") " Oct 02 01:54:13 crc kubenswrapper[4775]: I1002 01:54:13.681534 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb58aabc-fee6-480d-8d28-d7ebf0506266-bundle" (OuterVolumeSpecName: "bundle") pod "bb58aabc-fee6-480d-8d28-d7ebf0506266" (UID: "bb58aabc-fee6-480d-8d28-d7ebf0506266"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:54:13 crc kubenswrapper[4775]: I1002 01:54:13.690691 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb58aabc-fee6-480d-8d28-d7ebf0506266-kube-api-access-kd77q" (OuterVolumeSpecName: "kube-api-access-kd77q") pod "bb58aabc-fee6-480d-8d28-d7ebf0506266" (UID: "bb58aabc-fee6-480d-8d28-d7ebf0506266"). InnerVolumeSpecName "kube-api-access-kd77q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:54:13 crc kubenswrapper[4775]: I1002 01:54:13.695431 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb58aabc-fee6-480d-8d28-d7ebf0506266-util" (OuterVolumeSpecName: "util") pod "bb58aabc-fee6-480d-8d28-d7ebf0506266" (UID: "bb58aabc-fee6-480d-8d28-d7ebf0506266"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:54:13 crc kubenswrapper[4775]: I1002 01:54:13.781005 4775 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb58aabc-fee6-480d-8d28-d7ebf0506266-util\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:13 crc kubenswrapper[4775]: I1002 01:54:13.781047 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kd77q\" (UniqueName: \"kubernetes.io/projected/bb58aabc-fee6-480d-8d28-d7ebf0506266-kube-api-access-kd77q\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:13 crc kubenswrapper[4775]: I1002 01:54:13.781072 4775 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb58aabc-fee6-480d-8d28-d7ebf0506266-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:14 crc kubenswrapper[4775]: I1002 01:54:14.195643 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" event={"ID":"bb58aabc-fee6-480d-8d28-d7ebf0506266","Type":"ContainerDied","Data":"845cb221490123b99b159682636169e3cdd7008aee010c4c04b8e2e3a94783b2"} Oct 02 01:54:14 crc kubenswrapper[4775]: I1002 01:54:14.195679 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l" Oct 02 01:54:14 crc kubenswrapper[4775]: I1002 01:54:14.195699 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="845cb221490123b99b159682636169e3cdd7008aee010c4c04b8e2e3a94783b2" Oct 02 01:54:14 crc kubenswrapper[4775]: I1002 01:54:14.198298 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6749" event={"ID":"d520a768-2d91-45f0-a1d5-825b2c03853f","Type":"ContainerStarted","Data":"9b667b9e0f6367a0f4bb14f671271a59a2d865e6ede2eb8e304205cfee3b1f0c"} Oct 02 01:54:14 crc kubenswrapper[4775]: I1002 01:54:14.230110 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-b6749" podStartSLOduration=2.7811916180000003 podStartE2EDuration="5.230075389s" podCreationTimestamp="2025-10-02 01:54:09 +0000 UTC" firstStartedPulling="2025-10-02 01:54:11.157835323 +0000 UTC m=+788.324579403" lastFinishedPulling="2025-10-02 01:54:13.606719134 +0000 UTC m=+790.773463174" observedRunningTime="2025-10-02 01:54:14.223744745 +0000 UTC m=+791.390488845" watchObservedRunningTime="2025-10-02 01:54:14.230075389 +0000 UTC m=+791.396819429" Oct 02 01:54:15 crc kubenswrapper[4775]: I1002 01:54:15.481548 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:15 crc kubenswrapper[4775]: I1002 01:54:15.481984 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:15 crc kubenswrapper[4775]: I1002 01:54:15.551187 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:16 crc kubenswrapper[4775]: I1002 01:54:16.272114 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:18 crc kubenswrapper[4775]: I1002 01:54:18.686058 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rvpz5"] Oct 02 01:54:19 crc kubenswrapper[4775]: I1002 01:54:19.253506 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rvpz5" podUID="6b212cd7-598e-43b2-b2c0-de3e89358cde" containerName="registry-server" containerID="cri-o://2f6dec9987abc1444e3ad9103693e8dc153e7d7437fd8f8b26d5d7b4741f1f63" gracePeriod=2 Oct 02 01:54:19 crc kubenswrapper[4775]: I1002 01:54:19.702520 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:19 crc kubenswrapper[4775]: I1002 01:54:19.764085 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsrlb\" (UniqueName: \"kubernetes.io/projected/6b212cd7-598e-43b2-b2c0-de3e89358cde-kube-api-access-jsrlb\") pod \"6b212cd7-598e-43b2-b2c0-de3e89358cde\" (UID: \"6b212cd7-598e-43b2-b2c0-de3e89358cde\") " Oct 02 01:54:19 crc kubenswrapper[4775]: I1002 01:54:19.764149 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b212cd7-598e-43b2-b2c0-de3e89358cde-catalog-content\") pod \"6b212cd7-598e-43b2-b2c0-de3e89358cde\" (UID: \"6b212cd7-598e-43b2-b2c0-de3e89358cde\") " Oct 02 01:54:19 crc kubenswrapper[4775]: I1002 01:54:19.764220 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b212cd7-598e-43b2-b2c0-de3e89358cde-utilities\") pod \"6b212cd7-598e-43b2-b2c0-de3e89358cde\" (UID: \"6b212cd7-598e-43b2-b2c0-de3e89358cde\") " Oct 02 01:54:19 crc kubenswrapper[4775]: I1002 01:54:19.765171 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b212cd7-598e-43b2-b2c0-de3e89358cde-utilities" (OuterVolumeSpecName: "utilities") pod "6b212cd7-598e-43b2-b2c0-de3e89358cde" (UID: "6b212cd7-598e-43b2-b2c0-de3e89358cde"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:54:19 crc kubenswrapper[4775]: I1002 01:54:19.775383 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b212cd7-598e-43b2-b2c0-de3e89358cde-kube-api-access-jsrlb" (OuterVolumeSpecName: "kube-api-access-jsrlb") pod "6b212cd7-598e-43b2-b2c0-de3e89358cde" (UID: "6b212cd7-598e-43b2-b2c0-de3e89358cde"). InnerVolumeSpecName "kube-api-access-jsrlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:54:19 crc kubenswrapper[4775]: I1002 01:54:19.827871 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b212cd7-598e-43b2-b2c0-de3e89358cde-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b212cd7-598e-43b2-b2c0-de3e89358cde" (UID: "6b212cd7-598e-43b2-b2c0-de3e89358cde"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:54:19 crc kubenswrapper[4775]: I1002 01:54:19.865285 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b212cd7-598e-43b2-b2c0-de3e89358cde-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:19 crc kubenswrapper[4775]: I1002 01:54:19.865323 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsrlb\" (UniqueName: \"kubernetes.io/projected/6b212cd7-598e-43b2-b2c0-de3e89358cde-kube-api-access-jsrlb\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:19 crc kubenswrapper[4775]: I1002 01:54:19.865336 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b212cd7-598e-43b2-b2c0-de3e89358cde-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.019909 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.020694 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.100570 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.260138 4775 generic.go:334] "Generic (PLEG): container finished" podID="6b212cd7-598e-43b2-b2c0-de3e89358cde" containerID="2f6dec9987abc1444e3ad9103693e8dc153e7d7437fd8f8b26d5d7b4741f1f63" exitCode=0 Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.260241 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rvpz5" Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.260247 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rvpz5" event={"ID":"6b212cd7-598e-43b2-b2c0-de3e89358cde","Type":"ContainerDied","Data":"2f6dec9987abc1444e3ad9103693e8dc153e7d7437fd8f8b26d5d7b4741f1f63"} Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.260639 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rvpz5" event={"ID":"6b212cd7-598e-43b2-b2c0-de3e89358cde","Type":"ContainerDied","Data":"92ad2cc4be7ab97cd5d81a7287ce77d1cfc32a85c98ff75dc8b481339cb0613c"} Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.260667 4775 scope.go:117] "RemoveContainer" containerID="2f6dec9987abc1444e3ad9103693e8dc153e7d7437fd8f8b26d5d7b4741f1f63" Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.288535 4775 scope.go:117] "RemoveContainer" containerID="fa0b4c826756674926bb52677e6b8069c60a1ccdcba60fdc04da84badb9425b0" Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.308157 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rvpz5"] Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.311281 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rvpz5"] Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.334197 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.338641 4775 scope.go:117] "RemoveContainer" containerID="64ab50e35efaef6dd952d0de82b6860917df599072f66bea86f9ffcbd65b8b55" Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.352421 4775 scope.go:117] "RemoveContainer" containerID="2f6dec9987abc1444e3ad9103693e8dc153e7d7437fd8f8b26d5d7b4741f1f63" Oct 02 01:54:20 crc kubenswrapper[4775]: E1002 01:54:20.352782 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f6dec9987abc1444e3ad9103693e8dc153e7d7437fd8f8b26d5d7b4741f1f63\": container with ID starting with 2f6dec9987abc1444e3ad9103693e8dc153e7d7437fd8f8b26d5d7b4741f1f63 not found: ID does not exist" containerID="2f6dec9987abc1444e3ad9103693e8dc153e7d7437fd8f8b26d5d7b4741f1f63" Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.352813 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f6dec9987abc1444e3ad9103693e8dc153e7d7437fd8f8b26d5d7b4741f1f63"} err="failed to get container status \"2f6dec9987abc1444e3ad9103693e8dc153e7d7437fd8f8b26d5d7b4741f1f63\": rpc error: code = NotFound desc = could not find container \"2f6dec9987abc1444e3ad9103693e8dc153e7d7437fd8f8b26d5d7b4741f1f63\": container with ID starting with 2f6dec9987abc1444e3ad9103693e8dc153e7d7437fd8f8b26d5d7b4741f1f63 not found: ID does not exist" Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.352832 4775 scope.go:117] "RemoveContainer" containerID="fa0b4c826756674926bb52677e6b8069c60a1ccdcba60fdc04da84badb9425b0" Oct 02 01:54:20 crc kubenswrapper[4775]: E1002 01:54:20.353211 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa0b4c826756674926bb52677e6b8069c60a1ccdcba60fdc04da84badb9425b0\": container with ID starting with fa0b4c826756674926bb52677e6b8069c60a1ccdcba60fdc04da84badb9425b0 not found: ID does not exist" containerID="fa0b4c826756674926bb52677e6b8069c60a1ccdcba60fdc04da84badb9425b0" Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.353232 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa0b4c826756674926bb52677e6b8069c60a1ccdcba60fdc04da84badb9425b0"} err="failed to get container status \"fa0b4c826756674926bb52677e6b8069c60a1ccdcba60fdc04da84badb9425b0\": rpc error: code = NotFound desc = could not find container \"fa0b4c826756674926bb52677e6b8069c60a1ccdcba60fdc04da84badb9425b0\": container with ID starting with fa0b4c826756674926bb52677e6b8069c60a1ccdcba60fdc04da84badb9425b0 not found: ID does not exist" Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.353243 4775 scope.go:117] "RemoveContainer" containerID="64ab50e35efaef6dd952d0de82b6860917df599072f66bea86f9ffcbd65b8b55" Oct 02 01:54:20 crc kubenswrapper[4775]: E1002 01:54:20.354015 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64ab50e35efaef6dd952d0de82b6860917df599072f66bea86f9ffcbd65b8b55\": container with ID starting with 64ab50e35efaef6dd952d0de82b6860917df599072f66bea86f9ffcbd65b8b55 not found: ID does not exist" containerID="64ab50e35efaef6dd952d0de82b6860917df599072f66bea86f9ffcbd65b8b55" Oct 02 01:54:20 crc kubenswrapper[4775]: I1002 01:54:20.354060 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64ab50e35efaef6dd952d0de82b6860917df599072f66bea86f9ffcbd65b8b55"} err="failed to get container status \"64ab50e35efaef6dd952d0de82b6860917df599072f66bea86f9ffcbd65b8b55\": rpc error: code = NotFound desc = could not find container \"64ab50e35efaef6dd952d0de82b6860917df599072f66bea86f9ffcbd65b8b55\": container with ID starting with 64ab50e35efaef6dd952d0de82b6860917df599072f66bea86f9ffcbd65b8b55 not found: ID does not exist" Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.775082 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b212cd7-598e-43b2-b2c0-de3e89358cde" path="/var/lib/kubelet/pods/6b212cd7-598e-43b2-b2c0-de3e89358cde/volumes" Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.896095 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-584458db67-zj7kv"] Oct 02 01:54:21 crc kubenswrapper[4775]: E1002 01:54:21.896335 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb58aabc-fee6-480d-8d28-d7ebf0506266" containerName="pull" Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.896356 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb58aabc-fee6-480d-8d28-d7ebf0506266" containerName="pull" Oct 02 01:54:21 crc kubenswrapper[4775]: E1002 01:54:21.896375 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b212cd7-598e-43b2-b2c0-de3e89358cde" containerName="extract-utilities" Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.896383 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b212cd7-598e-43b2-b2c0-de3e89358cde" containerName="extract-utilities" Oct 02 01:54:21 crc kubenswrapper[4775]: E1002 01:54:21.896394 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb58aabc-fee6-480d-8d28-d7ebf0506266" containerName="extract" Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.896402 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb58aabc-fee6-480d-8d28-d7ebf0506266" containerName="extract" Oct 02 01:54:21 crc kubenswrapper[4775]: E1002 01:54:21.896415 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b212cd7-598e-43b2-b2c0-de3e89358cde" containerName="extract-content" Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.896424 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b212cd7-598e-43b2-b2c0-de3e89358cde" containerName="extract-content" Oct 02 01:54:21 crc kubenswrapper[4775]: E1002 01:54:21.896439 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b212cd7-598e-43b2-b2c0-de3e89358cde" containerName="registry-server" Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.896446 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b212cd7-598e-43b2-b2c0-de3e89358cde" containerName="registry-server" Oct 02 01:54:21 crc kubenswrapper[4775]: E1002 01:54:21.896460 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb58aabc-fee6-480d-8d28-d7ebf0506266" containerName="util" Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.896467 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb58aabc-fee6-480d-8d28-d7ebf0506266" containerName="util" Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.896596 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b212cd7-598e-43b2-b2c0-de3e89358cde" containerName="registry-server" Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.896609 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb58aabc-fee6-480d-8d28-d7ebf0506266" containerName="extract" Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.897058 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" Oct 02 01:54:21 crc kubenswrapper[4775]: W1002 01:54:21.899882 4775 reflector.go:561] object-"metallb-system"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 02 01:54:21 crc kubenswrapper[4775]: E1002 01:54:21.899921 4775 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 01:54:21 crc kubenswrapper[4775]: W1002 01:54:21.900113 4775 reflector.go:561] object-"metallb-system"/"metallb-operator-controller-manager-service-cert": failed to list *v1.Secret: secrets "metallb-operator-controller-manager-service-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 02 01:54:21 crc kubenswrapper[4775]: E1002 01:54:21.900159 4775 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-operator-controller-manager-service-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-operator-controller-manager-service-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 01:54:21 crc kubenswrapper[4775]: W1002 01:54:21.900260 4775 reflector.go:561] object-"metallb-system"/"metallb-operator-webhook-server-cert": failed to list *v1.Secret: secrets "metallb-operator-webhook-server-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 02 01:54:21 crc kubenswrapper[4775]: E1002 01:54:21.900285 4775 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-operator-webhook-server-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-operator-webhook-server-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 01:54:21 crc kubenswrapper[4775]: W1002 01:54:21.901104 4775 reflector.go:561] object-"metallb-system"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 02 01:54:21 crc kubenswrapper[4775]: E1002 01:54:21.901135 4775 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 01:54:21 crc kubenswrapper[4775]: W1002 01:54:21.903548 4775 reflector.go:561] object-"metallb-system"/"manager-account-dockercfg-8dkxl": failed to list *v1.Secret: secrets "manager-account-dockercfg-8dkxl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 02 01:54:21 crc kubenswrapper[4775]: E1002 01:54:21.903606 4775 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"manager-account-dockercfg-8dkxl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"manager-account-dockercfg-8dkxl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.918664 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-584458db67-zj7kv"] Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.993764 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8e0db7b3-7942-4d66-8ab1-5cfc51e5e355-webhook-cert\") pod \"metallb-operator-controller-manager-584458db67-zj7kv\" (UID: \"8e0db7b3-7942-4d66-8ab1-5cfc51e5e355\") " pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.993807 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8e0db7b3-7942-4d66-8ab1-5cfc51e5e355-apiservice-cert\") pod \"metallb-operator-controller-manager-584458db67-zj7kv\" (UID: \"8e0db7b3-7942-4d66-8ab1-5cfc51e5e355\") " pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" Oct 02 01:54:21 crc kubenswrapper[4775]: I1002 01:54:21.993849 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84jmj\" (UniqueName: \"kubernetes.io/projected/8e0db7b3-7942-4d66-8ab1-5cfc51e5e355-kube-api-access-84jmj\") pod \"metallb-operator-controller-manager-584458db67-zj7kv\" (UID: \"8e0db7b3-7942-4d66-8ab1-5cfc51e5e355\") " pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.095171 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8e0db7b3-7942-4d66-8ab1-5cfc51e5e355-webhook-cert\") pod \"metallb-operator-controller-manager-584458db67-zj7kv\" (UID: \"8e0db7b3-7942-4d66-8ab1-5cfc51e5e355\") " pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.095218 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8e0db7b3-7942-4d66-8ab1-5cfc51e5e355-apiservice-cert\") pod \"metallb-operator-controller-manager-584458db67-zj7kv\" (UID: \"8e0db7b3-7942-4d66-8ab1-5cfc51e5e355\") " pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.095254 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84jmj\" (UniqueName: \"kubernetes.io/projected/8e0db7b3-7942-4d66-8ab1-5cfc51e5e355-kube-api-access-84jmj\") pod \"metallb-operator-controller-manager-584458db67-zj7kv\" (UID: \"8e0db7b3-7942-4d66-8ab1-5cfc51e5e355\") " pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.135841 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx"] Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.136812 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.138512 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.138808 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-4hhw7" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.139464 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.159520 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx"] Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.196296 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/921014ce-ad65-4bb9-b673-0c2883ba98cd-apiservice-cert\") pod \"metallb-operator-webhook-server-5c6cc6d6c-c2kfx\" (UID: \"921014ce-ad65-4bb9-b673-0c2883ba98cd\") " pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.196339 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/921014ce-ad65-4bb9-b673-0c2883ba98cd-webhook-cert\") pod \"metallb-operator-webhook-server-5c6cc6d6c-c2kfx\" (UID: \"921014ce-ad65-4bb9-b673-0c2883ba98cd\") " pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.196434 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz45x\" (UniqueName: \"kubernetes.io/projected/921014ce-ad65-4bb9-b673-0c2883ba98cd-kube-api-access-pz45x\") pod \"metallb-operator-webhook-server-5c6cc6d6c-c2kfx\" (UID: \"921014ce-ad65-4bb9-b673-0c2883ba98cd\") " pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.298016 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/921014ce-ad65-4bb9-b673-0c2883ba98cd-apiservice-cert\") pod \"metallb-operator-webhook-server-5c6cc6d6c-c2kfx\" (UID: \"921014ce-ad65-4bb9-b673-0c2883ba98cd\") " pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.298068 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/921014ce-ad65-4bb9-b673-0c2883ba98cd-webhook-cert\") pod \"metallb-operator-webhook-server-5c6cc6d6c-c2kfx\" (UID: \"921014ce-ad65-4bb9-b673-0c2883ba98cd\") " pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.298151 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz45x\" (UniqueName: \"kubernetes.io/projected/921014ce-ad65-4bb9-b673-0c2883ba98cd-kube-api-access-pz45x\") pod \"metallb-operator-webhook-server-5c6cc6d6c-c2kfx\" (UID: \"921014ce-ad65-4bb9-b673-0c2883ba98cd\") " pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.320241 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/921014ce-ad65-4bb9-b673-0c2883ba98cd-webhook-cert\") pod \"metallb-operator-webhook-server-5c6cc6d6c-c2kfx\" (UID: \"921014ce-ad65-4bb9-b673-0c2883ba98cd\") " pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.320718 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/921014ce-ad65-4bb9-b673-0c2883ba98cd-apiservice-cert\") pod \"metallb-operator-webhook-server-5c6cc6d6c-c2kfx\" (UID: \"921014ce-ad65-4bb9-b673-0c2883ba98cd\") " pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.719238 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-8dkxl" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.851779 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.862698 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8e0db7b3-7942-4d66-8ab1-5cfc51e5e355-apiservice-cert\") pod \"metallb-operator-controller-manager-584458db67-zj7kv\" (UID: \"8e0db7b3-7942-4d66-8ab1-5cfc51e5e355\") " pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" Oct 02 01:54:22 crc kubenswrapper[4775]: I1002 01:54:22.863085 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8e0db7b3-7942-4d66-8ab1-5cfc51e5e355-webhook-cert\") pod \"metallb-operator-controller-manager-584458db67-zj7kv\" (UID: \"8e0db7b3-7942-4d66-8ab1-5cfc51e5e355\") " pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" Oct 02 01:54:23 crc kubenswrapper[4775]: E1002 01:54:23.114261 4775 projected.go:288] Couldn't get configMap metallb-system/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 02 01:54:23 crc kubenswrapper[4775]: I1002 01:54:23.188582 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 02 01:54:23 crc kubenswrapper[4775]: I1002 01:54:23.374642 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 02 01:54:23 crc kubenswrapper[4775]: I1002 01:54:23.382612 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz45x\" (UniqueName: \"kubernetes.io/projected/921014ce-ad65-4bb9-b673-0c2883ba98cd-kube-api-access-pz45x\") pod \"metallb-operator-webhook-server-5c6cc6d6c-c2kfx\" (UID: \"921014ce-ad65-4bb9-b673-0c2883ba98cd\") " pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" Oct 02 01:54:23 crc kubenswrapper[4775]: E1002 01:54:23.384792 4775 projected.go:194] Error preparing data for projected volume kube-api-access-84jmj for pod metallb-system/metallb-operator-controller-manager-584458db67-zj7kv: failed to sync configmap cache: timed out waiting for the condition Oct 02 01:54:23 crc kubenswrapper[4775]: E1002 01:54:23.385170 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/8e0db7b3-7942-4d66-8ab1-5cfc51e5e355-kube-api-access-84jmj podName:8e0db7b3-7942-4d66-8ab1-5cfc51e5e355 nodeName:}" failed. No retries permitted until 2025-10-02 01:54:23.885132546 +0000 UTC m=+801.051876626 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-84jmj" (UniqueName: "kubernetes.io/projected/8e0db7b3-7942-4d66-8ab1-5cfc51e5e355-kube-api-access-84jmj") pod "metallb-operator-controller-manager-584458db67-zj7kv" (UID: "8e0db7b3-7942-4d66-8ab1-5cfc51e5e355") : failed to sync configmap cache: timed out waiting for the condition Oct 02 01:54:23 crc kubenswrapper[4775]: I1002 01:54:23.437109 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 02 01:54:23 crc kubenswrapper[4775]: I1002 01:54:23.653648 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" Oct 02 01:54:23 crc kubenswrapper[4775]: I1002 01:54:23.899567 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b6749"] Oct 02 01:54:23 crc kubenswrapper[4775]: I1002 01:54:23.900064 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-b6749" podUID="d520a768-2d91-45f0-a1d5-825b2c03853f" containerName="registry-server" containerID="cri-o://9b667b9e0f6367a0f4bb14f671271a59a2d865e6ede2eb8e304205cfee3b1f0c" gracePeriod=2 Oct 02 01:54:23 crc kubenswrapper[4775]: I1002 01:54:23.928151 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84jmj\" (UniqueName: \"kubernetes.io/projected/8e0db7b3-7942-4d66-8ab1-5cfc51e5e355-kube-api-access-84jmj\") pod \"metallb-operator-controller-manager-584458db67-zj7kv\" (UID: \"8e0db7b3-7942-4d66-8ab1-5cfc51e5e355\") " pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" Oct 02 01:54:23 crc kubenswrapper[4775]: I1002 01:54:23.935024 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84jmj\" (UniqueName: \"kubernetes.io/projected/8e0db7b3-7942-4d66-8ab1-5cfc51e5e355-kube-api-access-84jmj\") pod \"metallb-operator-controller-manager-584458db67-zj7kv\" (UID: \"8e0db7b3-7942-4d66-8ab1-5cfc51e5e355\") " pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" Oct 02 01:54:24 crc kubenswrapper[4775]: I1002 01:54:24.011600 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" Oct 02 01:54:24 crc kubenswrapper[4775]: I1002 01:54:24.179164 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx"] Oct 02 01:54:24 crc kubenswrapper[4775]: I1002 01:54:24.281764 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" event={"ID":"921014ce-ad65-4bb9-b673-0c2883ba98cd","Type":"ContainerStarted","Data":"dc62042944d05988b26d6130a71e0accbe97c9602279e438d8b531afc6c6ebd7"} Oct 02 01:54:24 crc kubenswrapper[4775]: I1002 01:54:24.334036 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-584458db67-zj7kv"] Oct 02 01:54:24 crc kubenswrapper[4775]: W1002 01:54:24.368438 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e0db7b3_7942_4d66_8ab1_5cfc51e5e355.slice/crio-0eb77fac88578f47df4e52cd43f113ff311696185f71fdf344e1f0d7dcdec34c WatchSource:0}: Error finding container 0eb77fac88578f47df4e52cd43f113ff311696185f71fdf344e1f0d7dcdec34c: Status 404 returned error can't find the container with id 0eb77fac88578f47df4e52cd43f113ff311696185f71fdf344e1f0d7dcdec34c Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.016265 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.144061 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d520a768-2d91-45f0-a1d5-825b2c03853f-utilities\") pod \"d520a768-2d91-45f0-a1d5-825b2c03853f\" (UID: \"d520a768-2d91-45f0-a1d5-825b2c03853f\") " Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.144175 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d520a768-2d91-45f0-a1d5-825b2c03853f-catalog-content\") pod \"d520a768-2d91-45f0-a1d5-825b2c03853f\" (UID: \"d520a768-2d91-45f0-a1d5-825b2c03853f\") " Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.144198 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9cch\" (UniqueName: \"kubernetes.io/projected/d520a768-2d91-45f0-a1d5-825b2c03853f-kube-api-access-s9cch\") pod \"d520a768-2d91-45f0-a1d5-825b2c03853f\" (UID: \"d520a768-2d91-45f0-a1d5-825b2c03853f\") " Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.145259 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d520a768-2d91-45f0-a1d5-825b2c03853f-utilities" (OuterVolumeSpecName: "utilities") pod "d520a768-2d91-45f0-a1d5-825b2c03853f" (UID: "d520a768-2d91-45f0-a1d5-825b2c03853f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.152652 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d520a768-2d91-45f0-a1d5-825b2c03853f-kube-api-access-s9cch" (OuterVolumeSpecName: "kube-api-access-s9cch") pod "d520a768-2d91-45f0-a1d5-825b2c03853f" (UID: "d520a768-2d91-45f0-a1d5-825b2c03853f"). InnerVolumeSpecName "kube-api-access-s9cch". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.219490 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d520a768-2d91-45f0-a1d5-825b2c03853f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d520a768-2d91-45f0-a1d5-825b2c03853f" (UID: "d520a768-2d91-45f0-a1d5-825b2c03853f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.245916 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d520a768-2d91-45f0-a1d5-825b2c03853f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.245986 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9cch\" (UniqueName: \"kubernetes.io/projected/d520a768-2d91-45f0-a1d5-825b2c03853f-kube-api-access-s9cch\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.246005 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d520a768-2d91-45f0-a1d5-825b2c03853f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.288465 4775 generic.go:334] "Generic (PLEG): container finished" podID="d520a768-2d91-45f0-a1d5-825b2c03853f" containerID="9b667b9e0f6367a0f4bb14f671271a59a2d865e6ede2eb8e304205cfee3b1f0c" exitCode=0 Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.288531 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-b6749" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.288554 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6749" event={"ID":"d520a768-2d91-45f0-a1d5-825b2c03853f","Type":"ContainerDied","Data":"9b667b9e0f6367a0f4bb14f671271a59a2d865e6ede2eb8e304205cfee3b1f0c"} Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.288630 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-b6749" event={"ID":"d520a768-2d91-45f0-a1d5-825b2c03853f","Type":"ContainerDied","Data":"eece375a8320b0b9bef235489168df836433fe74f008b208a9f55dd383238b80"} Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.288655 4775 scope.go:117] "RemoveContainer" containerID="9b667b9e0f6367a0f4bb14f671271a59a2d865e6ede2eb8e304205cfee3b1f0c" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.289735 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" event={"ID":"8e0db7b3-7942-4d66-8ab1-5cfc51e5e355","Type":"ContainerStarted","Data":"0eb77fac88578f47df4e52cd43f113ff311696185f71fdf344e1f0d7dcdec34c"} Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.316042 4775 scope.go:117] "RemoveContainer" containerID="f68c1eb2f8157eab2be8ad54a5121b589f40e622aeca3fa4e588dd8ff476ccef" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.324626 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-b6749"] Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.333326 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-b6749"] Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.348936 4775 scope.go:117] "RemoveContainer" containerID="4205b4f4a2be9423a16ea8258f9e206d92f602e28407b6a9d88f2d773d13e1c5" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.392472 4775 scope.go:117] "RemoveContainer" containerID="9b667b9e0f6367a0f4bb14f671271a59a2d865e6ede2eb8e304205cfee3b1f0c" Oct 02 01:54:25 crc kubenswrapper[4775]: E1002 01:54:25.392928 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b667b9e0f6367a0f4bb14f671271a59a2d865e6ede2eb8e304205cfee3b1f0c\": container with ID starting with 9b667b9e0f6367a0f4bb14f671271a59a2d865e6ede2eb8e304205cfee3b1f0c not found: ID does not exist" containerID="9b667b9e0f6367a0f4bb14f671271a59a2d865e6ede2eb8e304205cfee3b1f0c" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.392988 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b667b9e0f6367a0f4bb14f671271a59a2d865e6ede2eb8e304205cfee3b1f0c"} err="failed to get container status \"9b667b9e0f6367a0f4bb14f671271a59a2d865e6ede2eb8e304205cfee3b1f0c\": rpc error: code = NotFound desc = could not find container \"9b667b9e0f6367a0f4bb14f671271a59a2d865e6ede2eb8e304205cfee3b1f0c\": container with ID starting with 9b667b9e0f6367a0f4bb14f671271a59a2d865e6ede2eb8e304205cfee3b1f0c not found: ID does not exist" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.393014 4775 scope.go:117] "RemoveContainer" containerID="f68c1eb2f8157eab2be8ad54a5121b589f40e622aeca3fa4e588dd8ff476ccef" Oct 02 01:54:25 crc kubenswrapper[4775]: E1002 01:54:25.398383 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f68c1eb2f8157eab2be8ad54a5121b589f40e622aeca3fa4e588dd8ff476ccef\": container with ID starting with f68c1eb2f8157eab2be8ad54a5121b589f40e622aeca3fa4e588dd8ff476ccef not found: ID does not exist" containerID="f68c1eb2f8157eab2be8ad54a5121b589f40e622aeca3fa4e588dd8ff476ccef" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.398434 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68c1eb2f8157eab2be8ad54a5121b589f40e622aeca3fa4e588dd8ff476ccef"} err="failed to get container status \"f68c1eb2f8157eab2be8ad54a5121b589f40e622aeca3fa4e588dd8ff476ccef\": rpc error: code = NotFound desc = could not find container \"f68c1eb2f8157eab2be8ad54a5121b589f40e622aeca3fa4e588dd8ff476ccef\": container with ID starting with f68c1eb2f8157eab2be8ad54a5121b589f40e622aeca3fa4e588dd8ff476ccef not found: ID does not exist" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.398462 4775 scope.go:117] "RemoveContainer" containerID="4205b4f4a2be9423a16ea8258f9e206d92f602e28407b6a9d88f2d773d13e1c5" Oct 02 01:54:25 crc kubenswrapper[4775]: E1002 01:54:25.401290 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4205b4f4a2be9423a16ea8258f9e206d92f602e28407b6a9d88f2d773d13e1c5\": container with ID starting with 4205b4f4a2be9423a16ea8258f9e206d92f602e28407b6a9d88f2d773d13e1c5 not found: ID does not exist" containerID="4205b4f4a2be9423a16ea8258f9e206d92f602e28407b6a9d88f2d773d13e1c5" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.401318 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4205b4f4a2be9423a16ea8258f9e206d92f602e28407b6a9d88f2d773d13e1c5"} err="failed to get container status \"4205b4f4a2be9423a16ea8258f9e206d92f602e28407b6a9d88f2d773d13e1c5\": rpc error: code = NotFound desc = could not find container \"4205b4f4a2be9423a16ea8258f9e206d92f602e28407b6a9d88f2d773d13e1c5\": container with ID starting with 4205b4f4a2be9423a16ea8258f9e206d92f602e28407b6a9d88f2d773d13e1c5 not found: ID does not exist" Oct 02 01:54:25 crc kubenswrapper[4775]: I1002 01:54:25.785065 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d520a768-2d91-45f0-a1d5-825b2c03853f" path="/var/lib/kubelet/pods/d520a768-2d91-45f0-a1d5-825b2c03853f/volumes" Oct 02 01:54:29 crc kubenswrapper[4775]: I1002 01:54:29.322045 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" event={"ID":"8e0db7b3-7942-4d66-8ab1-5cfc51e5e355","Type":"ContainerStarted","Data":"24012248be00625736931a0e627a51ddd001f30dfb3ed8749b2ea77d74a2141c"} Oct 02 01:54:29 crc kubenswrapper[4775]: I1002 01:54:29.322620 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" Oct 02 01:54:29 crc kubenswrapper[4775]: I1002 01:54:29.349128 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" podStartSLOduration=3.661204519 podStartE2EDuration="8.349107801s" podCreationTimestamp="2025-10-02 01:54:21 +0000 UTC" firstStartedPulling="2025-10-02 01:54:24.370889327 +0000 UTC m=+801.537633357" lastFinishedPulling="2025-10-02 01:54:29.058792599 +0000 UTC m=+806.225536639" observedRunningTime="2025-10-02 01:54:29.34536385 +0000 UTC m=+806.512107890" watchObservedRunningTime="2025-10-02 01:54:29.349107801 +0000 UTC m=+806.515851841" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.330754 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" event={"ID":"921014ce-ad65-4bb9-b673-0c2883ba98cd","Type":"ContainerStarted","Data":"92fc3f4010d99e9478ae954528cf04eb811701f9b4929d358ad59ac2366b2ad7"} Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.330917 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.362324 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" podStartSLOduration=3.490698795 podStartE2EDuration="8.362304742s" podCreationTimestamp="2025-10-02 01:54:22 +0000 UTC" firstStartedPulling="2025-10-02 01:54:24.198919238 +0000 UTC m=+801.365663278" lastFinishedPulling="2025-10-02 01:54:29.070525185 +0000 UTC m=+806.237269225" observedRunningTime="2025-10-02 01:54:30.357525275 +0000 UTC m=+807.524269325" watchObservedRunningTime="2025-10-02 01:54:30.362304742 +0000 UTC m=+807.529048782" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.503289 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qfrqq"] Oct 02 01:54:30 crc kubenswrapper[4775]: E1002 01:54:30.503542 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d520a768-2d91-45f0-a1d5-825b2c03853f" containerName="extract-content" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.503557 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d520a768-2d91-45f0-a1d5-825b2c03853f" containerName="extract-content" Oct 02 01:54:30 crc kubenswrapper[4775]: E1002 01:54:30.503570 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d520a768-2d91-45f0-a1d5-825b2c03853f" containerName="extract-utilities" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.503579 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d520a768-2d91-45f0-a1d5-825b2c03853f" containerName="extract-utilities" Oct 02 01:54:30 crc kubenswrapper[4775]: E1002 01:54:30.503599 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d520a768-2d91-45f0-a1d5-825b2c03853f" containerName="registry-server" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.503608 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d520a768-2d91-45f0-a1d5-825b2c03853f" containerName="registry-server" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.503730 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d520a768-2d91-45f0-a1d5-825b2c03853f" containerName="registry-server" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.504735 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.516056 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qfrqq"] Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.618621 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e67413cc-5710-4056-8670-57b6fd8f68a7-catalog-content\") pod \"community-operators-qfrqq\" (UID: \"e67413cc-5710-4056-8670-57b6fd8f68a7\") " pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.618709 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqsvq\" (UniqueName: \"kubernetes.io/projected/e67413cc-5710-4056-8670-57b6fd8f68a7-kube-api-access-pqsvq\") pod \"community-operators-qfrqq\" (UID: \"e67413cc-5710-4056-8670-57b6fd8f68a7\") " pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.618871 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e67413cc-5710-4056-8670-57b6fd8f68a7-utilities\") pod \"community-operators-qfrqq\" (UID: \"e67413cc-5710-4056-8670-57b6fd8f68a7\") " pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.720440 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e67413cc-5710-4056-8670-57b6fd8f68a7-catalog-content\") pod \"community-operators-qfrqq\" (UID: \"e67413cc-5710-4056-8670-57b6fd8f68a7\") " pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.720500 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqsvq\" (UniqueName: \"kubernetes.io/projected/e67413cc-5710-4056-8670-57b6fd8f68a7-kube-api-access-pqsvq\") pod \"community-operators-qfrqq\" (UID: \"e67413cc-5710-4056-8670-57b6fd8f68a7\") " pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.720549 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e67413cc-5710-4056-8670-57b6fd8f68a7-utilities\") pod \"community-operators-qfrqq\" (UID: \"e67413cc-5710-4056-8670-57b6fd8f68a7\") " pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.720972 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e67413cc-5710-4056-8670-57b6fd8f68a7-utilities\") pod \"community-operators-qfrqq\" (UID: \"e67413cc-5710-4056-8670-57b6fd8f68a7\") " pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.729935 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e67413cc-5710-4056-8670-57b6fd8f68a7-catalog-content\") pod \"community-operators-qfrqq\" (UID: \"e67413cc-5710-4056-8670-57b6fd8f68a7\") " pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.753823 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqsvq\" (UniqueName: \"kubernetes.io/projected/e67413cc-5710-4056-8670-57b6fd8f68a7-kube-api-access-pqsvq\") pod \"community-operators-qfrqq\" (UID: \"e67413cc-5710-4056-8670-57b6fd8f68a7\") " pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:30 crc kubenswrapper[4775]: I1002 01:54:30.838206 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:31 crc kubenswrapper[4775]: I1002 01:54:31.331756 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qfrqq"] Oct 02 01:54:31 crc kubenswrapper[4775]: W1002 01:54:31.342213 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode67413cc_5710_4056_8670_57b6fd8f68a7.slice/crio-c00d1dbfeaa60514e8c738e378bada20d76d76240a8f6557db03e0dd992f2cc2 WatchSource:0}: Error finding container c00d1dbfeaa60514e8c738e378bada20d76d76240a8f6557db03e0dd992f2cc2: Status 404 returned error can't find the container with id c00d1dbfeaa60514e8c738e378bada20d76d76240a8f6557db03e0dd992f2cc2 Oct 02 01:54:32 crc kubenswrapper[4775]: I1002 01:54:32.343724 4775 generic.go:334] "Generic (PLEG): container finished" podID="e67413cc-5710-4056-8670-57b6fd8f68a7" containerID="44ef7eef8cbe06857319e3987a69a3fca27a8abf67191aa001a4a9378d4ef960" exitCode=0 Oct 02 01:54:32 crc kubenswrapper[4775]: I1002 01:54:32.343781 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qfrqq" event={"ID":"e67413cc-5710-4056-8670-57b6fd8f68a7","Type":"ContainerDied","Data":"44ef7eef8cbe06857319e3987a69a3fca27a8abf67191aa001a4a9378d4ef960"} Oct 02 01:54:32 crc kubenswrapper[4775]: I1002 01:54:32.343818 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qfrqq" event={"ID":"e67413cc-5710-4056-8670-57b6fd8f68a7","Type":"ContainerStarted","Data":"c00d1dbfeaa60514e8c738e378bada20d76d76240a8f6557db03e0dd992f2cc2"} Oct 02 01:54:33 crc kubenswrapper[4775]: I1002 01:54:33.351485 4775 generic.go:334] "Generic (PLEG): container finished" podID="e67413cc-5710-4056-8670-57b6fd8f68a7" containerID="865d9b5df7e5f6ae606986ded48786cffe7c79e7f1d5b0850f1d6427630b762b" exitCode=0 Oct 02 01:54:33 crc kubenswrapper[4775]: I1002 01:54:33.351549 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qfrqq" event={"ID":"e67413cc-5710-4056-8670-57b6fd8f68a7","Type":"ContainerDied","Data":"865d9b5df7e5f6ae606986ded48786cffe7c79e7f1d5b0850f1d6427630b762b"} Oct 02 01:54:34 crc kubenswrapper[4775]: I1002 01:54:34.359627 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qfrqq" event={"ID":"e67413cc-5710-4056-8670-57b6fd8f68a7","Type":"ContainerStarted","Data":"b318f4d98971080ace96b503aa2d619dee304ac4f94d3a22b272b4a41c5c4b6a"} Oct 02 01:54:34 crc kubenswrapper[4775]: I1002 01:54:34.388801 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qfrqq" podStartSLOduration=3.004563764 podStartE2EDuration="4.388780342s" podCreationTimestamp="2025-10-02 01:54:30 +0000 UTC" firstStartedPulling="2025-10-02 01:54:32.345886729 +0000 UTC m=+809.512630769" lastFinishedPulling="2025-10-02 01:54:33.730103317 +0000 UTC m=+810.896847347" observedRunningTime="2025-10-02 01:54:34.386097657 +0000 UTC m=+811.552841727" watchObservedRunningTime="2025-10-02 01:54:34.388780342 +0000 UTC m=+811.555524392" Oct 02 01:54:40 crc kubenswrapper[4775]: I1002 01:54:40.838671 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:40 crc kubenswrapper[4775]: I1002 01:54:40.839438 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:40 crc kubenswrapper[4775]: I1002 01:54:40.885824 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:41 crc kubenswrapper[4775]: I1002 01:54:41.466136 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:43 crc kubenswrapper[4775]: I1002 01:54:43.664517 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5c6cc6d6c-c2kfx" Oct 02 01:54:43 crc kubenswrapper[4775]: I1002 01:54:43.700753 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qfrqq"] Oct 02 01:54:43 crc kubenswrapper[4775]: I1002 01:54:43.701248 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qfrqq" podUID="e67413cc-5710-4056-8670-57b6fd8f68a7" containerName="registry-server" containerID="cri-o://b318f4d98971080ace96b503aa2d619dee304ac4f94d3a22b272b4a41c5c4b6a" gracePeriod=2 Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.154608 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.223515 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e67413cc-5710-4056-8670-57b6fd8f68a7-utilities\") pod \"e67413cc-5710-4056-8670-57b6fd8f68a7\" (UID: \"e67413cc-5710-4056-8670-57b6fd8f68a7\") " Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.223591 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e67413cc-5710-4056-8670-57b6fd8f68a7-catalog-content\") pod \"e67413cc-5710-4056-8670-57b6fd8f68a7\" (UID: \"e67413cc-5710-4056-8670-57b6fd8f68a7\") " Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.223636 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqsvq\" (UniqueName: \"kubernetes.io/projected/e67413cc-5710-4056-8670-57b6fd8f68a7-kube-api-access-pqsvq\") pod \"e67413cc-5710-4056-8670-57b6fd8f68a7\" (UID: \"e67413cc-5710-4056-8670-57b6fd8f68a7\") " Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.225197 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e67413cc-5710-4056-8670-57b6fd8f68a7-utilities" (OuterVolumeSpecName: "utilities") pod "e67413cc-5710-4056-8670-57b6fd8f68a7" (UID: "e67413cc-5710-4056-8670-57b6fd8f68a7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.231203 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e67413cc-5710-4056-8670-57b6fd8f68a7-kube-api-access-pqsvq" (OuterVolumeSpecName: "kube-api-access-pqsvq") pod "e67413cc-5710-4056-8670-57b6fd8f68a7" (UID: "e67413cc-5710-4056-8670-57b6fd8f68a7"). InnerVolumeSpecName "kube-api-access-pqsvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.325710 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqsvq\" (UniqueName: \"kubernetes.io/projected/e67413cc-5710-4056-8670-57b6fd8f68a7-kube-api-access-pqsvq\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.325758 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e67413cc-5710-4056-8670-57b6fd8f68a7-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.418911 4775 generic.go:334] "Generic (PLEG): container finished" podID="e67413cc-5710-4056-8670-57b6fd8f68a7" containerID="b318f4d98971080ace96b503aa2d619dee304ac4f94d3a22b272b4a41c5c4b6a" exitCode=0 Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.418948 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qfrqq" event={"ID":"e67413cc-5710-4056-8670-57b6fd8f68a7","Type":"ContainerDied","Data":"b318f4d98971080ace96b503aa2d619dee304ac4f94d3a22b272b4a41c5c4b6a"} Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.418979 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qfrqq" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.418998 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qfrqq" event={"ID":"e67413cc-5710-4056-8670-57b6fd8f68a7","Type":"ContainerDied","Data":"c00d1dbfeaa60514e8c738e378bada20d76d76240a8f6557db03e0dd992f2cc2"} Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.419015 4775 scope.go:117] "RemoveContainer" containerID="b318f4d98971080ace96b503aa2d619dee304ac4f94d3a22b272b4a41c5c4b6a" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.451180 4775 scope.go:117] "RemoveContainer" containerID="865d9b5df7e5f6ae606986ded48786cffe7c79e7f1d5b0850f1d6427630b762b" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.471109 4775 scope.go:117] "RemoveContainer" containerID="44ef7eef8cbe06857319e3987a69a3fca27a8abf67191aa001a4a9378d4ef960" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.486916 4775 scope.go:117] "RemoveContainer" containerID="b318f4d98971080ace96b503aa2d619dee304ac4f94d3a22b272b4a41c5c4b6a" Oct 02 01:54:44 crc kubenswrapper[4775]: E1002 01:54:44.487409 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b318f4d98971080ace96b503aa2d619dee304ac4f94d3a22b272b4a41c5c4b6a\": container with ID starting with b318f4d98971080ace96b503aa2d619dee304ac4f94d3a22b272b4a41c5c4b6a not found: ID does not exist" containerID="b318f4d98971080ace96b503aa2d619dee304ac4f94d3a22b272b4a41c5c4b6a" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.487439 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b318f4d98971080ace96b503aa2d619dee304ac4f94d3a22b272b4a41c5c4b6a"} err="failed to get container status \"b318f4d98971080ace96b503aa2d619dee304ac4f94d3a22b272b4a41c5c4b6a\": rpc error: code = NotFound desc = could not find container \"b318f4d98971080ace96b503aa2d619dee304ac4f94d3a22b272b4a41c5c4b6a\": container with ID starting with b318f4d98971080ace96b503aa2d619dee304ac4f94d3a22b272b4a41c5c4b6a not found: ID does not exist" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.487460 4775 scope.go:117] "RemoveContainer" containerID="865d9b5df7e5f6ae606986ded48786cffe7c79e7f1d5b0850f1d6427630b762b" Oct 02 01:54:44 crc kubenswrapper[4775]: E1002 01:54:44.487733 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"865d9b5df7e5f6ae606986ded48786cffe7c79e7f1d5b0850f1d6427630b762b\": container with ID starting with 865d9b5df7e5f6ae606986ded48786cffe7c79e7f1d5b0850f1d6427630b762b not found: ID does not exist" containerID="865d9b5df7e5f6ae606986ded48786cffe7c79e7f1d5b0850f1d6427630b762b" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.487761 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"865d9b5df7e5f6ae606986ded48786cffe7c79e7f1d5b0850f1d6427630b762b"} err="failed to get container status \"865d9b5df7e5f6ae606986ded48786cffe7c79e7f1d5b0850f1d6427630b762b\": rpc error: code = NotFound desc = could not find container \"865d9b5df7e5f6ae606986ded48786cffe7c79e7f1d5b0850f1d6427630b762b\": container with ID starting with 865d9b5df7e5f6ae606986ded48786cffe7c79e7f1d5b0850f1d6427630b762b not found: ID does not exist" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.487774 4775 scope.go:117] "RemoveContainer" containerID="44ef7eef8cbe06857319e3987a69a3fca27a8abf67191aa001a4a9378d4ef960" Oct 02 01:54:44 crc kubenswrapper[4775]: E1002 01:54:44.488195 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44ef7eef8cbe06857319e3987a69a3fca27a8abf67191aa001a4a9378d4ef960\": container with ID starting with 44ef7eef8cbe06857319e3987a69a3fca27a8abf67191aa001a4a9378d4ef960 not found: ID does not exist" containerID="44ef7eef8cbe06857319e3987a69a3fca27a8abf67191aa001a4a9378d4ef960" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.488216 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44ef7eef8cbe06857319e3987a69a3fca27a8abf67191aa001a4a9378d4ef960"} err="failed to get container status \"44ef7eef8cbe06857319e3987a69a3fca27a8abf67191aa001a4a9378d4ef960\": rpc error: code = NotFound desc = could not find container \"44ef7eef8cbe06857319e3987a69a3fca27a8abf67191aa001a4a9378d4ef960\": container with ID starting with 44ef7eef8cbe06857319e3987a69a3fca27a8abf67191aa001a4a9378d4ef960 not found: ID does not exist" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.850273 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e67413cc-5710-4056-8670-57b6fd8f68a7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e67413cc-5710-4056-8670-57b6fd8f68a7" (UID: "e67413cc-5710-4056-8670-57b6fd8f68a7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:54:44 crc kubenswrapper[4775]: I1002 01:54:44.933801 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e67413cc-5710-4056-8670-57b6fd8f68a7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:54:45 crc kubenswrapper[4775]: I1002 01:54:45.061812 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qfrqq"] Oct 02 01:54:45 crc kubenswrapper[4775]: I1002 01:54:45.071924 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qfrqq"] Oct 02 01:54:45 crc kubenswrapper[4775]: I1002 01:54:45.775158 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e67413cc-5710-4056-8670-57b6fd8f68a7" path="/var/lib/kubelet/pods/e67413cc-5710-4056-8670-57b6fd8f68a7/volumes" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.707928 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-slrcg"] Oct 02 01:54:56 crc kubenswrapper[4775]: E1002 01:54:56.708843 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e67413cc-5710-4056-8670-57b6fd8f68a7" containerName="extract-content" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.708864 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e67413cc-5710-4056-8670-57b6fd8f68a7" containerName="extract-content" Oct 02 01:54:56 crc kubenswrapper[4775]: E1002 01:54:56.708884 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e67413cc-5710-4056-8670-57b6fd8f68a7" containerName="extract-utilities" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.708897 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e67413cc-5710-4056-8670-57b6fd8f68a7" containerName="extract-utilities" Oct 02 01:54:56 crc kubenswrapper[4775]: E1002 01:54:56.708921 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e67413cc-5710-4056-8670-57b6fd8f68a7" containerName="registry-server" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.708933 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e67413cc-5710-4056-8670-57b6fd8f68a7" containerName="registry-server" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.709155 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e67413cc-5710-4056-8670-57b6fd8f68a7" containerName="registry-server" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.710567 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.728619 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-slrcg"] Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.827793 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-catalog-content\") pod \"redhat-marketplace-slrcg\" (UID: \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\") " pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.827886 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-utilities\") pod \"redhat-marketplace-slrcg\" (UID: \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\") " pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.828061 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kknk8\" (UniqueName: \"kubernetes.io/projected/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-kube-api-access-kknk8\") pod \"redhat-marketplace-slrcg\" (UID: \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\") " pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.929157 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-utilities\") pod \"redhat-marketplace-slrcg\" (UID: \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\") " pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.929288 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kknk8\" (UniqueName: \"kubernetes.io/projected/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-kube-api-access-kknk8\") pod \"redhat-marketplace-slrcg\" (UID: \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\") " pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.929361 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-catalog-content\") pod \"redhat-marketplace-slrcg\" (UID: \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\") " pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.929652 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-utilities\") pod \"redhat-marketplace-slrcg\" (UID: \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\") " pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.929910 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-catalog-content\") pod \"redhat-marketplace-slrcg\" (UID: \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\") " pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:54:56 crc kubenswrapper[4775]: I1002 01:54:56.967216 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kknk8\" (UniqueName: \"kubernetes.io/projected/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-kube-api-access-kknk8\") pod \"redhat-marketplace-slrcg\" (UID: \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\") " pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:54:57 crc kubenswrapper[4775]: I1002 01:54:57.035058 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:54:57 crc kubenswrapper[4775]: I1002 01:54:57.323069 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-slrcg"] Oct 02 01:54:57 crc kubenswrapper[4775]: I1002 01:54:57.506564 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-slrcg" event={"ID":"2608f8d9-5fa1-49f6-b384-6c3f20bcb727","Type":"ContainerStarted","Data":"8519d7ccaac2628501a80ea4727426d4264b39160b9c56374785f4dfccb21747"} Oct 02 01:54:58 crc kubenswrapper[4775]: I1002 01:54:58.518898 4775 generic.go:334] "Generic (PLEG): container finished" podID="2608f8d9-5fa1-49f6-b384-6c3f20bcb727" containerID="f9aa6414263f90595c92d675db22505a0e771ac174745127803258569c315767" exitCode=0 Oct 02 01:54:58 crc kubenswrapper[4775]: I1002 01:54:58.518979 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-slrcg" event={"ID":"2608f8d9-5fa1-49f6-b384-6c3f20bcb727","Type":"ContainerDied","Data":"f9aa6414263f90595c92d675db22505a0e771ac174745127803258569c315767"} Oct 02 01:54:59 crc kubenswrapper[4775]: I1002 01:54:59.529821 4775 generic.go:334] "Generic (PLEG): container finished" podID="2608f8d9-5fa1-49f6-b384-6c3f20bcb727" containerID="f8030d0cd7e83168308fcdd69a530fd38446d722d9c044d0d88808927b3e1d48" exitCode=0 Oct 02 01:54:59 crc kubenswrapper[4775]: I1002 01:54:59.529932 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-slrcg" event={"ID":"2608f8d9-5fa1-49f6-b384-6c3f20bcb727","Type":"ContainerDied","Data":"f8030d0cd7e83168308fcdd69a530fd38446d722d9c044d0d88808927b3e1d48"} Oct 02 01:55:00 crc kubenswrapper[4775]: I1002 01:55:00.538883 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-slrcg" event={"ID":"2608f8d9-5fa1-49f6-b384-6c3f20bcb727","Type":"ContainerStarted","Data":"ba696ed0433d6e148cf1cc98130c586f60dca71450a8247c7c3d8a19f65644c1"} Oct 02 01:55:00 crc kubenswrapper[4775]: I1002 01:55:00.565076 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-slrcg" podStartSLOduration=3.131893715 podStartE2EDuration="4.565051544s" podCreationTimestamp="2025-10-02 01:54:56 +0000 UTC" firstStartedPulling="2025-10-02 01:54:58.52125763 +0000 UTC m=+835.688001710" lastFinishedPulling="2025-10-02 01:54:59.954415489 +0000 UTC m=+837.121159539" observedRunningTime="2025-10-02 01:55:00.563377553 +0000 UTC m=+837.730121623" watchObservedRunningTime="2025-10-02 01:55:00.565051544 +0000 UTC m=+837.731795614" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.016673 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-584458db67-zj7kv" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.835665 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r"] Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.836742 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.843377 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-28dbf" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.848580 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-g9pzp"] Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.851283 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.853870 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.853902 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.854034 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.858640 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r"] Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.917249 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-tq2ld"] Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.918585 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-tq2ld" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.920107 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-tntlr" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.920433 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.920566 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.925539 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.929317 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-42v88"] Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.930309 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-42v88" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.931850 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.942835 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15825320-b57d-4195-b410-65ca428465fa-metrics-certs\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.942879 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/15825320-b57d-4195-b410-65ca428465fa-frr-startup\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.942898 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/15825320-b57d-4195-b410-65ca428465fa-frr-conf\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.943013 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/15825320-b57d-4195-b410-65ca428465fa-metrics\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.943090 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6gcf\" (UniqueName: \"kubernetes.io/projected/15825320-b57d-4195-b410-65ca428465fa-kube-api-access-c6gcf\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.943129 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/15825320-b57d-4195-b410-65ca428465fa-frr-sockets\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.943155 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfwgg\" (UniqueName: \"kubernetes.io/projected/3a402158-a679-4d43-8e16-7b3ccfe91452-kube-api-access-lfwgg\") pod \"frr-k8s-webhook-server-64bf5d555-7d29r\" (UID: \"3a402158-a679-4d43-8e16-7b3ccfe91452\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.943206 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a402158-a679-4d43-8e16-7b3ccfe91452-cert\") pod \"frr-k8s-webhook-server-64bf5d555-7d29r\" (UID: \"3a402158-a679-4d43-8e16-7b3ccfe91452\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.943279 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/15825320-b57d-4195-b410-65ca428465fa-reloader\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:04 crc kubenswrapper[4775]: I1002 01:55:04.944030 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-42v88"] Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.044904 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/15825320-b57d-4195-b410-65ca428465fa-frr-sockets\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.045451 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6gcf\" (UniqueName: \"kubernetes.io/projected/15825320-b57d-4195-b410-65ca428465fa-kube-api-access-c6gcf\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.045531 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfwgg\" (UniqueName: \"kubernetes.io/projected/3a402158-a679-4d43-8e16-7b3ccfe91452-kube-api-access-lfwgg\") pod \"frr-k8s-webhook-server-64bf5d555-7d29r\" (UID: \"3a402158-a679-4d43-8e16-7b3ccfe91452\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.045636 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8d6246a6-7cbf-4d65-9d84-ee9649d3fed6-cert\") pod \"controller-68d546b9d8-42v88\" (UID: \"8d6246a6-7cbf-4d65-9d84-ee9649d3fed6\") " pod="metallb-system/controller-68d546b9d8-42v88" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.045760 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a402158-a679-4d43-8e16-7b3ccfe91452-cert\") pod \"frr-k8s-webhook-server-64bf5d555-7d29r\" (UID: \"3a402158-a679-4d43-8e16-7b3ccfe91452\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.045850 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpx66\" (UniqueName: \"kubernetes.io/projected/8d6246a6-7cbf-4d65-9d84-ee9649d3fed6-kube-api-access-rpx66\") pod \"controller-68d546b9d8-42v88\" (UID: \"8d6246a6-7cbf-4d65-9d84-ee9649d3fed6\") " pod="metallb-system/controller-68d546b9d8-42v88" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.045937 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a77c5c49-2e87-45ae-b2ab-5994102a0f38-metrics-certs\") pod \"speaker-tq2ld\" (UID: \"a77c5c49-2e87-45ae-b2ab-5994102a0f38\") " pod="metallb-system/speaker-tq2ld" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.046041 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d6246a6-7cbf-4d65-9d84-ee9649d3fed6-metrics-certs\") pod \"controller-68d546b9d8-42v88\" (UID: \"8d6246a6-7cbf-4d65-9d84-ee9649d3fed6\") " pod="metallb-system/controller-68d546b9d8-42v88" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.046149 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a77c5c49-2e87-45ae-b2ab-5994102a0f38-memberlist\") pod \"speaker-tq2ld\" (UID: \"a77c5c49-2e87-45ae-b2ab-5994102a0f38\") " pod="metallb-system/speaker-tq2ld" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.046039 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/15825320-b57d-4195-b410-65ca428465fa-frr-sockets\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.046281 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/15825320-b57d-4195-b410-65ca428465fa-reloader\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.046478 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15825320-b57d-4195-b410-65ca428465fa-metrics-certs\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.046533 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/15825320-b57d-4195-b410-65ca428465fa-frr-startup\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.046555 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/15825320-b57d-4195-b410-65ca428465fa-frr-conf\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.046585 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fvmg8\" (UniqueName: \"kubernetes.io/projected/a77c5c49-2e87-45ae-b2ab-5994102a0f38-kube-api-access-fvmg8\") pod \"speaker-tq2ld\" (UID: \"a77c5c49-2e87-45ae-b2ab-5994102a0f38\") " pod="metallb-system/speaker-tq2ld" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.046635 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/15825320-b57d-4195-b410-65ca428465fa-metrics\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.046677 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a77c5c49-2e87-45ae-b2ab-5994102a0f38-metallb-excludel2\") pod \"speaker-tq2ld\" (UID: \"a77c5c49-2e87-45ae-b2ab-5994102a0f38\") " pod="metallb-system/speaker-tq2ld" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.046928 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/15825320-b57d-4195-b410-65ca428465fa-reloader\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.047413 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/15825320-b57d-4195-b410-65ca428465fa-metrics\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.047540 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/15825320-b57d-4195-b410-65ca428465fa-frr-conf\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.047744 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/15825320-b57d-4195-b410-65ca428465fa-frr-startup\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.052694 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3a402158-a679-4d43-8e16-7b3ccfe91452-cert\") pod \"frr-k8s-webhook-server-64bf5d555-7d29r\" (UID: \"3a402158-a679-4d43-8e16-7b3ccfe91452\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.061548 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/15825320-b57d-4195-b410-65ca428465fa-metrics-certs\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.064419 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6gcf\" (UniqueName: \"kubernetes.io/projected/15825320-b57d-4195-b410-65ca428465fa-kube-api-access-c6gcf\") pod \"frr-k8s-g9pzp\" (UID: \"15825320-b57d-4195-b410-65ca428465fa\") " pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.067078 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfwgg\" (UniqueName: \"kubernetes.io/projected/3a402158-a679-4d43-8e16-7b3ccfe91452-kube-api-access-lfwgg\") pod \"frr-k8s-webhook-server-64bf5d555-7d29r\" (UID: \"3a402158-a679-4d43-8e16-7b3ccfe91452\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.147745 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fvmg8\" (UniqueName: \"kubernetes.io/projected/a77c5c49-2e87-45ae-b2ab-5994102a0f38-kube-api-access-fvmg8\") pod \"speaker-tq2ld\" (UID: \"a77c5c49-2e87-45ae-b2ab-5994102a0f38\") " pod="metallb-system/speaker-tq2ld" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.148141 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a77c5c49-2e87-45ae-b2ab-5994102a0f38-metallb-excludel2\") pod \"speaker-tq2ld\" (UID: \"a77c5c49-2e87-45ae-b2ab-5994102a0f38\") " pod="metallb-system/speaker-tq2ld" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.148207 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8d6246a6-7cbf-4d65-9d84-ee9649d3fed6-cert\") pod \"controller-68d546b9d8-42v88\" (UID: \"8d6246a6-7cbf-4d65-9d84-ee9649d3fed6\") " pod="metallb-system/controller-68d546b9d8-42v88" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.148268 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpx66\" (UniqueName: \"kubernetes.io/projected/8d6246a6-7cbf-4d65-9d84-ee9649d3fed6-kube-api-access-rpx66\") pod \"controller-68d546b9d8-42v88\" (UID: \"8d6246a6-7cbf-4d65-9d84-ee9649d3fed6\") " pod="metallb-system/controller-68d546b9d8-42v88" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.148302 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a77c5c49-2e87-45ae-b2ab-5994102a0f38-metrics-certs\") pod \"speaker-tq2ld\" (UID: \"a77c5c49-2e87-45ae-b2ab-5994102a0f38\") " pod="metallb-system/speaker-tq2ld" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.148342 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d6246a6-7cbf-4d65-9d84-ee9649d3fed6-metrics-certs\") pod \"controller-68d546b9d8-42v88\" (UID: \"8d6246a6-7cbf-4d65-9d84-ee9649d3fed6\") " pod="metallb-system/controller-68d546b9d8-42v88" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.148414 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a77c5c49-2e87-45ae-b2ab-5994102a0f38-memberlist\") pod \"speaker-tq2ld\" (UID: \"a77c5c49-2e87-45ae-b2ab-5994102a0f38\") " pod="metallb-system/speaker-tq2ld" Oct 02 01:55:05 crc kubenswrapper[4775]: E1002 01:55:05.148657 4775 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 01:55:05 crc kubenswrapper[4775]: E1002 01:55:05.148754 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a77c5c49-2e87-45ae-b2ab-5994102a0f38-memberlist podName:a77c5c49-2e87-45ae-b2ab-5994102a0f38 nodeName:}" failed. No retries permitted until 2025-10-02 01:55:05.648725917 +0000 UTC m=+842.815469997 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a77c5c49-2e87-45ae-b2ab-5994102a0f38-memberlist") pod "speaker-tq2ld" (UID: "a77c5c49-2e87-45ae-b2ab-5994102a0f38") : secret "metallb-memberlist" not found Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.148819 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a77c5c49-2e87-45ae-b2ab-5994102a0f38-metallb-excludel2\") pod \"speaker-tq2ld\" (UID: \"a77c5c49-2e87-45ae-b2ab-5994102a0f38\") " pod="metallb-system/speaker-tq2ld" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.150602 4775 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.152556 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8d6246a6-7cbf-4d65-9d84-ee9649d3fed6-metrics-certs\") pod \"controller-68d546b9d8-42v88\" (UID: \"8d6246a6-7cbf-4d65-9d84-ee9649d3fed6\") " pod="metallb-system/controller-68d546b9d8-42v88" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.152915 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a77c5c49-2e87-45ae-b2ab-5994102a0f38-metrics-certs\") pod \"speaker-tq2ld\" (UID: \"a77c5c49-2e87-45ae-b2ab-5994102a0f38\") " pod="metallb-system/speaker-tq2ld" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.155402 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.172593 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.173860 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fvmg8\" (UniqueName: \"kubernetes.io/projected/a77c5c49-2e87-45ae-b2ab-5994102a0f38-kube-api-access-fvmg8\") pod \"speaker-tq2ld\" (UID: \"a77c5c49-2e87-45ae-b2ab-5994102a0f38\") " pod="metallb-system/speaker-tq2ld" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.176182 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpx66\" (UniqueName: \"kubernetes.io/projected/8d6246a6-7cbf-4d65-9d84-ee9649d3fed6-kube-api-access-rpx66\") pod \"controller-68d546b9d8-42v88\" (UID: \"8d6246a6-7cbf-4d65-9d84-ee9649d3fed6\") " pod="metallb-system/controller-68d546b9d8-42v88" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.182419 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8d6246a6-7cbf-4d65-9d84-ee9649d3fed6-cert\") pod \"controller-68d546b9d8-42v88\" (UID: \"8d6246a6-7cbf-4d65-9d84-ee9649d3fed6\") " pod="metallb-system/controller-68d546b9d8-42v88" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.245268 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-42v88" Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.567158 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r"] Oct 02 01:55:05 crc kubenswrapper[4775]: W1002 01:55:05.574779 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a402158_a679_4d43_8e16_7b3ccfe91452.slice/crio-6f59ce0965bffd387acff9142c922495da9b4da5e4eaa4d0bf158273d7a0467b WatchSource:0}: Error finding container 6f59ce0965bffd387acff9142c922495da9b4da5e4eaa4d0bf158273d7a0467b: Status 404 returned error can't find the container with id 6f59ce0965bffd387acff9142c922495da9b4da5e4eaa4d0bf158273d7a0467b Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.576252 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g9pzp" event={"ID":"15825320-b57d-4195-b410-65ca428465fa","Type":"ContainerStarted","Data":"ea2c10bcf7fe76b0a21bcbe4097fc255a16e2b3a1c53f1bee423e5a3b9445905"} Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.655429 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a77c5c49-2e87-45ae-b2ab-5994102a0f38-memberlist\") pod \"speaker-tq2ld\" (UID: \"a77c5c49-2e87-45ae-b2ab-5994102a0f38\") " pod="metallb-system/speaker-tq2ld" Oct 02 01:55:05 crc kubenswrapper[4775]: E1002 01:55:05.655585 4775 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 02 01:55:05 crc kubenswrapper[4775]: E1002 01:55:05.655679 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a77c5c49-2e87-45ae-b2ab-5994102a0f38-memberlist podName:a77c5c49-2e87-45ae-b2ab-5994102a0f38 nodeName:}" failed. No retries permitted until 2025-10-02 01:55:06.655655412 +0000 UTC m=+843.822399462 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a77c5c49-2e87-45ae-b2ab-5994102a0f38-memberlist") pod "speaker-tq2ld" (UID: "a77c5c49-2e87-45ae-b2ab-5994102a0f38") : secret "metallb-memberlist" not found Oct 02 01:55:05 crc kubenswrapper[4775]: I1002 01:55:05.690636 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-42v88"] Oct 02 01:55:05 crc kubenswrapper[4775]: W1002 01:55:05.696199 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d6246a6_7cbf_4d65_9d84_ee9649d3fed6.slice/crio-aa03473691460d8df117b05adff593c838dc6460efc2a1fd71ca34cd18b8a91a WatchSource:0}: Error finding container aa03473691460d8df117b05adff593c838dc6460efc2a1fd71ca34cd18b8a91a: Status 404 returned error can't find the container with id aa03473691460d8df117b05adff593c838dc6460efc2a1fd71ca34cd18b8a91a Oct 02 01:55:06 crc kubenswrapper[4775]: I1002 01:55:06.585382 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r" event={"ID":"3a402158-a679-4d43-8e16-7b3ccfe91452","Type":"ContainerStarted","Data":"6f59ce0965bffd387acff9142c922495da9b4da5e4eaa4d0bf158273d7a0467b"} Oct 02 01:55:06 crc kubenswrapper[4775]: I1002 01:55:06.589558 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-42v88" event={"ID":"8d6246a6-7cbf-4d65-9d84-ee9649d3fed6","Type":"ContainerStarted","Data":"afb25251863bda525499f3f2d4a20f005c6ef8e3237e9d283ea30b338dc62660"} Oct 02 01:55:06 crc kubenswrapper[4775]: I1002 01:55:06.589622 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-42v88" event={"ID":"8d6246a6-7cbf-4d65-9d84-ee9649d3fed6","Type":"ContainerStarted","Data":"036ebb019b6ea570faf250fda00bfb1051f58e27a0e3d269103167d519a28eb0"} Oct 02 01:55:06 crc kubenswrapper[4775]: I1002 01:55:06.589641 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-42v88" event={"ID":"8d6246a6-7cbf-4d65-9d84-ee9649d3fed6","Type":"ContainerStarted","Data":"aa03473691460d8df117b05adff593c838dc6460efc2a1fd71ca34cd18b8a91a"} Oct 02 01:55:06 crc kubenswrapper[4775]: I1002 01:55:06.590453 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-42v88" Oct 02 01:55:06 crc kubenswrapper[4775]: I1002 01:55:06.610155 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-42v88" podStartSLOduration=2.6101364030000003 podStartE2EDuration="2.610136403s" podCreationTimestamp="2025-10-02 01:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:55:06.608431305 +0000 UTC m=+843.775175395" watchObservedRunningTime="2025-10-02 01:55:06.610136403 +0000 UTC m=+843.776880443" Oct 02 01:55:06 crc kubenswrapper[4775]: I1002 01:55:06.668066 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a77c5c49-2e87-45ae-b2ab-5994102a0f38-memberlist\") pod \"speaker-tq2ld\" (UID: \"a77c5c49-2e87-45ae-b2ab-5994102a0f38\") " pod="metallb-system/speaker-tq2ld" Oct 02 01:55:06 crc kubenswrapper[4775]: I1002 01:55:06.675482 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a77c5c49-2e87-45ae-b2ab-5994102a0f38-memberlist\") pod \"speaker-tq2ld\" (UID: \"a77c5c49-2e87-45ae-b2ab-5994102a0f38\") " pod="metallb-system/speaker-tq2ld" Oct 02 01:55:06 crc kubenswrapper[4775]: I1002 01:55:06.732921 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-tq2ld" Oct 02 01:55:06 crc kubenswrapper[4775]: W1002 01:55:06.784860 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda77c5c49_2e87_45ae_b2ab_5994102a0f38.slice/crio-16256685a256cfc83a6529659d250e95318612485d2c5ce7fdbba4f1bbc6c0f3 WatchSource:0}: Error finding container 16256685a256cfc83a6529659d250e95318612485d2c5ce7fdbba4f1bbc6c0f3: Status 404 returned error can't find the container with id 16256685a256cfc83a6529659d250e95318612485d2c5ce7fdbba4f1bbc6c0f3 Oct 02 01:55:07 crc kubenswrapper[4775]: I1002 01:55:07.039182 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:55:07 crc kubenswrapper[4775]: I1002 01:55:07.039238 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:55:07 crc kubenswrapper[4775]: I1002 01:55:07.079626 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:55:07 crc kubenswrapper[4775]: I1002 01:55:07.598289 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tq2ld" event={"ID":"a77c5c49-2e87-45ae-b2ab-5994102a0f38","Type":"ContainerStarted","Data":"f7989145d0d59a46aef83e6ecb4bdbfac4e69e31852d5c46636b7c379d2fb751"} Oct 02 01:55:07 crc kubenswrapper[4775]: I1002 01:55:07.598345 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tq2ld" event={"ID":"a77c5c49-2e87-45ae-b2ab-5994102a0f38","Type":"ContainerStarted","Data":"4e98fb6a95109e296d308706dc92a01113f52e662368c407260fb2339f71d07d"} Oct 02 01:55:07 crc kubenswrapper[4775]: I1002 01:55:07.598357 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tq2ld" event={"ID":"a77c5c49-2e87-45ae-b2ab-5994102a0f38","Type":"ContainerStarted","Data":"16256685a256cfc83a6529659d250e95318612485d2c5ce7fdbba4f1bbc6c0f3"} Oct 02 01:55:07 crc kubenswrapper[4775]: I1002 01:55:07.598592 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-tq2ld" Oct 02 01:55:07 crc kubenswrapper[4775]: I1002 01:55:07.659186 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:55:07 crc kubenswrapper[4775]: I1002 01:55:07.678029 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-tq2ld" podStartSLOduration=3.67801463 podStartE2EDuration="3.67801463s" podCreationTimestamp="2025-10-02 01:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:55:07.619481641 +0000 UTC m=+844.786225681" watchObservedRunningTime="2025-10-02 01:55:07.67801463 +0000 UTC m=+844.844758660" Oct 02 01:55:09 crc kubenswrapper[4775]: I1002 01:55:09.485997 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-slrcg"] Oct 02 01:55:09 crc kubenswrapper[4775]: I1002 01:55:09.610381 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-slrcg" podUID="2608f8d9-5fa1-49f6-b384-6c3f20bcb727" containerName="registry-server" containerID="cri-o://ba696ed0433d6e148cf1cc98130c586f60dca71450a8247c7c3d8a19f65644c1" gracePeriod=2 Oct 02 01:55:10 crc kubenswrapper[4775]: I1002 01:55:10.618438 4775 generic.go:334] "Generic (PLEG): container finished" podID="2608f8d9-5fa1-49f6-b384-6c3f20bcb727" containerID="ba696ed0433d6e148cf1cc98130c586f60dca71450a8247c7c3d8a19f65644c1" exitCode=0 Oct 02 01:55:10 crc kubenswrapper[4775]: I1002 01:55:10.618504 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-slrcg" event={"ID":"2608f8d9-5fa1-49f6-b384-6c3f20bcb727","Type":"ContainerDied","Data":"ba696ed0433d6e148cf1cc98130c586f60dca71450a8247c7c3d8a19f65644c1"} Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.483063 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.567376 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kknk8\" (UniqueName: \"kubernetes.io/projected/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-kube-api-access-kknk8\") pod \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\" (UID: \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\") " Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.567451 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-catalog-content\") pod \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\" (UID: \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\") " Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.567515 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-utilities\") pod \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\" (UID: \"2608f8d9-5fa1-49f6-b384-6c3f20bcb727\") " Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.568717 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-utilities" (OuterVolumeSpecName: "utilities") pod "2608f8d9-5fa1-49f6-b384-6c3f20bcb727" (UID: "2608f8d9-5fa1-49f6-b384-6c3f20bcb727"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.572427 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-kube-api-access-kknk8" (OuterVolumeSpecName: "kube-api-access-kknk8") pod "2608f8d9-5fa1-49f6-b384-6c3f20bcb727" (UID: "2608f8d9-5fa1-49f6-b384-6c3f20bcb727"). InnerVolumeSpecName "kube-api-access-kknk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.595754 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2608f8d9-5fa1-49f6-b384-6c3f20bcb727" (UID: "2608f8d9-5fa1-49f6-b384-6c3f20bcb727"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.631718 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-slrcg" event={"ID":"2608f8d9-5fa1-49f6-b384-6c3f20bcb727","Type":"ContainerDied","Data":"8519d7ccaac2628501a80ea4727426d4264b39160b9c56374785f4dfccb21747"} Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.631870 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-slrcg" Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.631986 4775 scope.go:117] "RemoveContainer" containerID="ba696ed0433d6e148cf1cc98130c586f60dca71450a8247c7c3d8a19f65644c1" Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.668734 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kknk8\" (UniqueName: \"kubernetes.io/projected/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-kube-api-access-kknk8\") on node \"crc\" DevicePath \"\"" Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.668763 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.668776 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2608f8d9-5fa1-49f6-b384-6c3f20bcb727-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.690162 4775 scope.go:117] "RemoveContainer" containerID="f8030d0cd7e83168308fcdd69a530fd38446d722d9c044d0d88808927b3e1d48" Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.721307 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-slrcg"] Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.726071 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-slrcg"] Oct 02 01:55:12 crc kubenswrapper[4775]: I1002 01:55:12.727234 4775 scope.go:117] "RemoveContainer" containerID="f9aa6414263f90595c92d675db22505a0e771ac174745127803258569c315767" Oct 02 01:55:13 crc kubenswrapper[4775]: I1002 01:55:13.641845 4775 generic.go:334] "Generic (PLEG): container finished" podID="15825320-b57d-4195-b410-65ca428465fa" containerID="8536ebef5b02c7e4182b5c40b92593f77ee696167a586f2d277da5a1c12cdb12" exitCode=0 Oct 02 01:55:13 crc kubenswrapper[4775]: I1002 01:55:13.642010 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g9pzp" event={"ID":"15825320-b57d-4195-b410-65ca428465fa","Type":"ContainerDied","Data":"8536ebef5b02c7e4182b5c40b92593f77ee696167a586f2d277da5a1c12cdb12"} Oct 02 01:55:13 crc kubenswrapper[4775]: I1002 01:55:13.646277 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r" event={"ID":"3a402158-a679-4d43-8e16-7b3ccfe91452","Type":"ContainerStarted","Data":"2cf63c5f8ec0b29e8f825d82baa0a12a3ece04292efb21d991cc0847d1e76f9a"} Oct 02 01:55:13 crc kubenswrapper[4775]: I1002 01:55:13.646531 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r" Oct 02 01:55:13 crc kubenswrapper[4775]: I1002 01:55:13.709827 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r" podStartSLOduration=2.700420152 podStartE2EDuration="9.709793488s" podCreationTimestamp="2025-10-02 01:55:04 +0000 UTC" firstStartedPulling="2025-10-02 01:55:05.577644306 +0000 UTC m=+842.744388386" lastFinishedPulling="2025-10-02 01:55:12.587017682 +0000 UTC m=+849.753761722" observedRunningTime="2025-10-02 01:55:13.7045328 +0000 UTC m=+850.871276900" watchObservedRunningTime="2025-10-02 01:55:13.709793488 +0000 UTC m=+850.876537598" Oct 02 01:55:13 crc kubenswrapper[4775]: I1002 01:55:13.776922 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2608f8d9-5fa1-49f6-b384-6c3f20bcb727" path="/var/lib/kubelet/pods/2608f8d9-5fa1-49f6-b384-6c3f20bcb727/volumes" Oct 02 01:55:14 crc kubenswrapper[4775]: I1002 01:55:14.657739 4775 generic.go:334] "Generic (PLEG): container finished" podID="15825320-b57d-4195-b410-65ca428465fa" containerID="9ba3ef840f705d201b528101fdd2e961398b7f1a14fa733c0eb57211cbcc5971" exitCode=0 Oct 02 01:55:14 crc kubenswrapper[4775]: I1002 01:55:14.657864 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g9pzp" event={"ID":"15825320-b57d-4195-b410-65ca428465fa","Type":"ContainerDied","Data":"9ba3ef840f705d201b528101fdd2e961398b7f1a14fa733c0eb57211cbcc5971"} Oct 02 01:55:15 crc kubenswrapper[4775]: I1002 01:55:15.254004 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-42v88" Oct 02 01:55:15 crc kubenswrapper[4775]: I1002 01:55:15.665818 4775 generic.go:334] "Generic (PLEG): container finished" podID="15825320-b57d-4195-b410-65ca428465fa" containerID="03e2f114f543095f75004210030186f0c01bb53f3cbe305b51614b7bdeb00024" exitCode=0 Oct 02 01:55:15 crc kubenswrapper[4775]: I1002 01:55:15.665861 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g9pzp" event={"ID":"15825320-b57d-4195-b410-65ca428465fa","Type":"ContainerDied","Data":"03e2f114f543095f75004210030186f0c01bb53f3cbe305b51614b7bdeb00024"} Oct 02 01:55:16 crc kubenswrapper[4775]: I1002 01:55:16.681775 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g9pzp" event={"ID":"15825320-b57d-4195-b410-65ca428465fa","Type":"ContainerStarted","Data":"26cf4003130c92fcc64f8109dc9a08705d8b208cc130b13cdce8246a1901e262"} Oct 02 01:55:16 crc kubenswrapper[4775]: I1002 01:55:16.682195 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g9pzp" event={"ID":"15825320-b57d-4195-b410-65ca428465fa","Type":"ContainerStarted","Data":"ee859e92612864eca5391f773f2d4fc74c353c71b853a19227be019eaaed061f"} Oct 02 01:55:16 crc kubenswrapper[4775]: I1002 01:55:16.682212 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g9pzp" event={"ID":"15825320-b57d-4195-b410-65ca428465fa","Type":"ContainerStarted","Data":"c4e3bdd8e98233f2500449c3cca81298fff23e35169ca7b5112b30bed95dc04b"} Oct 02 01:55:16 crc kubenswrapper[4775]: I1002 01:55:16.682238 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g9pzp" event={"ID":"15825320-b57d-4195-b410-65ca428465fa","Type":"ContainerStarted","Data":"3964eb9c900ad4744f58ffb99f9115a2789f26a4006206050584862e9a4b534b"} Oct 02 01:55:16 crc kubenswrapper[4775]: I1002 01:55:16.682250 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g9pzp" event={"ID":"15825320-b57d-4195-b410-65ca428465fa","Type":"ContainerStarted","Data":"e32fdd031116d05cd226493175f748a34caf79f7a93a5d95326648825a2a1765"} Oct 02 01:55:17 crc kubenswrapper[4775]: I1002 01:55:17.698888 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-g9pzp" event={"ID":"15825320-b57d-4195-b410-65ca428465fa","Type":"ContainerStarted","Data":"aae93e9396a2e1b5b587becdcb12ade73215dd00cfd87c6e3143514f74f1828d"} Oct 02 01:55:17 crc kubenswrapper[4775]: I1002 01:55:17.699264 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:17 crc kubenswrapper[4775]: I1002 01:55:17.734528 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-g9pzp" podStartSLOduration=6.522477866 podStartE2EDuration="13.734506628s" podCreationTimestamp="2025-10-02 01:55:04 +0000 UTC" firstStartedPulling="2025-10-02 01:55:05.355414622 +0000 UTC m=+842.522158662" lastFinishedPulling="2025-10-02 01:55:12.567443374 +0000 UTC m=+849.734187424" observedRunningTime="2025-10-02 01:55:17.731296206 +0000 UTC m=+854.898040256" watchObservedRunningTime="2025-10-02 01:55:17.734506628 +0000 UTC m=+854.901250698" Oct 02 01:55:20 crc kubenswrapper[4775]: I1002 01:55:20.173525 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:20 crc kubenswrapper[4775]: I1002 01:55:20.240648 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:25 crc kubenswrapper[4775]: I1002 01:55:25.163584 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-7d29r" Oct 02 01:55:25 crc kubenswrapper[4775]: I1002 01:55:25.179780 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-g9pzp" Oct 02 01:55:26 crc kubenswrapper[4775]: I1002 01:55:26.744765 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-tq2ld" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.357305 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h"] Oct 02 01:55:28 crc kubenswrapper[4775]: E1002 01:55:28.357603 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2608f8d9-5fa1-49f6-b384-6c3f20bcb727" containerName="extract-content" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.357618 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2608f8d9-5fa1-49f6-b384-6c3f20bcb727" containerName="extract-content" Oct 02 01:55:28 crc kubenswrapper[4775]: E1002 01:55:28.357631 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2608f8d9-5fa1-49f6-b384-6c3f20bcb727" containerName="extract-utilities" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.357640 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2608f8d9-5fa1-49f6-b384-6c3f20bcb727" containerName="extract-utilities" Oct 02 01:55:28 crc kubenswrapper[4775]: E1002 01:55:28.357651 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2608f8d9-5fa1-49f6-b384-6c3f20bcb727" containerName="registry-server" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.357659 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2608f8d9-5fa1-49f6-b384-6c3f20bcb727" containerName="registry-server" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.357779 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2608f8d9-5fa1-49f6-b384-6c3f20bcb727" containerName="registry-server" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.360534 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.362660 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.370572 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h"] Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.427519 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9802ba9a-2a80-44d8-b475-27bfa5044497-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h\" (UID: \"9802ba9a-2a80-44d8-b475-27bfa5044497\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.427845 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfrnx\" (UniqueName: \"kubernetes.io/projected/9802ba9a-2a80-44d8-b475-27bfa5044497-kube-api-access-wfrnx\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h\" (UID: \"9802ba9a-2a80-44d8-b475-27bfa5044497\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.427867 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9802ba9a-2a80-44d8-b475-27bfa5044497-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h\" (UID: \"9802ba9a-2a80-44d8-b475-27bfa5044497\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.529422 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfrnx\" (UniqueName: \"kubernetes.io/projected/9802ba9a-2a80-44d8-b475-27bfa5044497-kube-api-access-wfrnx\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h\" (UID: \"9802ba9a-2a80-44d8-b475-27bfa5044497\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.529469 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9802ba9a-2a80-44d8-b475-27bfa5044497-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h\" (UID: \"9802ba9a-2a80-44d8-b475-27bfa5044497\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.529513 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9802ba9a-2a80-44d8-b475-27bfa5044497-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h\" (UID: \"9802ba9a-2a80-44d8-b475-27bfa5044497\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.530023 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9802ba9a-2a80-44d8-b475-27bfa5044497-util\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h\" (UID: \"9802ba9a-2a80-44d8-b475-27bfa5044497\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.530042 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9802ba9a-2a80-44d8-b475-27bfa5044497-bundle\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h\" (UID: \"9802ba9a-2a80-44d8-b475-27bfa5044497\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.574976 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfrnx\" (UniqueName: \"kubernetes.io/projected/9802ba9a-2a80-44d8-b475-27bfa5044497-kube-api-access-wfrnx\") pod \"695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h\" (UID: \"9802ba9a-2a80-44d8-b475-27bfa5044497\") " pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" Oct 02 01:55:28 crc kubenswrapper[4775]: I1002 01:55:28.698120 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" Oct 02 01:55:29 crc kubenswrapper[4775]: I1002 01:55:29.134768 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h"] Oct 02 01:55:29 crc kubenswrapper[4775]: I1002 01:55:29.798189 4775 generic.go:334] "Generic (PLEG): container finished" podID="9802ba9a-2a80-44d8-b475-27bfa5044497" containerID="7e2b91c63dd6864b0d22de24d0715199d0d9052e6b85b27374c81c74689e41d7" exitCode=0 Oct 02 01:55:29 crc kubenswrapper[4775]: I1002 01:55:29.798325 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" event={"ID":"9802ba9a-2a80-44d8-b475-27bfa5044497","Type":"ContainerDied","Data":"7e2b91c63dd6864b0d22de24d0715199d0d9052e6b85b27374c81c74689e41d7"} Oct 02 01:55:29 crc kubenswrapper[4775]: I1002 01:55:29.798441 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" event={"ID":"9802ba9a-2a80-44d8-b475-27bfa5044497","Type":"ContainerStarted","Data":"988a75bc45711588091e648e6d6b353eb7a9cbb7aa207a3558d3febf039be0de"} Oct 02 01:55:33 crc kubenswrapper[4775]: I1002 01:55:33.825075 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" event={"ID":"9802ba9a-2a80-44d8-b475-27bfa5044497","Type":"ContainerStarted","Data":"b97549711295f6b33eb224b22fc7db107e9b1fd5c48ecd0f15088607c5717bbd"} Oct 02 01:55:34 crc kubenswrapper[4775]: I1002 01:55:34.833874 4775 generic.go:334] "Generic (PLEG): container finished" podID="9802ba9a-2a80-44d8-b475-27bfa5044497" containerID="b97549711295f6b33eb224b22fc7db107e9b1fd5c48ecd0f15088607c5717bbd" exitCode=0 Oct 02 01:55:34 crc kubenswrapper[4775]: I1002 01:55:34.833919 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" event={"ID":"9802ba9a-2a80-44d8-b475-27bfa5044497","Type":"ContainerDied","Data":"b97549711295f6b33eb224b22fc7db107e9b1fd5c48ecd0f15088607c5717bbd"} Oct 02 01:55:35 crc kubenswrapper[4775]: I1002 01:55:35.850526 4775 generic.go:334] "Generic (PLEG): container finished" podID="9802ba9a-2a80-44d8-b475-27bfa5044497" containerID="11f234ebcf79cf473f78ea0ae6fbdd411baf867e2fcf73dcd235b8408f8cff9f" exitCode=0 Oct 02 01:55:35 crc kubenswrapper[4775]: I1002 01:55:35.850691 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" event={"ID":"9802ba9a-2a80-44d8-b475-27bfa5044497","Type":"ContainerDied","Data":"11f234ebcf79cf473f78ea0ae6fbdd411baf867e2fcf73dcd235b8408f8cff9f"} Oct 02 01:55:37 crc kubenswrapper[4775]: I1002 01:55:37.245928 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" Oct 02 01:55:37 crc kubenswrapper[4775]: I1002 01:55:37.373716 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9802ba9a-2a80-44d8-b475-27bfa5044497-util\") pod \"9802ba9a-2a80-44d8-b475-27bfa5044497\" (UID: \"9802ba9a-2a80-44d8-b475-27bfa5044497\") " Oct 02 01:55:37 crc kubenswrapper[4775]: I1002 01:55:37.373860 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfrnx\" (UniqueName: \"kubernetes.io/projected/9802ba9a-2a80-44d8-b475-27bfa5044497-kube-api-access-wfrnx\") pod \"9802ba9a-2a80-44d8-b475-27bfa5044497\" (UID: \"9802ba9a-2a80-44d8-b475-27bfa5044497\") " Oct 02 01:55:37 crc kubenswrapper[4775]: I1002 01:55:37.373914 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9802ba9a-2a80-44d8-b475-27bfa5044497-bundle\") pod \"9802ba9a-2a80-44d8-b475-27bfa5044497\" (UID: \"9802ba9a-2a80-44d8-b475-27bfa5044497\") " Oct 02 01:55:37 crc kubenswrapper[4775]: I1002 01:55:37.375503 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9802ba9a-2a80-44d8-b475-27bfa5044497-bundle" (OuterVolumeSpecName: "bundle") pod "9802ba9a-2a80-44d8-b475-27bfa5044497" (UID: "9802ba9a-2a80-44d8-b475-27bfa5044497"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:55:37 crc kubenswrapper[4775]: I1002 01:55:37.382059 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9802ba9a-2a80-44d8-b475-27bfa5044497-kube-api-access-wfrnx" (OuterVolumeSpecName: "kube-api-access-wfrnx") pod "9802ba9a-2a80-44d8-b475-27bfa5044497" (UID: "9802ba9a-2a80-44d8-b475-27bfa5044497"). InnerVolumeSpecName "kube-api-access-wfrnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:55:37 crc kubenswrapper[4775]: I1002 01:55:37.383446 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9802ba9a-2a80-44d8-b475-27bfa5044497-util" (OuterVolumeSpecName: "util") pod "9802ba9a-2a80-44d8-b475-27bfa5044497" (UID: "9802ba9a-2a80-44d8-b475-27bfa5044497"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:55:37 crc kubenswrapper[4775]: I1002 01:55:37.475774 4775 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9802ba9a-2a80-44d8-b475-27bfa5044497-util\") on node \"crc\" DevicePath \"\"" Oct 02 01:55:37 crc kubenswrapper[4775]: I1002 01:55:37.475810 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfrnx\" (UniqueName: \"kubernetes.io/projected/9802ba9a-2a80-44d8-b475-27bfa5044497-kube-api-access-wfrnx\") on node \"crc\" DevicePath \"\"" Oct 02 01:55:37 crc kubenswrapper[4775]: I1002 01:55:37.475820 4775 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9802ba9a-2a80-44d8-b475-27bfa5044497-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:55:37 crc kubenswrapper[4775]: I1002 01:55:37.869500 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" event={"ID":"9802ba9a-2a80-44d8-b475-27bfa5044497","Type":"ContainerDied","Data":"988a75bc45711588091e648e6d6b353eb7a9cbb7aa207a3558d3febf039be0de"} Oct 02 01:55:37 crc kubenswrapper[4775]: I1002 01:55:37.869551 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="988a75bc45711588091e648e6d6b353eb7a9cbb7aa207a3558d3febf039be0de" Oct 02 01:55:37 crc kubenswrapper[4775]: I1002 01:55:37.869612 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h" Oct 02 01:55:42 crc kubenswrapper[4775]: I1002 01:55:42.493382 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4wqh4"] Oct 02 01:55:42 crc kubenswrapper[4775]: E1002 01:55:42.494183 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9802ba9a-2a80-44d8-b475-27bfa5044497" containerName="util" Oct 02 01:55:42 crc kubenswrapper[4775]: I1002 01:55:42.494198 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9802ba9a-2a80-44d8-b475-27bfa5044497" containerName="util" Oct 02 01:55:42 crc kubenswrapper[4775]: E1002 01:55:42.494215 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9802ba9a-2a80-44d8-b475-27bfa5044497" containerName="extract" Oct 02 01:55:42 crc kubenswrapper[4775]: I1002 01:55:42.494222 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9802ba9a-2a80-44d8-b475-27bfa5044497" containerName="extract" Oct 02 01:55:42 crc kubenswrapper[4775]: E1002 01:55:42.494232 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9802ba9a-2a80-44d8-b475-27bfa5044497" containerName="pull" Oct 02 01:55:42 crc kubenswrapper[4775]: I1002 01:55:42.494240 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9802ba9a-2a80-44d8-b475-27bfa5044497" containerName="pull" Oct 02 01:55:42 crc kubenswrapper[4775]: I1002 01:55:42.494377 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="9802ba9a-2a80-44d8-b475-27bfa5044497" containerName="extract" Oct 02 01:55:42 crc kubenswrapper[4775]: I1002 01:55:42.494863 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4wqh4" Oct 02 01:55:42 crc kubenswrapper[4775]: I1002 01:55:42.497036 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Oct 02 01:55:42 crc kubenswrapper[4775]: I1002 01:55:42.497192 4775 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-5tgs8" Oct 02 01:55:42 crc kubenswrapper[4775]: I1002 01:55:42.498633 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Oct 02 01:55:42 crc kubenswrapper[4775]: I1002 01:55:42.509283 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4wqh4"] Oct 02 01:55:42 crc kubenswrapper[4775]: I1002 01:55:42.637276 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd9g6\" (UniqueName: \"kubernetes.io/projected/570f69cd-4e1a-4197-a992-307b5c42868d-kube-api-access-vd9g6\") pod \"cert-manager-operator-controller-manager-57cd46d6d-4wqh4\" (UID: \"570f69cd-4e1a-4197-a992-307b5c42868d\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4wqh4" Oct 02 01:55:42 crc kubenswrapper[4775]: I1002 01:55:42.738521 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd9g6\" (UniqueName: \"kubernetes.io/projected/570f69cd-4e1a-4197-a992-307b5c42868d-kube-api-access-vd9g6\") pod \"cert-manager-operator-controller-manager-57cd46d6d-4wqh4\" (UID: \"570f69cd-4e1a-4197-a992-307b5c42868d\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4wqh4" Oct 02 01:55:42 crc kubenswrapper[4775]: I1002 01:55:42.776947 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd9g6\" (UniqueName: \"kubernetes.io/projected/570f69cd-4e1a-4197-a992-307b5c42868d-kube-api-access-vd9g6\") pod \"cert-manager-operator-controller-manager-57cd46d6d-4wqh4\" (UID: \"570f69cd-4e1a-4197-a992-307b5c42868d\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4wqh4" Oct 02 01:55:42 crc kubenswrapper[4775]: I1002 01:55:42.857127 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4wqh4" Oct 02 01:55:43 crc kubenswrapper[4775]: I1002 01:55:43.279369 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4wqh4"] Oct 02 01:55:43 crc kubenswrapper[4775]: W1002 01:55:43.293654 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod570f69cd_4e1a_4197_a992_307b5c42868d.slice/crio-4d4af32c902c0bcebfde8d84f5811ef89cce2d4f6f89a7ff8f5ddca77c29c4f0 WatchSource:0}: Error finding container 4d4af32c902c0bcebfde8d84f5811ef89cce2d4f6f89a7ff8f5ddca77c29c4f0: Status 404 returned error can't find the container with id 4d4af32c902c0bcebfde8d84f5811ef89cce2d4f6f89a7ff8f5ddca77c29c4f0 Oct 02 01:55:43 crc kubenswrapper[4775]: I1002 01:55:43.913659 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4wqh4" event={"ID":"570f69cd-4e1a-4197-a992-307b5c42868d","Type":"ContainerStarted","Data":"4d4af32c902c0bcebfde8d84f5811ef89cce2d4f6f89a7ff8f5ddca77c29c4f0"} Oct 02 01:55:50 crc kubenswrapper[4775]: I1002 01:55:50.990993 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4wqh4" event={"ID":"570f69cd-4e1a-4197-a992-307b5c42868d","Type":"ContainerStarted","Data":"f9c78427bc59572d962124713e60862cc62dd866dea9fd8858cae5b53aefd963"} Oct 02 01:55:51 crc kubenswrapper[4775]: I1002 01:55:51.019592 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-57cd46d6d-4wqh4" podStartSLOduration=2.171904897 podStartE2EDuration="9.019569254s" podCreationTimestamp="2025-10-02 01:55:42 +0000 UTC" firstStartedPulling="2025-10-02 01:55:43.297241341 +0000 UTC m=+880.463985401" lastFinishedPulling="2025-10-02 01:55:50.144905688 +0000 UTC m=+887.311649758" observedRunningTime="2025-10-02 01:55:51.016172048 +0000 UTC m=+888.182916078" watchObservedRunningTime="2025-10-02 01:55:51.019569254 +0000 UTC m=+888.186313334" Oct 02 01:55:53 crc kubenswrapper[4775]: I1002 01:55:53.155507 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-m456n"] Oct 02 01:55:53 crc kubenswrapper[4775]: I1002 01:55:53.156652 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-m456n" Oct 02 01:55:53 crc kubenswrapper[4775]: I1002 01:55:53.160434 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 02 01:55:53 crc kubenswrapper[4775]: I1002 01:55:53.160752 4775 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-47g7b" Oct 02 01:55:53 crc kubenswrapper[4775]: I1002 01:55:53.162592 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 02 01:55:53 crc kubenswrapper[4775]: I1002 01:55:53.167561 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-m456n"] Oct 02 01:55:53 crc kubenswrapper[4775]: I1002 01:55:53.306034 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnl2g\" (UniqueName: \"kubernetes.io/projected/d130200c-29a8-45ca-8497-fec6226297a3-kube-api-access-tnl2g\") pod \"cert-manager-webhook-d969966f-m456n\" (UID: \"d130200c-29a8-45ca-8497-fec6226297a3\") " pod="cert-manager/cert-manager-webhook-d969966f-m456n" Oct 02 01:55:53 crc kubenswrapper[4775]: I1002 01:55:53.306692 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d130200c-29a8-45ca-8497-fec6226297a3-bound-sa-token\") pod \"cert-manager-webhook-d969966f-m456n\" (UID: \"d130200c-29a8-45ca-8497-fec6226297a3\") " pod="cert-manager/cert-manager-webhook-d969966f-m456n" Oct 02 01:55:53 crc kubenswrapper[4775]: I1002 01:55:53.408140 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnl2g\" (UniqueName: \"kubernetes.io/projected/d130200c-29a8-45ca-8497-fec6226297a3-kube-api-access-tnl2g\") pod \"cert-manager-webhook-d969966f-m456n\" (UID: \"d130200c-29a8-45ca-8497-fec6226297a3\") " pod="cert-manager/cert-manager-webhook-d969966f-m456n" Oct 02 01:55:53 crc kubenswrapper[4775]: I1002 01:55:53.408294 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d130200c-29a8-45ca-8497-fec6226297a3-bound-sa-token\") pod \"cert-manager-webhook-d969966f-m456n\" (UID: \"d130200c-29a8-45ca-8497-fec6226297a3\") " pod="cert-manager/cert-manager-webhook-d969966f-m456n" Oct 02 01:55:53 crc kubenswrapper[4775]: I1002 01:55:53.429789 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnl2g\" (UniqueName: \"kubernetes.io/projected/d130200c-29a8-45ca-8497-fec6226297a3-kube-api-access-tnl2g\") pod \"cert-manager-webhook-d969966f-m456n\" (UID: \"d130200c-29a8-45ca-8497-fec6226297a3\") " pod="cert-manager/cert-manager-webhook-d969966f-m456n" Oct 02 01:55:53 crc kubenswrapper[4775]: I1002 01:55:53.429796 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d130200c-29a8-45ca-8497-fec6226297a3-bound-sa-token\") pod \"cert-manager-webhook-d969966f-m456n\" (UID: \"d130200c-29a8-45ca-8497-fec6226297a3\") " pod="cert-manager/cert-manager-webhook-d969966f-m456n" Oct 02 01:55:53 crc kubenswrapper[4775]: I1002 01:55:53.475710 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-d969966f-m456n" Oct 02 01:55:53 crc kubenswrapper[4775]: I1002 01:55:53.981862 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-d969966f-m456n"] Oct 02 01:55:55 crc kubenswrapper[4775]: I1002 01:55:55.014994 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-m456n" event={"ID":"d130200c-29a8-45ca-8497-fec6226297a3","Type":"ContainerStarted","Data":"a88a4903edf3482500fded579f6afa8b0ffcb928ccd417fb4c30c73bad621875"} Oct 02 01:55:55 crc kubenswrapper[4775]: I1002 01:55:55.227888 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-ljzw5"] Oct 02 01:55:55 crc kubenswrapper[4775]: I1002 01:55:55.229152 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ljzw5" Oct 02 01:55:55 crc kubenswrapper[4775]: I1002 01:55:55.231835 4775 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-rb8m2" Oct 02 01:55:55 crc kubenswrapper[4775]: I1002 01:55:55.245078 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-ljzw5"] Oct 02 01:55:55 crc kubenswrapper[4775]: I1002 01:55:55.340775 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-865gn\" (UniqueName: \"kubernetes.io/projected/0a404f98-18c9-437f-9da0-b3075699cc4a-kube-api-access-865gn\") pod \"cert-manager-cainjector-7d9f95dbf-ljzw5\" (UID: \"0a404f98-18c9-437f-9da0-b3075699cc4a\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ljzw5" Oct 02 01:55:55 crc kubenswrapper[4775]: I1002 01:55:55.340830 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0a404f98-18c9-437f-9da0-b3075699cc4a-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-ljzw5\" (UID: \"0a404f98-18c9-437f-9da0-b3075699cc4a\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ljzw5" Oct 02 01:55:55 crc kubenswrapper[4775]: I1002 01:55:55.441500 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-865gn\" (UniqueName: \"kubernetes.io/projected/0a404f98-18c9-437f-9da0-b3075699cc4a-kube-api-access-865gn\") pod \"cert-manager-cainjector-7d9f95dbf-ljzw5\" (UID: \"0a404f98-18c9-437f-9da0-b3075699cc4a\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ljzw5" Oct 02 01:55:55 crc kubenswrapper[4775]: I1002 01:55:55.441563 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0a404f98-18c9-437f-9da0-b3075699cc4a-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-ljzw5\" (UID: \"0a404f98-18c9-437f-9da0-b3075699cc4a\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ljzw5" Oct 02 01:55:55 crc kubenswrapper[4775]: I1002 01:55:55.466135 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-865gn\" (UniqueName: \"kubernetes.io/projected/0a404f98-18c9-437f-9da0-b3075699cc4a-kube-api-access-865gn\") pod \"cert-manager-cainjector-7d9f95dbf-ljzw5\" (UID: \"0a404f98-18c9-437f-9da0-b3075699cc4a\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ljzw5" Oct 02 01:55:55 crc kubenswrapper[4775]: I1002 01:55:55.469315 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0a404f98-18c9-437f-9da0-b3075699cc4a-bound-sa-token\") pod \"cert-manager-cainjector-7d9f95dbf-ljzw5\" (UID: \"0a404f98-18c9-437f-9da0-b3075699cc4a\") " pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ljzw5" Oct 02 01:55:55 crc kubenswrapper[4775]: I1002 01:55:55.552323 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ljzw5" Oct 02 01:55:56 crc kubenswrapper[4775]: I1002 01:55:56.017466 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7d9f95dbf-ljzw5"] Oct 02 01:55:56 crc kubenswrapper[4775]: W1002 01:55:56.042324 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a404f98_18c9_437f_9da0_b3075699cc4a.slice/crio-bc6c9af01eba50bfcb0e65b5149da5df27c31614a4e4af6c51401a12a998248a WatchSource:0}: Error finding container bc6c9af01eba50bfcb0e65b5149da5df27c31614a4e4af6c51401a12a998248a: Status 404 returned error can't find the container with id bc6c9af01eba50bfcb0e65b5149da5df27c31614a4e4af6c51401a12a998248a Oct 02 01:55:57 crc kubenswrapper[4775]: I1002 01:55:57.028533 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ljzw5" event={"ID":"0a404f98-18c9-437f-9da0-b3075699cc4a","Type":"ContainerStarted","Data":"bc6c9af01eba50bfcb0e65b5149da5df27c31614a4e4af6c51401a12a998248a"} Oct 02 01:55:59 crc kubenswrapper[4775]: I1002 01:55:59.055386 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-d969966f-m456n" event={"ID":"d130200c-29a8-45ca-8497-fec6226297a3","Type":"ContainerStarted","Data":"a8ff03d616142fc9ca3c4e830050be1dfc020496076e3e7f2e456d2aef0b0d24"} Oct 02 01:55:59 crc kubenswrapper[4775]: I1002 01:55:59.056722 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-d969966f-m456n" Oct 02 01:55:59 crc kubenswrapper[4775]: I1002 01:55:59.058542 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ljzw5" event={"ID":"0a404f98-18c9-437f-9da0-b3075699cc4a","Type":"ContainerStarted","Data":"16081554619c8c615b40c05ac76afc7dad39ebc63eee07fbf355978c9786dc38"} Oct 02 01:55:59 crc kubenswrapper[4775]: I1002 01:55:59.093418 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7d9f95dbf-ljzw5" podStartSLOduration=1.368596186 podStartE2EDuration="4.093404894s" podCreationTimestamp="2025-10-02 01:55:55 +0000 UTC" firstStartedPulling="2025-10-02 01:55:56.05481132 +0000 UTC m=+893.221555360" lastFinishedPulling="2025-10-02 01:55:58.779619988 +0000 UTC m=+895.946364068" observedRunningTime="2025-10-02 01:55:59.0901304 +0000 UTC m=+896.256874440" watchObservedRunningTime="2025-10-02 01:55:59.093404894 +0000 UTC m=+896.260148934" Oct 02 01:55:59 crc kubenswrapper[4775]: I1002 01:55:59.093801 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-d969966f-m456n" podStartSLOduration=1.359266801 podStartE2EDuration="6.093796654s" podCreationTimestamp="2025-10-02 01:55:53 +0000 UTC" firstStartedPulling="2025-10-02 01:55:54.012622197 +0000 UTC m=+891.179366267" lastFinishedPulling="2025-10-02 01:55:58.74715205 +0000 UTC m=+895.913896120" observedRunningTime="2025-10-02 01:55:59.076397355 +0000 UTC m=+896.243141395" watchObservedRunningTime="2025-10-02 01:55:59.093796654 +0000 UTC m=+896.260540694" Oct 02 01:56:03 crc kubenswrapper[4775]: I1002 01:56:03.479192 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-d969966f-m456n" Oct 02 01:56:03 crc kubenswrapper[4775]: I1002 01:56:03.878306 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-6lsvs"] Oct 02 01:56:03 crc kubenswrapper[4775]: I1002 01:56:03.879759 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-6lsvs" Oct 02 01:56:03 crc kubenswrapper[4775]: I1002 01:56:03.882553 4775 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-jskpr" Oct 02 01:56:03 crc kubenswrapper[4775]: I1002 01:56:03.887364 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-6lsvs"] Oct 02 01:56:03 crc kubenswrapper[4775]: I1002 01:56:03.983358 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwqgg\" (UniqueName: \"kubernetes.io/projected/8c968ad9-3556-49c9-859b-323223695d9e-kube-api-access-rwqgg\") pod \"cert-manager-7d4cc89fcb-6lsvs\" (UID: \"8c968ad9-3556-49c9-859b-323223695d9e\") " pod="cert-manager/cert-manager-7d4cc89fcb-6lsvs" Oct 02 01:56:03 crc kubenswrapper[4775]: I1002 01:56:03.983568 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c968ad9-3556-49c9-859b-323223695d9e-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-6lsvs\" (UID: \"8c968ad9-3556-49c9-859b-323223695d9e\") " pod="cert-manager/cert-manager-7d4cc89fcb-6lsvs" Oct 02 01:56:04 crc kubenswrapper[4775]: I1002 01:56:04.085033 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c968ad9-3556-49c9-859b-323223695d9e-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-6lsvs\" (UID: \"8c968ad9-3556-49c9-859b-323223695d9e\") " pod="cert-manager/cert-manager-7d4cc89fcb-6lsvs" Oct 02 01:56:04 crc kubenswrapper[4775]: I1002 01:56:04.085222 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwqgg\" (UniqueName: \"kubernetes.io/projected/8c968ad9-3556-49c9-859b-323223695d9e-kube-api-access-rwqgg\") pod \"cert-manager-7d4cc89fcb-6lsvs\" (UID: \"8c968ad9-3556-49c9-859b-323223695d9e\") " pod="cert-manager/cert-manager-7d4cc89fcb-6lsvs" Oct 02 01:56:04 crc kubenswrapper[4775]: I1002 01:56:04.107118 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8c968ad9-3556-49c9-859b-323223695d9e-bound-sa-token\") pod \"cert-manager-7d4cc89fcb-6lsvs\" (UID: \"8c968ad9-3556-49c9-859b-323223695d9e\") " pod="cert-manager/cert-manager-7d4cc89fcb-6lsvs" Oct 02 01:56:04 crc kubenswrapper[4775]: I1002 01:56:04.113490 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwqgg\" (UniqueName: \"kubernetes.io/projected/8c968ad9-3556-49c9-859b-323223695d9e-kube-api-access-rwqgg\") pod \"cert-manager-7d4cc89fcb-6lsvs\" (UID: \"8c968ad9-3556-49c9-859b-323223695d9e\") " pod="cert-manager/cert-manager-7d4cc89fcb-6lsvs" Oct 02 01:56:04 crc kubenswrapper[4775]: I1002 01:56:04.203918 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-7d4cc89fcb-6lsvs" Oct 02 01:56:04 crc kubenswrapper[4775]: I1002 01:56:04.702756 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-7d4cc89fcb-6lsvs"] Oct 02 01:56:05 crc kubenswrapper[4775]: I1002 01:56:05.100909 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-6lsvs" event={"ID":"8c968ad9-3556-49c9-859b-323223695d9e","Type":"ContainerStarted","Data":"1373217a788ff442c6fdfe644ffa56de1b41fede93156b6d38f15eb3f681b0a9"} Oct 02 01:56:07 crc kubenswrapper[4775]: I1002 01:56:07.233675 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:56:07 crc kubenswrapper[4775]: I1002 01:56:07.234031 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:56:09 crc kubenswrapper[4775]: I1002 01:56:09.138403 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-7d4cc89fcb-6lsvs" event={"ID":"8c968ad9-3556-49c9-859b-323223695d9e","Type":"ContainerStarted","Data":"b091861980072e0a0020ad16d88bceefdf14278e052053dc244c104d2d7a1f75"} Oct 02 01:56:17 crc kubenswrapper[4775]: I1002 01:56:17.242150 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-7d4cc89fcb-6lsvs" podStartSLOduration=14.242125304 podStartE2EDuration="14.242125304s" podCreationTimestamp="2025-10-02 01:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:56:09.187383587 +0000 UTC m=+906.354127657" watchObservedRunningTime="2025-10-02 01:56:17.242125304 +0000 UTC m=+914.408869384" Oct 02 01:56:17 crc kubenswrapper[4775]: I1002 01:56:17.245740 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-7qtw6"] Oct 02 01:56:17 crc kubenswrapper[4775]: I1002 01:56:17.246983 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7qtw6" Oct 02 01:56:17 crc kubenswrapper[4775]: I1002 01:56:17.250652 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-68x6t" Oct 02 01:56:17 crc kubenswrapper[4775]: I1002 01:56:17.256444 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 02 01:56:17 crc kubenswrapper[4775]: I1002 01:56:17.257095 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 02 01:56:17 crc kubenswrapper[4775]: I1002 01:56:17.272349 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7qtw6"] Oct 02 01:56:17 crc kubenswrapper[4775]: I1002 01:56:17.279395 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrghz\" (UniqueName: \"kubernetes.io/projected/044c5add-4c62-4809-bbf1-bf874be969fd-kube-api-access-vrghz\") pod \"openstack-operator-index-7qtw6\" (UID: \"044c5add-4c62-4809-bbf1-bf874be969fd\") " pod="openstack-operators/openstack-operator-index-7qtw6" Oct 02 01:56:17 crc kubenswrapper[4775]: I1002 01:56:17.380233 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrghz\" (UniqueName: \"kubernetes.io/projected/044c5add-4c62-4809-bbf1-bf874be969fd-kube-api-access-vrghz\") pod \"openstack-operator-index-7qtw6\" (UID: \"044c5add-4c62-4809-bbf1-bf874be969fd\") " pod="openstack-operators/openstack-operator-index-7qtw6" Oct 02 01:56:17 crc kubenswrapper[4775]: I1002 01:56:17.402582 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrghz\" (UniqueName: \"kubernetes.io/projected/044c5add-4c62-4809-bbf1-bf874be969fd-kube-api-access-vrghz\") pod \"openstack-operator-index-7qtw6\" (UID: \"044c5add-4c62-4809-bbf1-bf874be969fd\") " pod="openstack-operators/openstack-operator-index-7qtw6" Oct 02 01:56:17 crc kubenswrapper[4775]: I1002 01:56:17.571450 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7qtw6" Oct 02 01:56:18 crc kubenswrapper[4775]: I1002 01:56:18.039672 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-7qtw6"] Oct 02 01:56:18 crc kubenswrapper[4775]: W1002 01:56:18.046198 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod044c5add_4c62_4809_bbf1_bf874be969fd.slice/crio-92a4db0480f838b0fae4825b2ba3c468acc20bcf104c858df75d3268e036eb4e WatchSource:0}: Error finding container 92a4db0480f838b0fae4825b2ba3c468acc20bcf104c858df75d3268e036eb4e: Status 404 returned error can't find the container with id 92a4db0480f838b0fae4825b2ba3c468acc20bcf104c858df75d3268e036eb4e Oct 02 01:56:18 crc kubenswrapper[4775]: I1002 01:56:18.207103 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7qtw6" event={"ID":"044c5add-4c62-4809-bbf1-bf874be969fd","Type":"ContainerStarted","Data":"92a4db0480f838b0fae4825b2ba3c468acc20bcf104c858df75d3268e036eb4e"} Oct 02 01:56:20 crc kubenswrapper[4775]: I1002 01:56:20.602152 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7qtw6"] Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.210732 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-f8jlc"] Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.212618 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-f8jlc" Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.225913 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-f8jlc"] Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.236859 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnmfd\" (UniqueName: \"kubernetes.io/projected/b5b65461-e031-4f60-aa2b-a63a906053c4-kube-api-access-nnmfd\") pod \"openstack-operator-index-f8jlc\" (UID: \"b5b65461-e031-4f60-aa2b-a63a906053c4\") " pod="openstack-operators/openstack-operator-index-f8jlc" Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.245767 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7qtw6" event={"ID":"044c5add-4c62-4809-bbf1-bf874be969fd","Type":"ContainerStarted","Data":"2280f31e35ae486384ed9bb2e9bbc7eca74ae5754529c5812377ef655c1017ff"} Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.245901 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-7qtw6" podUID="044c5add-4c62-4809-bbf1-bf874be969fd" containerName="registry-server" containerID="cri-o://2280f31e35ae486384ed9bb2e9bbc7eca74ae5754529c5812377ef655c1017ff" gracePeriod=2 Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.280294 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-7qtw6" podStartSLOduration=1.848916509 podStartE2EDuration="4.280267505s" podCreationTimestamp="2025-10-02 01:56:17 +0000 UTC" firstStartedPulling="2025-10-02 01:56:18.049324635 +0000 UTC m=+915.216068715" lastFinishedPulling="2025-10-02 01:56:20.480675671 +0000 UTC m=+917.647419711" observedRunningTime="2025-10-02 01:56:21.278239273 +0000 UTC m=+918.444983353" watchObservedRunningTime="2025-10-02 01:56:21.280267505 +0000 UTC m=+918.447011585" Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.337909 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnmfd\" (UniqueName: \"kubernetes.io/projected/b5b65461-e031-4f60-aa2b-a63a906053c4-kube-api-access-nnmfd\") pod \"openstack-operator-index-f8jlc\" (UID: \"b5b65461-e031-4f60-aa2b-a63a906053c4\") " pod="openstack-operators/openstack-operator-index-f8jlc" Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.369803 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnmfd\" (UniqueName: \"kubernetes.io/projected/b5b65461-e031-4f60-aa2b-a63a906053c4-kube-api-access-nnmfd\") pod \"openstack-operator-index-f8jlc\" (UID: \"b5b65461-e031-4f60-aa2b-a63a906053c4\") " pod="openstack-operators/openstack-operator-index-f8jlc" Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.542320 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-f8jlc" Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.687527 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7qtw6" Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.845151 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrghz\" (UniqueName: \"kubernetes.io/projected/044c5add-4c62-4809-bbf1-bf874be969fd-kube-api-access-vrghz\") pod \"044c5add-4c62-4809-bbf1-bf874be969fd\" (UID: \"044c5add-4c62-4809-bbf1-bf874be969fd\") " Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.852484 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/044c5add-4c62-4809-bbf1-bf874be969fd-kube-api-access-vrghz" (OuterVolumeSpecName: "kube-api-access-vrghz") pod "044c5add-4c62-4809-bbf1-bf874be969fd" (UID: "044c5add-4c62-4809-bbf1-bf874be969fd"). InnerVolumeSpecName "kube-api-access-vrghz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.947040 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrghz\" (UniqueName: \"kubernetes.io/projected/044c5add-4c62-4809-bbf1-bf874be969fd-kube-api-access-vrghz\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:21 crc kubenswrapper[4775]: I1002 01:56:21.975670 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-f8jlc"] Oct 02 01:56:21 crc kubenswrapper[4775]: W1002 01:56:21.986326 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5b65461_e031_4f60_aa2b_a63a906053c4.slice/crio-33245c814f56fbb8b9aae6ab385477e3aad93a30354078a8dd6276270db2c25f WatchSource:0}: Error finding container 33245c814f56fbb8b9aae6ab385477e3aad93a30354078a8dd6276270db2c25f: Status 404 returned error can't find the container with id 33245c814f56fbb8b9aae6ab385477e3aad93a30354078a8dd6276270db2c25f Oct 02 01:56:22 crc kubenswrapper[4775]: I1002 01:56:22.257777 4775 generic.go:334] "Generic (PLEG): container finished" podID="044c5add-4c62-4809-bbf1-bf874be969fd" containerID="2280f31e35ae486384ed9bb2e9bbc7eca74ae5754529c5812377ef655c1017ff" exitCode=0 Oct 02 01:56:22 crc kubenswrapper[4775]: I1002 01:56:22.258170 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-7qtw6" Oct 02 01:56:22 crc kubenswrapper[4775]: I1002 01:56:22.258066 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7qtw6" event={"ID":"044c5add-4c62-4809-bbf1-bf874be969fd","Type":"ContainerDied","Data":"2280f31e35ae486384ed9bb2e9bbc7eca74ae5754529c5812377ef655c1017ff"} Oct 02 01:56:22 crc kubenswrapper[4775]: I1002 01:56:22.258315 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-7qtw6" event={"ID":"044c5add-4c62-4809-bbf1-bf874be969fd","Type":"ContainerDied","Data":"92a4db0480f838b0fae4825b2ba3c468acc20bcf104c858df75d3268e036eb4e"} Oct 02 01:56:22 crc kubenswrapper[4775]: I1002 01:56:22.258347 4775 scope.go:117] "RemoveContainer" containerID="2280f31e35ae486384ed9bb2e9bbc7eca74ae5754529c5812377ef655c1017ff" Oct 02 01:56:22 crc kubenswrapper[4775]: I1002 01:56:22.261002 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-f8jlc" event={"ID":"b5b65461-e031-4f60-aa2b-a63a906053c4","Type":"ContainerStarted","Data":"f92f5fe63ba6d73f5c6a47dc2f4fd533f50753a09fb43f7c9e831abaa7afb2b1"} Oct 02 01:56:22 crc kubenswrapper[4775]: I1002 01:56:22.261045 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-f8jlc" event={"ID":"b5b65461-e031-4f60-aa2b-a63a906053c4","Type":"ContainerStarted","Data":"33245c814f56fbb8b9aae6ab385477e3aad93a30354078a8dd6276270db2c25f"} Oct 02 01:56:22 crc kubenswrapper[4775]: I1002 01:56:22.276410 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-f8jlc" podStartSLOduration=1.215646297 podStartE2EDuration="1.276388416s" podCreationTimestamp="2025-10-02 01:56:21 +0000 UTC" firstStartedPulling="2025-10-02 01:56:21.994634618 +0000 UTC m=+919.161378688" lastFinishedPulling="2025-10-02 01:56:22.055376727 +0000 UTC m=+919.222120807" observedRunningTime="2025-10-02 01:56:22.27538639 +0000 UTC m=+919.442130470" watchObservedRunningTime="2025-10-02 01:56:22.276388416 +0000 UTC m=+919.443132506" Oct 02 01:56:22 crc kubenswrapper[4775]: I1002 01:56:22.289838 4775 scope.go:117] "RemoveContainer" containerID="2280f31e35ae486384ed9bb2e9bbc7eca74ae5754529c5812377ef655c1017ff" Oct 02 01:56:22 crc kubenswrapper[4775]: E1002 01:56:22.290551 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2280f31e35ae486384ed9bb2e9bbc7eca74ae5754529c5812377ef655c1017ff\": container with ID starting with 2280f31e35ae486384ed9bb2e9bbc7eca74ae5754529c5812377ef655c1017ff not found: ID does not exist" containerID="2280f31e35ae486384ed9bb2e9bbc7eca74ae5754529c5812377ef655c1017ff" Oct 02 01:56:22 crc kubenswrapper[4775]: I1002 01:56:22.290633 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2280f31e35ae486384ed9bb2e9bbc7eca74ae5754529c5812377ef655c1017ff"} err="failed to get container status \"2280f31e35ae486384ed9bb2e9bbc7eca74ae5754529c5812377ef655c1017ff\": rpc error: code = NotFound desc = could not find container \"2280f31e35ae486384ed9bb2e9bbc7eca74ae5754529c5812377ef655c1017ff\": container with ID starting with 2280f31e35ae486384ed9bb2e9bbc7eca74ae5754529c5812377ef655c1017ff not found: ID does not exist" Oct 02 01:56:22 crc kubenswrapper[4775]: I1002 01:56:22.302990 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-7qtw6"] Oct 02 01:56:22 crc kubenswrapper[4775]: I1002 01:56:22.309678 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-7qtw6"] Oct 02 01:56:23 crc kubenswrapper[4775]: I1002 01:56:23.779583 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="044c5add-4c62-4809-bbf1-bf874be969fd" path="/var/lib/kubelet/pods/044c5add-4c62-4809-bbf1-bf874be969fd/volumes" Oct 02 01:56:31 crc kubenswrapper[4775]: I1002 01:56:31.543609 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-f8jlc" Oct 02 01:56:31 crc kubenswrapper[4775]: I1002 01:56:31.544472 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-f8jlc" Oct 02 01:56:31 crc kubenswrapper[4775]: I1002 01:56:31.592588 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-f8jlc" Oct 02 01:56:32 crc kubenswrapper[4775]: I1002 01:56:32.382588 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-f8jlc" Oct 02 01:56:37 crc kubenswrapper[4775]: I1002 01:56:37.234423 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:56:37 crc kubenswrapper[4775]: I1002 01:56:37.235144 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.050033 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm"] Oct 02 01:56:39 crc kubenswrapper[4775]: E1002 01:56:39.050685 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="044c5add-4c62-4809-bbf1-bf874be969fd" containerName="registry-server" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.050704 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="044c5add-4c62-4809-bbf1-bf874be969fd" containerName="registry-server" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.050901 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="044c5add-4c62-4809-bbf1-bf874be969fd" containerName="registry-server" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.052220 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.054450 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-bfkfx" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.061638 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm"] Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.201111 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/790fe7be-506d-4eb7-b754-8aa3e2856236-bundle\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm\" (UID: \"790fe7be-506d-4eb7-b754-8aa3e2856236\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.201160 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97bzs\" (UniqueName: \"kubernetes.io/projected/790fe7be-506d-4eb7-b754-8aa3e2856236-kube-api-access-97bzs\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm\" (UID: \"790fe7be-506d-4eb7-b754-8aa3e2856236\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.201202 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/790fe7be-506d-4eb7-b754-8aa3e2856236-util\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm\" (UID: \"790fe7be-506d-4eb7-b754-8aa3e2856236\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.303284 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/790fe7be-506d-4eb7-b754-8aa3e2856236-util\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm\" (UID: \"790fe7be-506d-4eb7-b754-8aa3e2856236\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.303384 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/790fe7be-506d-4eb7-b754-8aa3e2856236-bundle\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm\" (UID: \"790fe7be-506d-4eb7-b754-8aa3e2856236\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.303411 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97bzs\" (UniqueName: \"kubernetes.io/projected/790fe7be-506d-4eb7-b754-8aa3e2856236-kube-api-access-97bzs\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm\" (UID: \"790fe7be-506d-4eb7-b754-8aa3e2856236\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.303925 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/790fe7be-506d-4eb7-b754-8aa3e2856236-bundle\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm\" (UID: \"790fe7be-506d-4eb7-b754-8aa3e2856236\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.304016 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/790fe7be-506d-4eb7-b754-8aa3e2856236-util\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm\" (UID: \"790fe7be-506d-4eb7-b754-8aa3e2856236\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.336562 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97bzs\" (UniqueName: \"kubernetes.io/projected/790fe7be-506d-4eb7-b754-8aa3e2856236-kube-api-access-97bzs\") pod \"cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm\" (UID: \"790fe7be-506d-4eb7-b754-8aa3e2856236\") " pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.377102 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" Oct 02 01:56:39 crc kubenswrapper[4775]: I1002 01:56:39.613702 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm"] Oct 02 01:56:40 crc kubenswrapper[4775]: I1002 01:56:40.403265 4775 generic.go:334] "Generic (PLEG): container finished" podID="790fe7be-506d-4eb7-b754-8aa3e2856236" containerID="bbafeef4d5f076eda57f5be3fb0a80c280a5d9eaa08ca51dd74934fd6827b481" exitCode=0 Oct 02 01:56:40 crc kubenswrapper[4775]: I1002 01:56:40.403302 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" event={"ID":"790fe7be-506d-4eb7-b754-8aa3e2856236","Type":"ContainerDied","Data":"bbafeef4d5f076eda57f5be3fb0a80c280a5d9eaa08ca51dd74934fd6827b481"} Oct 02 01:56:40 crc kubenswrapper[4775]: I1002 01:56:40.403326 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" event={"ID":"790fe7be-506d-4eb7-b754-8aa3e2856236","Type":"ContainerStarted","Data":"f685d2ac6bd03a17c1c10b02d934258e744366dee5e47bb4aa88cb13da90172c"} Oct 02 01:56:41 crc kubenswrapper[4775]: I1002 01:56:41.411464 4775 generic.go:334] "Generic (PLEG): container finished" podID="790fe7be-506d-4eb7-b754-8aa3e2856236" containerID="f5a57ab9fe68ea628c980e41f6020979075457ebad893787c9917459d0a87bf6" exitCode=0 Oct 02 01:56:41 crc kubenswrapper[4775]: I1002 01:56:41.411798 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" event={"ID":"790fe7be-506d-4eb7-b754-8aa3e2856236","Type":"ContainerDied","Data":"f5a57ab9fe68ea628c980e41f6020979075457ebad893787c9917459d0a87bf6"} Oct 02 01:56:42 crc kubenswrapper[4775]: I1002 01:56:42.420337 4775 generic.go:334] "Generic (PLEG): container finished" podID="790fe7be-506d-4eb7-b754-8aa3e2856236" containerID="6b6887bbd7dc2454f733494c600d89e6c81f938b57b79c76718dfe9832cd270c" exitCode=0 Oct 02 01:56:42 crc kubenswrapper[4775]: I1002 01:56:42.420397 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" event={"ID":"790fe7be-506d-4eb7-b754-8aa3e2856236","Type":"ContainerDied","Data":"6b6887bbd7dc2454f733494c600d89e6c81f938b57b79c76718dfe9832cd270c"} Oct 02 01:56:43 crc kubenswrapper[4775]: I1002 01:56:43.716370 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" Oct 02 01:56:43 crc kubenswrapper[4775]: I1002 01:56:43.769068 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/790fe7be-506d-4eb7-b754-8aa3e2856236-util\") pod \"790fe7be-506d-4eb7-b754-8aa3e2856236\" (UID: \"790fe7be-506d-4eb7-b754-8aa3e2856236\") " Oct 02 01:56:43 crc kubenswrapper[4775]: I1002 01:56:43.769129 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97bzs\" (UniqueName: \"kubernetes.io/projected/790fe7be-506d-4eb7-b754-8aa3e2856236-kube-api-access-97bzs\") pod \"790fe7be-506d-4eb7-b754-8aa3e2856236\" (UID: \"790fe7be-506d-4eb7-b754-8aa3e2856236\") " Oct 02 01:56:43 crc kubenswrapper[4775]: I1002 01:56:43.769158 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/790fe7be-506d-4eb7-b754-8aa3e2856236-bundle\") pod \"790fe7be-506d-4eb7-b754-8aa3e2856236\" (UID: \"790fe7be-506d-4eb7-b754-8aa3e2856236\") " Oct 02 01:56:43 crc kubenswrapper[4775]: I1002 01:56:43.772648 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/790fe7be-506d-4eb7-b754-8aa3e2856236-bundle" (OuterVolumeSpecName: "bundle") pod "790fe7be-506d-4eb7-b754-8aa3e2856236" (UID: "790fe7be-506d-4eb7-b754-8aa3e2856236"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:56:43 crc kubenswrapper[4775]: I1002 01:56:43.778855 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/790fe7be-506d-4eb7-b754-8aa3e2856236-kube-api-access-97bzs" (OuterVolumeSpecName: "kube-api-access-97bzs") pod "790fe7be-506d-4eb7-b754-8aa3e2856236" (UID: "790fe7be-506d-4eb7-b754-8aa3e2856236"). InnerVolumeSpecName "kube-api-access-97bzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:56:43 crc kubenswrapper[4775]: I1002 01:56:43.790341 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/790fe7be-506d-4eb7-b754-8aa3e2856236-util" (OuterVolumeSpecName: "util") pod "790fe7be-506d-4eb7-b754-8aa3e2856236" (UID: "790fe7be-506d-4eb7-b754-8aa3e2856236"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:56:43 crc kubenswrapper[4775]: I1002 01:56:43.870130 4775 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/790fe7be-506d-4eb7-b754-8aa3e2856236-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:43 crc kubenswrapper[4775]: I1002 01:56:43.870161 4775 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/790fe7be-506d-4eb7-b754-8aa3e2856236-util\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:43 crc kubenswrapper[4775]: I1002 01:56:43.870170 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97bzs\" (UniqueName: \"kubernetes.io/projected/790fe7be-506d-4eb7-b754-8aa3e2856236-kube-api-access-97bzs\") on node \"crc\" DevicePath \"\"" Oct 02 01:56:44 crc kubenswrapper[4775]: I1002 01:56:44.441301 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" event={"ID":"790fe7be-506d-4eb7-b754-8aa3e2856236","Type":"ContainerDied","Data":"f685d2ac6bd03a17c1c10b02d934258e744366dee5e47bb4aa88cb13da90172c"} Oct 02 01:56:44 crc kubenswrapper[4775]: I1002 01:56:44.441374 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f685d2ac6bd03a17c1c10b02d934258e744366dee5e47bb4aa88cb13da90172c" Oct 02 01:56:44 crc kubenswrapper[4775]: I1002 01:56:44.441385 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm" Oct 02 01:56:46 crc kubenswrapper[4775]: I1002 01:56:46.161002 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6bcc7488c-bprqn"] Oct 02 01:56:46 crc kubenswrapper[4775]: E1002 01:56:46.161602 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790fe7be-506d-4eb7-b754-8aa3e2856236" containerName="pull" Oct 02 01:56:46 crc kubenswrapper[4775]: I1002 01:56:46.161621 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="790fe7be-506d-4eb7-b754-8aa3e2856236" containerName="pull" Oct 02 01:56:46 crc kubenswrapper[4775]: E1002 01:56:46.161680 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790fe7be-506d-4eb7-b754-8aa3e2856236" containerName="util" Oct 02 01:56:46 crc kubenswrapper[4775]: I1002 01:56:46.161692 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="790fe7be-506d-4eb7-b754-8aa3e2856236" containerName="util" Oct 02 01:56:46 crc kubenswrapper[4775]: E1002 01:56:46.161714 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="790fe7be-506d-4eb7-b754-8aa3e2856236" containerName="extract" Oct 02 01:56:46 crc kubenswrapper[4775]: I1002 01:56:46.161726 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="790fe7be-506d-4eb7-b754-8aa3e2856236" containerName="extract" Oct 02 01:56:46 crc kubenswrapper[4775]: I1002 01:56:46.161994 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="790fe7be-506d-4eb7-b754-8aa3e2856236" containerName="extract" Oct 02 01:56:46 crc kubenswrapper[4775]: I1002 01:56:46.163128 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bprqn" Oct 02 01:56:46 crc kubenswrapper[4775]: I1002 01:56:46.166368 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-nzbt8" Oct 02 01:56:46 crc kubenswrapper[4775]: I1002 01:56:46.179641 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6bcc7488c-bprqn"] Oct 02 01:56:46 crc kubenswrapper[4775]: I1002 01:56:46.199917 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pkmx\" (UniqueName: \"kubernetes.io/projected/f08455ed-dbee-4865-885e-c92d82da7457-kube-api-access-2pkmx\") pod \"openstack-operator-controller-operator-6bcc7488c-bprqn\" (UID: \"f08455ed-dbee-4865-885e-c92d82da7457\") " pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bprqn" Oct 02 01:56:46 crc kubenswrapper[4775]: I1002 01:56:46.301470 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pkmx\" (UniqueName: \"kubernetes.io/projected/f08455ed-dbee-4865-885e-c92d82da7457-kube-api-access-2pkmx\") pod \"openstack-operator-controller-operator-6bcc7488c-bprqn\" (UID: \"f08455ed-dbee-4865-885e-c92d82da7457\") " pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bprqn" Oct 02 01:56:46 crc kubenswrapper[4775]: I1002 01:56:46.325036 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pkmx\" (UniqueName: \"kubernetes.io/projected/f08455ed-dbee-4865-885e-c92d82da7457-kube-api-access-2pkmx\") pod \"openstack-operator-controller-operator-6bcc7488c-bprqn\" (UID: \"f08455ed-dbee-4865-885e-c92d82da7457\") " pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bprqn" Oct 02 01:56:46 crc kubenswrapper[4775]: I1002 01:56:46.482200 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bprqn" Oct 02 01:56:46 crc kubenswrapper[4775]: I1002 01:56:46.715569 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6bcc7488c-bprqn"] Oct 02 01:56:47 crc kubenswrapper[4775]: I1002 01:56:47.456855 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bprqn" event={"ID":"f08455ed-dbee-4865-885e-c92d82da7457","Type":"ContainerStarted","Data":"e5cef8cdaabaab7916318d3770c1ec4c1fe2442ce3df9387e07e556f8a08da0a"} Oct 02 01:56:52 crc kubenswrapper[4775]: I1002 01:56:52.505116 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bprqn" event={"ID":"f08455ed-dbee-4865-885e-c92d82da7457","Type":"ContainerStarted","Data":"9227ab4e968394bd6ae5f920395ac6015ebc9ea4a0c677747cfe467d0723d8db"} Oct 02 01:56:54 crc kubenswrapper[4775]: I1002 01:56:54.521675 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bprqn" event={"ID":"f08455ed-dbee-4865-885e-c92d82da7457","Type":"ContainerStarted","Data":"027e1c165fdebfce2365919dcfa6a5543de7e205160f2bdb20e97da6b0fe6ca7"} Oct 02 01:56:54 crc kubenswrapper[4775]: I1002 01:56:54.522183 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bprqn" Oct 02 01:56:54 crc kubenswrapper[4775]: I1002 01:56:54.557042 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bprqn" podStartSLOduration=1.417915323 podStartE2EDuration="8.557014026s" podCreationTimestamp="2025-10-02 01:56:46 +0000 UTC" firstStartedPulling="2025-10-02 01:56:46.725546799 +0000 UTC m=+943.892290849" lastFinishedPulling="2025-10-02 01:56:53.864645502 +0000 UTC m=+951.031389552" observedRunningTime="2025-10-02 01:56:54.553572508 +0000 UTC m=+951.720316548" watchObservedRunningTime="2025-10-02 01:56:54.557014026 +0000 UTC m=+951.723758106" Oct 02 01:56:56 crc kubenswrapper[4775]: I1002 01:56:56.485752 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6bcc7488c-bprqn" Oct 02 01:57:07 crc kubenswrapper[4775]: I1002 01:57:07.233845 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:57:07 crc kubenswrapper[4775]: I1002 01:57:07.234394 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:57:07 crc kubenswrapper[4775]: I1002 01:57:07.234445 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 01:57:07 crc kubenswrapper[4775]: I1002 01:57:07.235105 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d8df09487b33326d5ae12f2261c6dc85c52809ba9bfd51eac55d13c8fc891e6"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 01:57:07 crc kubenswrapper[4775]: I1002 01:57:07.235169 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://8d8df09487b33326d5ae12f2261c6dc85c52809ba9bfd51eac55d13c8fc891e6" gracePeriod=600 Oct 02 01:57:07 crc kubenswrapper[4775]: E1002 01:57:07.381228 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc390c5b3_f65b_4f2e_9d03_bb09cf613134.slice/crio-8d8df09487b33326d5ae12f2261c6dc85c52809ba9bfd51eac55d13c8fc891e6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc390c5b3_f65b_4f2e_9d03_bb09cf613134.slice/crio-conmon-8d8df09487b33326d5ae12f2261c6dc85c52809ba9bfd51eac55d13c8fc891e6.scope\": RecentStats: unable to find data in memory cache]" Oct 02 01:57:07 crc kubenswrapper[4775]: I1002 01:57:07.629539 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="8d8df09487b33326d5ae12f2261c6dc85c52809ba9bfd51eac55d13c8fc891e6" exitCode=0 Oct 02 01:57:07 crc kubenswrapper[4775]: I1002 01:57:07.629639 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"8d8df09487b33326d5ae12f2261c6dc85c52809ba9bfd51eac55d13c8fc891e6"} Oct 02 01:57:07 crc kubenswrapper[4775]: I1002 01:57:07.629816 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"61c8c995ae4362ef1e859023c2c2ffc70bf9e259d47718d71624878bd8dc02ec"} Oct 02 01:57:07 crc kubenswrapper[4775]: I1002 01:57:07.629837 4775 scope.go:117] "RemoveContainer" containerID="0af4ec8f881f09471035912f64ffe17497158c6365d4ee31b994b4425f60c6bb" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.406683 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-5wrms"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.408140 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-5wrms" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.410744 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-hh8bb" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.438565 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-mx8pj"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.439853 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-mx8pj" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.442065 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-bnw6w" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.442855 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-5wrms"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.462878 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-mx8pj"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.474579 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-md78z"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.475457 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-md78z" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.478923 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-p84wv" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.480359 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-4gjk2"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.481339 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-4gjk2" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.487578 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-8l5d4" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.502788 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-4gjk2"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.509883 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-md78z"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.541049 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.542059 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.545582 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nltsc\" (UniqueName: \"kubernetes.io/projected/c2757728-9f90-4b12-8a40-3a7845c4f461-kube-api-access-nltsc\") pod \"barbican-operator-controller-manager-6ff8b75857-5wrms\" (UID: \"c2757728-9f90-4b12-8a40-3a7845c4f461\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-5wrms" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.545663 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9knj\" (UniqueName: \"kubernetes.io/projected/63b4a9a2-62d3-48a8-a2ca-74496dcff908-kube-api-access-q9knj\") pod \"cinder-operator-controller-manager-644bddb6d8-mx8pj\" (UID: \"63b4a9a2-62d3-48a8-a2ca-74496dcff908\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-mx8pj" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.545921 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-5kj5j" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.555577 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.556585 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.559795 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-zmsqn" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.567127 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.573147 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.585272 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.586200 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.590645 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-tqds4" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.590801 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.605943 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.621448 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-f2khq"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.622359 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-f2khq" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.635462 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-c97qw" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.641474 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jvrxr"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.642359 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jvrxr" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.646397 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-f2khq"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.647394 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29hk7\" (UniqueName: \"kubernetes.io/projected/7e0f47de-d77c-4840-a382-ed6a911f20b1-kube-api-access-29hk7\") pod \"designate-operator-controller-manager-84f4f7b77b-4gjk2\" (UID: \"7e0f47de-d77c-4840-a382-ed6a911f20b1\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-4gjk2" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.647439 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jpht\" (UniqueName: \"kubernetes.io/projected/68306a2f-6d51-4de9-9645-d3f0e3bc6e61-kube-api-access-9jpht\") pod \"glance-operator-controller-manager-84958c4d49-md78z\" (UID: \"68306a2f-6d51-4de9-9645-d3f0e3bc6e61\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-md78z" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.647486 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9knj\" (UniqueName: \"kubernetes.io/projected/63b4a9a2-62d3-48a8-a2ca-74496dcff908-kube-api-access-q9knj\") pod \"cinder-operator-controller-manager-644bddb6d8-mx8pj\" (UID: \"63b4a9a2-62d3-48a8-a2ca-74496dcff908\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-mx8pj" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.647516 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctwj6\" (UniqueName: \"kubernetes.io/projected/c815738f-cb2a-4136-9320-f75ceea923c5-kube-api-access-ctwj6\") pod \"heat-operator-controller-manager-5d889d78cf-5nl49\" (UID: \"c815738f-cb2a-4136-9320-f75ceea923c5\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.647558 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz7gc\" (UniqueName: \"kubernetes.io/projected/03d2a01f-1564-46f8-9648-e4a826470b44-kube-api-access-sz7gc\") pod \"horizon-operator-controller-manager-9f4696d94-2mxjr\" (UID: \"03d2a01f-1564-46f8-9648-e4a826470b44\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.647589 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nltsc\" (UniqueName: \"kubernetes.io/projected/c2757728-9f90-4b12-8a40-3a7845c4f461-kube-api-access-nltsc\") pod \"barbican-operator-controller-manager-6ff8b75857-5wrms\" (UID: \"c2757728-9f90-4b12-8a40-3a7845c4f461\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-5wrms" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.653224 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-bq8h9" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.657556 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jvrxr"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.665896 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.666896 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.669427 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-gznjv" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.684882 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9knj\" (UniqueName: \"kubernetes.io/projected/63b4a9a2-62d3-48a8-a2ca-74496dcff908-kube-api-access-q9knj\") pod \"cinder-operator-controller-manager-644bddb6d8-mx8pj\" (UID: \"63b4a9a2-62d3-48a8-a2ca-74496dcff908\") " pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-mx8pj" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.690867 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.698514 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nltsc\" (UniqueName: \"kubernetes.io/projected/c2757728-9f90-4b12-8a40-3a7845c4f461-kube-api-access-nltsc\") pod \"barbican-operator-controller-manager-6ff8b75857-5wrms\" (UID: \"c2757728-9f90-4b12-8a40-3a7845c4f461\") " pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-5wrms" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.728252 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-5wrms" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.753041 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29hk7\" (UniqueName: \"kubernetes.io/projected/7e0f47de-d77c-4840-a382-ed6a911f20b1-kube-api-access-29hk7\") pod \"designate-operator-controller-manager-84f4f7b77b-4gjk2\" (UID: \"7e0f47de-d77c-4840-a382-ed6a911f20b1\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-4gjk2" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.753277 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jpht\" (UniqueName: \"kubernetes.io/projected/68306a2f-6d51-4de9-9645-d3f0e3bc6e61-kube-api-access-9jpht\") pod \"glance-operator-controller-manager-84958c4d49-md78z\" (UID: \"68306a2f-6d51-4de9-9645-d3f0e3bc6e61\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-md78z" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.753298 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk2bd\" (UniqueName: \"kubernetes.io/projected/4dd0da8e-118c-4125-987a-a1122b592002-kube-api-access-hk2bd\") pod \"keystone-operator-controller-manager-5bd55b4bff-jvrxr\" (UID: \"4dd0da8e-118c-4125-987a-a1122b592002\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jvrxr" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.753328 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctwj6\" (UniqueName: \"kubernetes.io/projected/c815738f-cb2a-4136-9320-f75ceea923c5-kube-api-access-ctwj6\") pod \"heat-operator-controller-manager-5d889d78cf-5nl49\" (UID: \"c815738f-cb2a-4136-9320-f75ceea923c5\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.753375 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/03a52fee-7a10-42b3-8316-c16a0dcc01aa-cert\") pod \"infra-operator-controller-manager-9d6c5db85-jt9hq\" (UID: \"03a52fee-7a10-42b3-8316-c16a0dcc01aa\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.753394 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz7gc\" (UniqueName: \"kubernetes.io/projected/03d2a01f-1564-46f8-9648-e4a826470b44-kube-api-access-sz7gc\") pod \"horizon-operator-controller-manager-9f4696d94-2mxjr\" (UID: \"03d2a01f-1564-46f8-9648-e4a826470b44\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.753422 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l5pc\" (UniqueName: \"kubernetes.io/projected/f1ef22ed-d1cd-4bcc-b399-716552b889b8-kube-api-access-8l5pc\") pod \"ironic-operator-controller-manager-5cd4858477-f2khq\" (UID: \"f1ef22ed-d1cd-4bcc-b399-716552b889b8\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-f2khq" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.753444 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8dt7\" (UniqueName: \"kubernetes.io/projected/03a52fee-7a10-42b3-8316-c16a0dcc01aa-kube-api-access-w8dt7\") pod \"infra-operator-controller-manager-9d6c5db85-jt9hq\" (UID: \"03a52fee-7a10-42b3-8316-c16a0dcc01aa\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.766510 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-mx8pj" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.795611 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29hk7\" (UniqueName: \"kubernetes.io/projected/7e0f47de-d77c-4840-a382-ed6a911f20b1-kube-api-access-29hk7\") pod \"designate-operator-controller-manager-84f4f7b77b-4gjk2\" (UID: \"7e0f47de-d77c-4840-a382-ed6a911f20b1\") " pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-4gjk2" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.856556 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz7gc\" (UniqueName: \"kubernetes.io/projected/03d2a01f-1564-46f8-9648-e4a826470b44-kube-api-access-sz7gc\") pod \"horizon-operator-controller-manager-9f4696d94-2mxjr\" (UID: \"03d2a01f-1564-46f8-9648-e4a826470b44\") " pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.858612 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jpht\" (UniqueName: \"kubernetes.io/projected/68306a2f-6d51-4de9-9645-d3f0e3bc6e61-kube-api-access-9jpht\") pod \"glance-operator-controller-manager-84958c4d49-md78z\" (UID: \"68306a2f-6d51-4de9-9645-d3f0e3bc6e61\") " pod="openstack-operators/glance-operator-controller-manager-84958c4d49-md78z" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.859044 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctwj6\" (UniqueName: \"kubernetes.io/projected/c815738f-cb2a-4136-9320-f75ceea923c5-kube-api-access-ctwj6\") pod \"heat-operator-controller-manager-5d889d78cf-5nl49\" (UID: \"c815738f-cb2a-4136-9320-f75ceea923c5\") " pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.859180 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-4gjk2" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.860760 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk2bd\" (UniqueName: \"kubernetes.io/projected/4dd0da8e-118c-4125-987a-a1122b592002-kube-api-access-hk2bd\") pod \"keystone-operator-controller-manager-5bd55b4bff-jvrxr\" (UID: \"4dd0da8e-118c-4125-987a-a1122b592002\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jvrxr" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.860799 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrnbq\" (UniqueName: \"kubernetes.io/projected/c066e712-5c8d-48f8-b2fe-daa26e9f53e6-kube-api-access-lrnbq\") pod \"manila-operator-controller-manager-6d68dbc695-kvhkb\" (UID: \"c066e712-5c8d-48f8-b2fe-daa26e9f53e6\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.860852 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/03a52fee-7a10-42b3-8316-c16a0dcc01aa-cert\") pod \"infra-operator-controller-manager-9d6c5db85-jt9hq\" (UID: \"03a52fee-7a10-42b3-8316-c16a0dcc01aa\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.860890 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l5pc\" (UniqueName: \"kubernetes.io/projected/f1ef22ed-d1cd-4bcc-b399-716552b889b8-kube-api-access-8l5pc\") pod \"ironic-operator-controller-manager-5cd4858477-f2khq\" (UID: \"f1ef22ed-d1cd-4bcc-b399-716552b889b8\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-f2khq" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.860912 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8dt7\" (UniqueName: \"kubernetes.io/projected/03a52fee-7a10-42b3-8316-c16a0dcc01aa-kube-api-access-w8dt7\") pod \"infra-operator-controller-manager-9d6c5db85-jt9hq\" (UID: \"03a52fee-7a10-42b3-8316-c16a0dcc01aa\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" Oct 02 01:57:11 crc kubenswrapper[4775]: E1002 01:57:11.861226 4775 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 02 01:57:11 crc kubenswrapper[4775]: E1002 01:57:11.861264 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/03a52fee-7a10-42b3-8316-c16a0dcc01aa-cert podName:03a52fee-7a10-42b3-8316-c16a0dcc01aa nodeName:}" failed. No retries permitted until 2025-10-02 01:57:12.361249191 +0000 UTC m=+969.527993231 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/03a52fee-7a10-42b3-8316-c16a0dcc01aa-cert") pod "infra-operator-controller-manager-9d6c5db85-jt9hq" (UID: "03a52fee-7a10-42b3-8316-c16a0dcc01aa") : secret "infra-operator-webhook-server-cert" not found Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.883919 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.886046 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-2zft5"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.886822 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.888760 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.889070 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2zft5" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.896748 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-bg6mg" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.896841 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-txbtz" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.898560 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.899494 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.901511 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.907003 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-2zft5"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.918585 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk2bd\" (UniqueName: \"kubernetes.io/projected/4dd0da8e-118c-4125-987a-a1122b592002-kube-api-access-hk2bd\") pod \"keystone-operator-controller-manager-5bd55b4bff-jvrxr\" (UID: \"4dd0da8e-118c-4125-987a-a1122b592002\") " pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jvrxr" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.918725 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-lxx84" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.930537 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8dt7\" (UniqueName: \"kubernetes.io/projected/03a52fee-7a10-42b3-8316-c16a0dcc01aa-kube-api-access-w8dt7\") pod \"infra-operator-controller-manager-9d6c5db85-jt9hq\" (UID: \"03a52fee-7a10-42b3-8316-c16a0dcc01aa\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.942437 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.948876 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l5pc\" (UniqueName: \"kubernetes.io/projected/f1ef22ed-d1cd-4bcc-b399-716552b889b8-kube-api-access-8l5pc\") pod \"ironic-operator-controller-manager-5cd4858477-f2khq\" (UID: \"f1ef22ed-d1cd-4bcc-b399-716552b889b8\") " pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-f2khq" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.954233 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65"] Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.962448 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrnbq\" (UniqueName: \"kubernetes.io/projected/c066e712-5c8d-48f8-b2fe-daa26e9f53e6-kube-api-access-lrnbq\") pod \"manila-operator-controller-manager-6d68dbc695-kvhkb\" (UID: \"c066e712-5c8d-48f8-b2fe-daa26e9f53e6\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.962512 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzqr7\" (UniqueName: \"kubernetes.io/projected/600073d0-68ea-45f5-87c2-0775377796b5-kube-api-access-bzqr7\") pod \"neutron-operator-controller-manager-849d5b9b84-2zft5\" (UID: \"600073d0-68ea-45f5-87c2-0775377796b5\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2zft5" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.962545 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w4bq\" (UniqueName: \"kubernetes.io/projected/52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2-kube-api-access-5w4bq\") pod \"nova-operator-controller-manager-64cd67b5cb-5vt65\" (UID: \"52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65" Oct 02 01:57:11 crc kubenswrapper[4775]: I1002 01:57:11.967642 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jvrxr" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.041230 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrnbq\" (UniqueName: \"kubernetes.io/projected/c066e712-5c8d-48f8-b2fe-daa26e9f53e6-kube-api-access-lrnbq\") pod \"manila-operator-controller-manager-6d68dbc695-kvhkb\" (UID: \"c066e712-5c8d-48f8-b2fe-daa26e9f53e6\") " pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.063672 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.064534 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzqr7\" (UniqueName: \"kubernetes.io/projected/600073d0-68ea-45f5-87c2-0775377796b5-kube-api-access-bzqr7\") pod \"neutron-operator-controller-manager-849d5b9b84-2zft5\" (UID: \"600073d0-68ea-45f5-87c2-0775377796b5\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2zft5" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.064586 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w4bq\" (UniqueName: \"kubernetes.io/projected/52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2-kube-api-access-5w4bq\") pod \"nova-operator-controller-manager-64cd67b5cb-5vt65\" (UID: \"52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.064626 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2p5bs\" (UniqueName: \"kubernetes.io/projected/faa566e8-9b97-4766-a414-4ed0fccf9a81-kube-api-access-2p5bs\") pod \"mariadb-operator-controller-manager-88c7-qkvrn\" (UID: \"faa566e8-9b97-4766-a414-4ed0fccf9a81\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.065162 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.086454 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.090666 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-zwwpn" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.097068 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w4bq\" (UniqueName: \"kubernetes.io/projected/52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2-kube-api-access-5w4bq\") pod \"nova-operator-controller-manager-64cd67b5cb-5vt65\" (UID: \"52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2\") " pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.103816 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzqr7\" (UniqueName: \"kubernetes.io/projected/600073d0-68ea-45f5-87c2-0775377796b5-kube-api-access-bzqr7\") pod \"neutron-operator-controller-manager-849d5b9b84-2zft5\" (UID: \"600073d0-68ea-45f5-87c2-0775377796b5\") " pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2zft5" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.103919 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.105762 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-md78z" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.108252 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.111924 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-tshtc" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.112771 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.113700 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.115992 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.116127 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-5wn8p" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.119946 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-kv2nm"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.122820 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-kv2nm" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.133269 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-rh27g" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.139019 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.142934 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.148126 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.149140 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.153266 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-5ql46" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.153375 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-kv2nm"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.163389 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.165474 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9d4b689-6ed6-4784-a010-1cea2d49f469-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-9qscs\" (UID: \"c9d4b689-6ed6-4784-a010-1cea2d49f469\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.165542 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skvmp\" (UniqueName: \"kubernetes.io/projected/749bf5d6-b257-47e7-b632-e0edf0321adf-kube-api-access-skvmp\") pod \"octavia-operator-controller-manager-7b787867f4-lqrg2\" (UID: \"749bf5d6-b257-47e7-b632-e0edf0321adf\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.165570 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgqnn\" (UniqueName: \"kubernetes.io/projected/c9d4b689-6ed6-4784-a010-1cea2d49f469-kube-api-access-zgqnn\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-9qscs\" (UID: \"c9d4b689-6ed6-4784-a010-1cea2d49f469\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.165608 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2p5bs\" (UniqueName: \"kubernetes.io/projected/faa566e8-9b97-4766-a414-4ed0fccf9a81-kube-api-access-2p5bs\") pod \"mariadb-operator-controller-manager-88c7-qkvrn\" (UID: \"faa566e8-9b97-4766-a414-4ed0fccf9a81\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.191107 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.192290 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.197721 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-dgnnn" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.201572 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-bpg5g"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.202160 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2p5bs\" (UniqueName: \"kubernetes.io/projected/faa566e8-9b97-4766-a414-4ed0fccf9a81-kube-api-access-2p5bs\") pod \"mariadb-operator-controller-manager-88c7-qkvrn\" (UID: \"faa566e8-9b97-4766-a414-4ed0fccf9a81\") " pod="openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.202688 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-bpg5g" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.205576 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-vkrqs" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.210693 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.214343 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-bpg5g"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.218121 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.219100 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.222035 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-xx7zr" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.224080 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.245231 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-f2khq" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.250264 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.251297 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.256782 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-dlpbd" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.257061 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.261319 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.262828 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.266326 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnvvq\" (UniqueName: \"kubernetes.io/projected/c058610a-487c-4f23-b906-b03b0a458534-kube-api-access-mnvvq\") pod \"test-operator-controller-manager-85777745bb-bpg5g\" (UID: \"c058610a-487c-4f23-b906-b03b0a458534\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-bpg5g" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.266355 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vhr7\" (UniqueName: \"kubernetes.io/projected/6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f-kube-api-access-4vhr7\") pod \"ovn-operator-controller-manager-9976ff44c-xhpvs\" (UID: \"6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.266377 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skvmp\" (UniqueName: \"kubernetes.io/projected/749bf5d6-b257-47e7-b632-e0edf0321adf-kube-api-access-skvmp\") pod \"octavia-operator-controller-manager-7b787867f4-lqrg2\" (UID: \"749bf5d6-b257-47e7-b632-e0edf0321adf\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.266402 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgqnn\" (UniqueName: \"kubernetes.io/projected/c9d4b689-6ed6-4784-a010-1cea2d49f469-kube-api-access-zgqnn\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-9qscs\" (UID: \"c9d4b689-6ed6-4784-a010-1cea2d49f469\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.266427 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tqdx\" (UniqueName: \"kubernetes.io/projected/4ea65180-a9ee-476e-a260-be2d9f276f8c-kube-api-access-2tqdx\") pod \"swift-operator-controller-manager-84d6b4b759-ptq59\" (UID: \"4ea65180-a9ee-476e-a260-be2d9f276f8c\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.266454 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gctxm\" (UniqueName: \"kubernetes.io/projected/c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f-kube-api-access-gctxm\") pod \"telemetry-operator-controller-manager-b8d54b5d7-vbs7k\" (UID: \"c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.266475 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgmbh\" (UniqueName: \"kubernetes.io/projected/35c85a80-497b-47f8-afdd-8ad771e4557d-kube-api-access-pgmbh\") pod \"placement-operator-controller-manager-589c58c6c-kv2nm\" (UID: \"35c85a80-497b-47f8-afdd-8ad771e4557d\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-kv2nm" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.266529 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9d4b689-6ed6-4784-a010-1cea2d49f469-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-9qscs\" (UID: \"c9d4b689-6ed6-4784-a010-1cea2d49f469\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" Oct 02 01:57:12 crc kubenswrapper[4775]: E1002 01:57:12.266630 4775 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 01:57:12 crc kubenswrapper[4775]: E1002 01:57:12.266689 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9d4b689-6ed6-4784-a010-1cea2d49f469-cert podName:c9d4b689-6ed6-4784-a010-1cea2d49f469 nodeName:}" failed. No retries permitted until 2025-10-02 01:57:12.766670843 +0000 UTC m=+969.933414883 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c9d4b689-6ed6-4784-a010-1cea2d49f469-cert") pod "openstack-baremetal-operator-controller-manager-5869cb545-9qscs" (UID: "c9d4b689-6ed6-4784-a010-1cea2d49f469") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.311457 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgqnn\" (UniqueName: \"kubernetes.io/projected/c9d4b689-6ed6-4784-a010-1cea2d49f469-kube-api-access-zgqnn\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-9qscs\" (UID: \"c9d4b689-6ed6-4784-a010-1cea2d49f469\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.319788 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2zft5" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.326057 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.327365 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skvmp\" (UniqueName: \"kubernetes.io/projected/749bf5d6-b257-47e7-b632-e0edf0321adf-kube-api-access-skvmp\") pod \"octavia-operator-controller-manager-7b787867f4-lqrg2\" (UID: \"749bf5d6-b257-47e7-b632-e0edf0321adf\") " pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.328433 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.329216 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.331871 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-4j4wj" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.349692 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.360667 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.367610 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gctxm\" (UniqueName: \"kubernetes.io/projected/c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f-kube-api-access-gctxm\") pod \"telemetry-operator-controller-manager-b8d54b5d7-vbs7k\" (UID: \"c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.367653 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgmbh\" (UniqueName: \"kubernetes.io/projected/35c85a80-497b-47f8-afdd-8ad771e4557d-kube-api-access-pgmbh\") pod \"placement-operator-controller-manager-589c58c6c-kv2nm\" (UID: \"35c85a80-497b-47f8-afdd-8ad771e4557d\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-kv2nm" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.367721 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0c5961b-3c87-4438-b862-e2d1a83b1d6f-cert\") pod \"openstack-operator-controller-manager-557f5d867b-2m767\" (UID: \"f0c5961b-3c87-4438-b862-e2d1a83b1d6f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.367787 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnmvw\" (UniqueName: \"kubernetes.io/projected/f0c5961b-3c87-4438-b862-e2d1a83b1d6f-kube-api-access-dnmvw\") pod \"openstack-operator-controller-manager-557f5d867b-2m767\" (UID: \"f0c5961b-3c87-4438-b862-e2d1a83b1d6f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.367855 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/03a52fee-7a10-42b3-8316-c16a0dcc01aa-cert\") pod \"infra-operator-controller-manager-9d6c5db85-jt9hq\" (UID: \"03a52fee-7a10-42b3-8316-c16a0dcc01aa\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.367880 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vhr7\" (UniqueName: \"kubernetes.io/projected/6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f-kube-api-access-4vhr7\") pod \"ovn-operator-controller-manager-9976ff44c-xhpvs\" (UID: \"6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.367896 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnvvq\" (UniqueName: \"kubernetes.io/projected/c058610a-487c-4f23-b906-b03b0a458534-kube-api-access-mnvvq\") pod \"test-operator-controller-manager-85777745bb-bpg5g\" (UID: \"c058610a-487c-4f23-b906-b03b0a458534\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-bpg5g" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.367937 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdd8p\" (UniqueName: \"kubernetes.io/projected/7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62-kube-api-access-wdd8p\") pod \"watcher-operator-controller-manager-6b9957f54f-mlksw\" (UID: \"7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.368012 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tqdx\" (UniqueName: \"kubernetes.io/projected/4ea65180-a9ee-476e-a260-be2d9f276f8c-kube-api-access-2tqdx\") pod \"swift-operator-controller-manager-84d6b4b759-ptq59\" (UID: \"4ea65180-a9ee-476e-a260-be2d9f276f8c\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.406455 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/03a52fee-7a10-42b3-8316-c16a0dcc01aa-cert\") pod \"infra-operator-controller-manager-9d6c5db85-jt9hq\" (UID: \"03a52fee-7a10-42b3-8316-c16a0dcc01aa\") " pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.420077 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gctxm\" (UniqueName: \"kubernetes.io/projected/c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f-kube-api-access-gctxm\") pod \"telemetry-operator-controller-manager-b8d54b5d7-vbs7k\" (UID: \"c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f\") " pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.420123 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnvvq\" (UniqueName: \"kubernetes.io/projected/c058610a-487c-4f23-b906-b03b0a458534-kube-api-access-mnvvq\") pod \"test-operator-controller-manager-85777745bb-bpg5g\" (UID: \"c058610a-487c-4f23-b906-b03b0a458534\") " pod="openstack-operators/test-operator-controller-manager-85777745bb-bpg5g" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.420675 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgmbh\" (UniqueName: \"kubernetes.io/projected/35c85a80-497b-47f8-afdd-8ad771e4557d-kube-api-access-pgmbh\") pod \"placement-operator-controller-manager-589c58c6c-kv2nm\" (UID: \"35c85a80-497b-47f8-afdd-8ad771e4557d\") " pod="openstack-operators/placement-operator-controller-manager-589c58c6c-kv2nm" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.421072 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vhr7\" (UniqueName: \"kubernetes.io/projected/6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f-kube-api-access-4vhr7\") pod \"ovn-operator-controller-manager-9976ff44c-xhpvs\" (UID: \"6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f\") " pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.422567 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tqdx\" (UniqueName: \"kubernetes.io/projected/4ea65180-a9ee-476e-a260-be2d9f276f8c-kube-api-access-2tqdx\") pod \"swift-operator-controller-manager-84d6b4b759-ptq59\" (UID: \"4ea65180-a9ee-476e-a260-be2d9f276f8c\") " pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.445068 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-85777745bb-bpg5g" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.468662 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdd8p\" (UniqueName: \"kubernetes.io/projected/7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62-kube-api-access-wdd8p\") pod \"watcher-operator-controller-manager-6b9957f54f-mlksw\" (UID: \"7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.468933 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0c5961b-3c87-4438-b862-e2d1a83b1d6f-cert\") pod \"openstack-operator-controller-manager-557f5d867b-2m767\" (UID: \"f0c5961b-3c87-4438-b862-e2d1a83b1d6f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.469087 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqvrp\" (UniqueName: \"kubernetes.io/projected/4175c381-6ab1-4e55-b786-c7b886d69b35-kube-api-access-rqvrp\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6\" (UID: \"4175c381-6ab1-4e55-b786-c7b886d69b35\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.469183 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnmvw\" (UniqueName: \"kubernetes.io/projected/f0c5961b-3c87-4438-b862-e2d1a83b1d6f-kube-api-access-dnmvw\") pod \"openstack-operator-controller-manager-557f5d867b-2m767\" (UID: \"f0c5961b-3c87-4438-b862-e2d1a83b1d6f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" Oct 02 01:57:12 crc kubenswrapper[4775]: E1002 01:57:12.469771 4775 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 02 01:57:12 crc kubenswrapper[4775]: E1002 01:57:12.469848 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f0c5961b-3c87-4438-b862-e2d1a83b1d6f-cert podName:f0c5961b-3c87-4438-b862-e2d1a83b1d6f nodeName:}" failed. No retries permitted until 2025-10-02 01:57:12.969827841 +0000 UTC m=+970.136571881 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/f0c5961b-3c87-4438-b862-e2d1a83b1d6f-cert") pod "openstack-operator-controller-manager-557f5d867b-2m767" (UID: "f0c5961b-3c87-4438-b862-e2d1a83b1d6f") : secret "webhook-server-cert" not found Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.490853 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnmvw\" (UniqueName: \"kubernetes.io/projected/f0c5961b-3c87-4438-b862-e2d1a83b1d6f-kube-api-access-dnmvw\") pod \"openstack-operator-controller-manager-557f5d867b-2m767\" (UID: \"f0c5961b-3c87-4438-b862-e2d1a83b1d6f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.497626 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdd8p\" (UniqueName: \"kubernetes.io/projected/7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62-kube-api-access-wdd8p\") pod \"watcher-operator-controller-manager-6b9957f54f-mlksw\" (UID: \"7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62\") " pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.515012 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.578228 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.579588 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqvrp\" (UniqueName: \"kubernetes.io/projected/4175c381-6ab1-4e55-b786-c7b886d69b35-kube-api-access-rqvrp\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6\" (UID: \"4175c381-6ab1-4e55-b786-c7b886d69b35\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.604053 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqvrp\" (UniqueName: \"kubernetes.io/projected/4175c381-6ab1-4e55-b786-c7b886d69b35-kube-api-access-rqvrp\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6\" (UID: \"4175c381-6ab1-4e55-b786-c7b886d69b35\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.614695 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.615171 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-6ff8b75857-5wrms"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.618733 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-644bddb6d8-mx8pj"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.621923 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-kv2nm" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.636383 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.653684 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.772090 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.786875 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9d4b689-6ed6-4784-a010-1cea2d49f469-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-9qscs\" (UID: \"c9d4b689-6ed6-4784-a010-1cea2d49f469\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.793178 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9d4b689-6ed6-4784-a010-1cea2d49f469-cert\") pod \"openstack-baremetal-operator-controller-manager-5869cb545-9qscs\" (UID: \"c9d4b689-6ed6-4784-a010-1cea2d49f469\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.815855 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.885248 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.897380 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr"] Oct 02 01:57:12 crc kubenswrapper[4775]: I1002 01:57:12.899447 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84958c4d49-md78z"] Oct 02 01:57:12 crc kubenswrapper[4775]: W1002 01:57:12.923029 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68306a2f_6d51_4de9_9645_d3f0e3bc6e61.slice/crio-a6b942ee17b6810dc74be6abed75b32622db2b6f596d17c1461ecd13ea6be4ef WatchSource:0}: Error finding container a6b942ee17b6810dc74be6abed75b32622db2b6f596d17c1461ecd13ea6be4ef: Status 404 returned error can't find the container with id a6b942ee17b6810dc74be6abed75b32622db2b6f596d17c1461ecd13ea6be4ef Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.002148 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0c5961b-3c87-4438-b862-e2d1a83b1d6f-cert\") pod \"openstack-operator-controller-manager-557f5d867b-2m767\" (UID: \"f0c5961b-3c87-4438-b862-e2d1a83b1d6f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.010535 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f0c5961b-3c87-4438-b862-e2d1a83b1d6f-cert\") pod \"openstack-operator-controller-manager-557f5d867b-2m767\" (UID: \"f0c5961b-3c87-4438-b862-e2d1a83b1d6f\") " pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.097433 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.148925 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65"] Oct 02 01:57:13 crc kubenswrapper[4775]: W1002 01:57:13.153358 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc815738f_cb2a_4136_9320_f75ceea923c5.slice/crio-54cf709507f79658d87258245f83c896869efc4d0f698de32a474e9a52bdee67 WatchSource:0}: Error finding container 54cf709507f79658d87258245f83c896869efc4d0f698de32a474e9a52bdee67: Status 404 returned error can't find the container with id 54cf709507f79658d87258245f83c896869efc4d0f698de32a474e9a52bdee67 Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.157303 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49"] Oct 02 01:57:13 crc kubenswrapper[4775]: W1002 01:57:13.162331 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52bcb0eb_a3c9_42a0_9c7e_5bb2ef5fabb2.slice/crio-da9e6f97dafa68d7893e9cc6c9a7a17e4fc725b56bd573d0dab14dfefb7cfe30 WatchSource:0}: Error finding container da9e6f97dafa68d7893e9cc6c9a7a17e4fc725b56bd573d0dab14dfefb7cfe30: Status 404 returned error can't find the container with id da9e6f97dafa68d7893e9cc6c9a7a17e4fc725b56bd573d0dab14dfefb7cfe30 Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.180447 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-84f4f7b77b-4gjk2"] Oct 02 01:57:13 crc kubenswrapper[4775]: W1002 01:57:13.222831 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e0f47de_d77c_4840_a382_ed6a911f20b1.slice/crio-04f137c457e246ef4fc350beee446002aeaa7097ab5a835ddc26b908cff52185 WatchSource:0}: Error finding container 04f137c457e246ef4fc350beee446002aeaa7097ab5a835ddc26b908cff52185: Status 404 returned error can't find the container with id 04f137c457e246ef4fc350beee446002aeaa7097ab5a835ddc26b908cff52185 Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.333447 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5cd4858477-f2khq"] Oct 02 01:57:13 crc kubenswrapper[4775]: W1002 01:57:13.342118 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1ef22ed_d1cd_4bcc_b399_716552b889b8.slice/crio-cbe014d4fc1a555bca6d1d80fbeb19b5b1a953c056605c2fc3609ba216587585 WatchSource:0}: Error finding container cbe014d4fc1a555bca6d1d80fbeb19b5b1a953c056605c2fc3609ba216587585: Status 404 returned error can't find the container with id cbe014d4fc1a555bca6d1d80fbeb19b5b1a953c056605c2fc3609ba216587585 Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.348892 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jvrxr"] Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.353876 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb"] Oct 02 01:57:13 crc kubenswrapper[4775]: W1002 01:57:13.358903 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod600073d0_68ea_45f5_87c2_0775377796b5.slice/crio-d64bcec01548bf1c72ffc1d06a744974eee22e383625d7dc25d7c67bb79c5347 WatchSource:0}: Error finding container d64bcec01548bf1c72ffc1d06a744974eee22e383625d7dc25d7c67bb79c5347: Status 404 returned error can't find the container with id d64bcec01548bf1c72ffc1d06a744974eee22e383625d7dc25d7c67bb79c5347 Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.359777 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq"] Oct 02 01:57:13 crc kubenswrapper[4775]: W1002 01:57:13.361917 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4dd0da8e_118c_4125_987a_a1122b592002.slice/crio-c62db0d81f83308b131ab4c10b86fb6dbff4ee18c2b47377977cf0a0d422459e WatchSource:0}: Error finding container c62db0d81f83308b131ab4c10b86fb6dbff4ee18c2b47377977cf0a0d422459e: Status 404 returned error can't find the container with id c62db0d81f83308b131ab4c10b86fb6dbff4ee18c2b47377977cf0a0d422459e Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.363591 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-849d5b9b84-2zft5"] Oct 02 01:57:13 crc kubenswrapper[4775]: W1002 01:57:13.363620 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc066e712_5c8d_48f8_b2fe_daa26e9f53e6.slice/crio-f99be7f6a2db55894868926c99ba587e14ea07e5ee71fbc32401b476e9511a3f WatchSource:0}: Error finding container f99be7f6a2db55894868926c99ba587e14ea07e5ee71fbc32401b476e9511a3f: Status 404 returned error can't find the container with id f99be7f6a2db55894868926c99ba587e14ea07e5ee71fbc32401b476e9511a3f Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.675700 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jvrxr" event={"ID":"4dd0da8e-118c-4125-987a-a1122b592002","Type":"ContainerStarted","Data":"c62db0d81f83308b131ab4c10b86fb6dbff4ee18c2b47377977cf0a0d422459e"} Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.677941 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr" event={"ID":"03d2a01f-1564-46f8-9648-e4a826470b44","Type":"ContainerStarted","Data":"cd659e16be598d776f4bbeef41a81eaf4dc0644424139fb68858f0d52a95839b"} Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.679069 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-5wrms" event={"ID":"c2757728-9f90-4b12-8a40-3a7845c4f461","Type":"ContainerStarted","Data":"26e7b95659f706f3d0ed49f5e0984e4b354aeb5290403574f261238bd45cafd7"} Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.680188 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-4gjk2" event={"ID":"7e0f47de-d77c-4840-a382-ed6a911f20b1","Type":"ContainerStarted","Data":"04f137c457e246ef4fc350beee446002aeaa7097ab5a835ddc26b908cff52185"} Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.684148 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65" event={"ID":"52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2","Type":"ContainerStarted","Data":"da9e6f97dafa68d7893e9cc6c9a7a17e4fc725b56bd573d0dab14dfefb7cfe30"} Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.685259 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-mx8pj" event={"ID":"63b4a9a2-62d3-48a8-a2ca-74496dcff908","Type":"ContainerStarted","Data":"5cdf1f8f48bc1af18149daaff45c63ac33f057afa8dd99e51459189849dc41cf"} Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.686655 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" event={"ID":"03a52fee-7a10-42b3-8316-c16a0dcc01aa","Type":"ContainerStarted","Data":"875b8098980e6fa232177217708e9612a33490e425ed11b76c7b302d3493c269"} Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.687796 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49" event={"ID":"c815738f-cb2a-4136-9320-f75ceea923c5","Type":"ContainerStarted","Data":"54cf709507f79658d87258245f83c896869efc4d0f698de32a474e9a52bdee67"} Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.688765 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb" event={"ID":"c066e712-5c8d-48f8-b2fe-daa26e9f53e6","Type":"ContainerStarted","Data":"f99be7f6a2db55894868926c99ba587e14ea07e5ee71fbc32401b476e9511a3f"} Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.689765 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2zft5" event={"ID":"600073d0-68ea-45f5-87c2-0775377796b5","Type":"ContainerStarted","Data":"d64bcec01548bf1c72ffc1d06a744974eee22e383625d7dc25d7c67bb79c5347"} Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.689800 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767"] Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.691012 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-md78z" event={"ID":"68306a2f-6d51-4de9-9645-d3f0e3bc6e61","Type":"ContainerStarted","Data":"a6b942ee17b6810dc74be6abed75b32622db2b6f596d17c1461ecd13ea6be4ef"} Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.691989 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-f2khq" event={"ID":"f1ef22ed-d1cd-4bcc-b399-716552b889b8","Type":"ContainerStarted","Data":"cbe014d4fc1a555bca6d1d80fbeb19b5b1a953c056605c2fc3609ba216587585"} Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.755744 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn"] Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.798533 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw"] Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.798577 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59"] Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.799471 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k"] Oct 02 01:57:13 crc kubenswrapper[4775]: W1002 01:57:13.803543 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35c85a80_497b_47f8_afdd_8ad771e4557d.slice/crio-f02a22f664a311df4abd102d06d8cde9cf1114e688e407e10f26b8e529faacd7 WatchSource:0}: Error finding container f02a22f664a311df4abd102d06d8cde9cf1114e688e407e10f26b8e529faacd7: Status 404 returned error can't find the container with id f02a22f664a311df4abd102d06d8cde9cf1114e688e407e10f26b8e529faacd7 Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.804659 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-589c58c6c-kv2nm"] Oct 02 01:57:13 crc kubenswrapper[4775]: W1002 01:57:13.810728 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fe9e47b_ad8f_4ebf_ae98_6adaba8cae62.slice/crio-ce85537522a9ea225450c4f591d4edd1955debdfe79b0fce881cc576eb263c8a WatchSource:0}: Error finding container ce85537522a9ea225450c4f591d4edd1955debdfe79b0fce881cc576eb263c8a: Status 404 returned error can't find the container with id ce85537522a9ea225450c4f591d4edd1955debdfe79b0fce881cc576eb263c8a Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.814333 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6"] Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.821340 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs"] Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.826420 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-85777745bb-bpg5g"] Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.843241 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2"] Oct 02 01:57:13 crc kubenswrapper[4775]: I1002 01:57:13.860520 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs"] Oct 02 01:57:13 crc kubenswrapper[4775]: E1002 01:57:13.877336 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rqvrp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6_openstack-operators(4175c381-6ab1-4e55-b786-c7b886d69b35): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 01:57:13 crc kubenswrapper[4775]: E1002 01:57:13.877657 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zgqnn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-5869cb545-9qscs_openstack-operators(c9d4b689-6ed6-4784-a010-1cea2d49f469): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 01:57:13 crc kubenswrapper[4775]: E1002 01:57:13.878716 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6" podUID="4175c381-6ab1-4e55-b786-c7b886d69b35" Oct 02 01:57:13 crc kubenswrapper[4775]: E1002 01:57:13.879458 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:15d7b5a365350a831ca59d984df67fadeccf89d599e487a7597b105afb82ce4a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2p5bs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-88c7-qkvrn_openstack-operators(faa566e8-9b97-4766-a414-4ed0fccf9a81): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 01:57:13 crc kubenswrapper[4775]: W1002 01:57:13.903648 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ac8ba4b_5cd5_41b8_92e1_fabd2543a33f.slice/crio-021e542542967995fc1996ca4999e09e4b32fd899f27d9098d9e30ae7f76ff91 WatchSource:0}: Error finding container 021e542542967995fc1996ca4999e09e4b32fd899f27d9098d9e30ae7f76ff91: Status 404 returned error can't find the container with id 021e542542967995fc1996ca4999e09e4b32fd899f27d9098d9e30ae7f76ff91 Oct 02 01:57:13 crc kubenswrapper[4775]: E1002 01:57:13.904903 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mnvvq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-85777745bb-bpg5g_openstack-operators(c058610a-487c-4f23-b906-b03b0a458534): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 01:57:13 crc kubenswrapper[4775]: E1002 01:57:13.907631 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2tqdx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-84d6b4b759-ptq59_openstack-operators(4ea65180-a9ee-476e-a260-be2d9f276f8c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 01:57:13 crc kubenswrapper[4775]: E1002 01:57:13.927228 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4vhr7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-9976ff44c-xhpvs_openstack-operators(6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 02 01:57:14 crc kubenswrapper[4775]: E1002 01:57:14.243859 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" podUID="c9d4b689-6ed6-4784-a010-1cea2d49f469" Oct 02 01:57:14 crc kubenswrapper[4775]: E1002 01:57:14.263773 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn" podUID="faa566e8-9b97-4766-a414-4ed0fccf9a81" Oct 02 01:57:14 crc kubenswrapper[4775]: E1002 01:57:14.264044 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs" podUID="6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f" Oct 02 01:57:14 crc kubenswrapper[4775]: E1002 01:57:14.461337 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-bpg5g" podUID="c058610a-487c-4f23-b906-b03b0a458534" Oct 02 01:57:14 crc kubenswrapper[4775]: E1002 01:57:14.541515 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59" podUID="4ea65180-a9ee-476e-a260-be2d9f276f8c" Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.729156 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs" event={"ID":"6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f","Type":"ContainerStarted","Data":"5f35a50c66c1e2f30fc9ff4be3983f544504f5381d22b4734f2f8e94d9eddf4e"} Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.729231 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs" event={"ID":"6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f","Type":"ContainerStarted","Data":"021e542542967995fc1996ca4999e09e4b32fd899f27d9098d9e30ae7f76ff91"} Oct 02 01:57:14 crc kubenswrapper[4775]: E1002 01:57:14.735156 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs" podUID="6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f" Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.750190 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59" event={"ID":"4ea65180-a9ee-476e-a260-be2d9f276f8c","Type":"ContainerStarted","Data":"3f8910bc8af38d89648fce6c218127646a8861fd3b365f4d96e9002173de430d"} Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.750232 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59" event={"ID":"4ea65180-a9ee-476e-a260-be2d9f276f8c","Type":"ContainerStarted","Data":"9386833eb491f362b94228da26217de60d277f630835418979083ed1830b69ff"} Oct 02 01:57:14 crc kubenswrapper[4775]: E1002 01:57:14.751457 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59" podUID="4ea65180-a9ee-476e-a260-be2d9f276f8c" Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.752123 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-kv2nm" event={"ID":"35c85a80-497b-47f8-afdd-8ad771e4557d","Type":"ContainerStarted","Data":"f02a22f664a311df4abd102d06d8cde9cf1114e688e407e10f26b8e529faacd7"} Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.766184 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn" event={"ID":"faa566e8-9b97-4766-a414-4ed0fccf9a81","Type":"ContainerStarted","Data":"18bffe0833fa9931e2510f8fd0ea9e484dfd200d7e99ba30e5902c1fba87c1ca"} Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.766233 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn" event={"ID":"faa566e8-9b97-4766-a414-4ed0fccf9a81","Type":"ContainerStarted","Data":"d68f6589f540bcd81dda86c4b28db138d626d9af0498ee93c650ecddc59d71b5"} Oct 02 01:57:14 crc kubenswrapper[4775]: E1002 01:57:14.767550 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:15d7b5a365350a831ca59d984df67fadeccf89d599e487a7597b105afb82ce4a\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn" podUID="faa566e8-9b97-4766-a414-4ed0fccf9a81" Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.791940 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-bpg5g" event={"ID":"c058610a-487c-4f23-b906-b03b0a458534","Type":"ContainerStarted","Data":"a44940859d7dd81ef12fefb86810a851c490d086b0e0b40399d7feb85bec70bf"} Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.792002 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-bpg5g" event={"ID":"c058610a-487c-4f23-b906-b03b0a458534","Type":"ContainerStarted","Data":"600098da55f0639be50560f7d60dac0330ad5dbf4abb65dbe691a84ec6157c32"} Oct 02 01:57:14 crc kubenswrapper[4775]: E1002 01:57:14.799196 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-bpg5g" podUID="c058610a-487c-4f23-b906-b03b0a458534" Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.832016 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" event={"ID":"c9d4b689-6ed6-4784-a010-1cea2d49f469","Type":"ContainerStarted","Data":"cb89692956076294aea3a91068dfc669a0d1bd8421d4957ce28a6e9776a52bcb"} Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.832074 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" event={"ID":"c9d4b689-6ed6-4784-a010-1cea2d49f469","Type":"ContainerStarted","Data":"8d4c4bee65edb4ae0db0e60bf8d87ca92441f0c435164e60e472df70d4eaec61"} Oct 02 01:57:14 crc kubenswrapper[4775]: E1002 01:57:14.835134 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" podUID="c9d4b689-6ed6-4784-a010-1cea2d49f469" Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.854393 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k" event={"ID":"c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f","Type":"ContainerStarted","Data":"b42a3a8100836ce6ec6ba2f47aa3d571ac739986dda9d187200653909df6861e"} Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.872348 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" event={"ID":"f0c5961b-3c87-4438-b862-e2d1a83b1d6f","Type":"ContainerStarted","Data":"c9819e4e72fe2d010f9ff5bc5971c4e62539d7ede18f161a56aa26eb02720d4e"} Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.872388 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" event={"ID":"f0c5961b-3c87-4438-b862-e2d1a83b1d6f","Type":"ContainerStarted","Data":"e302e6b9b509768586bf13d8e7a6c1753d403b99ee437235b1c02ca0395ab3a9"} Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.872413 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" event={"ID":"f0c5961b-3c87-4438-b862-e2d1a83b1d6f","Type":"ContainerStarted","Data":"aac4db6dfcfac360db2c9a30cf92e03f900622e38706a438e2d551d69bc0d342"} Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.873177 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.874421 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6" event={"ID":"4175c381-6ab1-4e55-b786-c7b886d69b35","Type":"ContainerStarted","Data":"bb469e63ac55af0516bac70cfcce779b05f28d84fdd79cedacec796ae19b4fe5"} Oct 02 01:57:14 crc kubenswrapper[4775]: E1002 01:57:14.875619 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6" podUID="4175c381-6ab1-4e55-b786-c7b886d69b35" Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.876534 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw" event={"ID":"7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62","Type":"ContainerStarted","Data":"ce85537522a9ea225450c4f591d4edd1955debdfe79b0fce881cc576eb263c8a"} Oct 02 01:57:14 crc kubenswrapper[4775]: I1002 01:57:14.894267 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2" event={"ID":"749bf5d6-b257-47e7-b632-e0edf0321adf","Type":"ContainerStarted","Data":"ceb67a1d32d89c3528ed9a1a7bfa3d48e5bc4fdd87e4e63158c22d294fba2f4e"} Oct 02 01:57:15 crc kubenswrapper[4775]: I1002 01:57:15.030840 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" podStartSLOduration=3.030821244 podStartE2EDuration="3.030821244s" podCreationTimestamp="2025-10-02 01:57:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:57:15.027441379 +0000 UTC m=+972.194185419" watchObservedRunningTime="2025-10-02 01:57:15.030821244 +0000 UTC m=+972.197565284" Oct 02 01:57:15 crc kubenswrapper[4775]: E1002 01:57:15.922456 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:f61fdfbfd12027ce6b4e7ad553ec0582f080de0cfb472de6dc04ad3078bb17e3\\\"\"" pod="openstack-operators/test-operator-controller-manager-85777745bb-bpg5g" podUID="c058610a-487c-4f23-b906-b03b0a458534" Oct 02 01:57:15 crc kubenswrapper[4775]: E1002 01:57:15.922655 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:bca053da8adc37a9a246b478949960ac7abef8fcc0c58a2a45045c59a62b5fe4\\\"\"" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59" podUID="4ea65180-a9ee-476e-a260-be2d9f276f8c" Oct 02 01:57:15 crc kubenswrapper[4775]: E1002 01:57:15.922702 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6" podUID="4175c381-6ab1-4e55-b786-c7b886d69b35" Oct 02 01:57:15 crc kubenswrapper[4775]: E1002 01:57:15.924005 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:15d7b5a365350a831ca59d984df67fadeccf89d599e487a7597b105afb82ce4a\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn" podUID="faa566e8-9b97-4766-a414-4ed0fccf9a81" Oct 02 01:57:15 crc kubenswrapper[4775]: E1002 01:57:15.924271 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:1051afc168038fb814f75e7a5f07c588b295a83ebd143dcd8b46d799e31ad302\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs" podUID="6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f" Oct 02 01:57:15 crc kubenswrapper[4775]: E1002 01:57:15.924527 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:e7cfed051c1cf801e651fd4035070e38698039f284ac0b2a0332769fdbb4a9c8\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" podUID="c9d4b689-6ed6-4784-a010-1cea2d49f469" Oct 02 01:57:23 crc kubenswrapper[4775]: I1002 01:57:23.107469 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-557f5d867b-2m767" Oct 02 01:57:28 crc kubenswrapper[4775]: E1002 01:57:28.482689 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:917e6dcc519277c46e42898bc9f0f066790fa7b9633fcde668cc8a68a547c13c" Oct 02 01:57:28 crc kubenswrapper[4775]: E1002 01:57:28.483415 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:917e6dcc519277c46e42898bc9f0f066790fa7b9633fcde668cc8a68a547c13c,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ctwj6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5d889d78cf-5nl49_openstack-operators(c815738f-cb2a-4136-9320-f75ceea923c5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 01:57:28 crc kubenswrapper[4775]: E1002 01:57:28.871418 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9" Oct 02 01:57:28 crc kubenswrapper[4775]: E1002 01:57:28.871642 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-skvmp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-7b787867f4-lqrg2_openstack-operators(749bf5d6-b257-47e7-b632-e0edf0321adf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 01:57:29 crc kubenswrapper[4775]: E1002 01:57:29.300769 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884" Oct 02 01:57:29 crc kubenswrapper[4775]: E1002 01:57:29.301009 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lrnbq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-6d68dbc695-kvhkb_openstack-operators(c066e712-5c8d-48f8-b2fe-daa26e9f53e6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 01:57:29 crc kubenswrapper[4775]: E1002 01:57:29.689271 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:f5f0d2eb534f763cf6578af513add1c21c1659b2cd75214dfddfedb9eebf6397" Oct 02 01:57:29 crc kubenswrapper[4775]: E1002 01:57:29.689639 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:f5f0d2eb534f763cf6578af513add1c21c1659b2cd75214dfddfedb9eebf6397,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sz7gc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-9f4696d94-2mxjr_openstack-operators(03d2a01f-1564-46f8-9648-e4a826470b44): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 01:57:30 crc kubenswrapper[4775]: E1002 01:57:30.195586 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f" Oct 02 01:57:30 crc kubenswrapper[4775]: E1002 01:57:30.195802 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gctxm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-b8d54b5d7-vbs7k_openstack-operators(c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 01:57:30 crc kubenswrapper[4775]: E1002 01:57:30.678054 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06" Oct 02 01:57:30 crc kubenswrapper[4775]: E1002 01:57:30.678196 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wdd8p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6b9957f54f-mlksw_openstack-operators(7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 01:57:31 crc kubenswrapper[4775]: E1002 01:57:31.098396 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f" Oct 02 01:57:31 crc kubenswrapper[4775]: E1002 01:57:31.098599 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5w4bq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-64cd67b5cb-5vt65_openstack-operators(52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 01:57:34 crc kubenswrapper[4775]: E1002 01:57:34.169019 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb" podUID="c066e712-5c8d-48f8-b2fe-daa26e9f53e6" Oct 02 01:57:34 crc kubenswrapper[4775]: E1002 01:57:34.333856 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2" podUID="749bf5d6-b257-47e7-b632-e0edf0321adf" Oct 02 01:57:34 crc kubenswrapper[4775]: E1002 01:57:34.363570 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49" podUID="c815738f-cb2a-4136-9320-f75ceea923c5" Oct 02 01:57:34 crc kubenswrapper[4775]: E1002 01:57:34.370106 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k" podUID="c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f" Oct 02 01:57:34 crc kubenswrapper[4775]: E1002 01:57:34.423335 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr" podUID="03d2a01f-1564-46f8-9648-e4a826470b44" Oct 02 01:57:34 crc kubenswrapper[4775]: E1002 01:57:34.519320 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw" podUID="7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62" Oct 02 01:57:34 crc kubenswrapper[4775]: E1002 01:57:34.624643 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65" podUID="52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.085420 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65" event={"ID":"52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2","Type":"ContainerStarted","Data":"30bc0b7017a5da2651819aba9ac6a156c86c14b66cf762f2d655bdd495165638"} Oct 02 01:57:35 crc kubenswrapper[4775]: E1002 01:57:35.087047 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f\\\"\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65" podUID="52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.089038 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr" event={"ID":"03d2a01f-1564-46f8-9648-e4a826470b44","Type":"ContainerStarted","Data":"83c47e161c7497bf903941697db15c8c76dd74a0ec8b1e0363ed534a77377cff"} Oct 02 01:57:35 crc kubenswrapper[4775]: E1002 01:57:35.091167 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:f5f0d2eb534f763cf6578af513add1c21c1659b2cd75214dfddfedb9eebf6397\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr" podUID="03d2a01f-1564-46f8-9648-e4a826470b44" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.092270 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49" event={"ID":"c815738f-cb2a-4136-9320-f75ceea923c5","Type":"ContainerStarted","Data":"dd5c7bcf133f0cf0306d292bfd6e84591d9b2ea75df0158bcbbb0048c3648e12"} Oct 02 01:57:35 crc kubenswrapper[4775]: E1002 01:57:35.094160 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:917e6dcc519277c46e42898bc9f0f066790fa7b9633fcde668cc8a68a547c13c\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49" podUID="c815738f-cb2a-4136-9320-f75ceea923c5" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.094934 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-85777745bb-bpg5g" event={"ID":"c058610a-487c-4f23-b906-b03b0a458534","Type":"ContainerStarted","Data":"c8142b9264d0f545c4ed953b9c654c480200be384aa8a957d7be417951ed8840"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.095322 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-85777745bb-bpg5g" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.096309 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-md78z" event={"ID":"68306a2f-6d51-4de9-9645-d3f0e3bc6e61","Type":"ContainerStarted","Data":"8697de396d1dad422efd37304678defc3f4a10e9763ba3c83d696c4a793a7c38"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.098560 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k" event={"ID":"c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f","Type":"ContainerStarted","Data":"089b8fe17f55a07e29319aa3a8fc7b32a81e0d359a15efe343f3ccc71aab870b"} Oct 02 01:57:35 crc kubenswrapper[4775]: E1002 01:57:35.099445 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k" podUID="c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.101506 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59" event={"ID":"4ea65180-a9ee-476e-a260-be2d9f276f8c","Type":"ContainerStarted","Data":"d1e753fc19269a48af52ba9283225d9d551499de4ab1a2ae98965575a7b23804"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.101837 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.102709 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2zft5" event={"ID":"600073d0-68ea-45f5-87c2-0775377796b5","Type":"ContainerStarted","Data":"c2fcf13cd4672c78dd1051e18e8a93c0b224c0a5e76e60bf8e3636bbe027710a"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.103719 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw" event={"ID":"7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62","Type":"ContainerStarted","Data":"4398ef546727b9f84be822120dbe8a154aded6751526756b55c022ccf55a059d"} Oct 02 01:57:35 crc kubenswrapper[4775]: E1002 01:57:35.104373 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw" podUID="7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.105197 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-5wrms" event={"ID":"c2757728-9f90-4b12-8a40-3a7845c4f461","Type":"ContainerStarted","Data":"8faaa4765179cbf1b4d16bbb5cbf78ccce242e4f0883fff92784c6468e767b3e"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.109239 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-kv2nm" event={"ID":"35c85a80-497b-47f8-afdd-8ad771e4557d","Type":"ContainerStarted","Data":"07d1d2f4c8f0658b897d6510624276c87cd86fed7f25d3ca2c2e3e8941160479"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.111351 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" event={"ID":"c9d4b689-6ed6-4784-a010-1cea2d49f469","Type":"ContainerStarted","Data":"126eb38f48534975e33708a438f81d0b38898e46025241f223c807ef822cd029"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.111698 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.119460 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-4gjk2" event={"ID":"7e0f47de-d77c-4840-a382-ed6a911f20b1","Type":"ContainerStarted","Data":"8e3eb1f58ebf724fa38d2bf6262357def053cdd92b9e13d16b0880e4f8610740"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.121945 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-mx8pj" event={"ID":"63b4a9a2-62d3-48a8-a2ca-74496dcff908","Type":"ContainerStarted","Data":"e67306316b8368ea42aebac5a5ecae7dcc8df24dd9781272e1034f0c345efb80"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.131068 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb" event={"ID":"c066e712-5c8d-48f8-b2fe-daa26e9f53e6","Type":"ContainerStarted","Data":"ea4b2aa6097eb5fe65a95fe15484909287fe64e8db5d2ea67da60c48a9eab0eb"} Oct 02 01:57:35 crc kubenswrapper[4775]: E1002 01:57:35.137106 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb" podUID="c066e712-5c8d-48f8-b2fe-daa26e9f53e6" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.139368 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-f2khq" event={"ID":"f1ef22ed-d1cd-4bcc-b399-716552b889b8","Type":"ContainerStarted","Data":"0dd479c6511359627f342153b8fcc12373272bb453368e69eb02924637a73177"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.148668 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jvrxr" event={"ID":"4dd0da8e-118c-4125-987a-a1122b592002","Type":"ContainerStarted","Data":"1efc283df6c727b6c48a022db615b832dd736551cb8a8748b7ee8c7041549fee"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.149688 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2" event={"ID":"749bf5d6-b257-47e7-b632-e0edf0321adf","Type":"ContainerStarted","Data":"e4d1ac068d8be144c87f7070d7b38b797f7df04f13e21fbe61e79ce2142e9ac6"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.150475 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-85777745bb-bpg5g" podStartSLOduration=4.073148858 podStartE2EDuration="24.150460467s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.904807559 +0000 UTC m=+971.071551599" lastFinishedPulling="2025-10-02 01:57:33.982119158 +0000 UTC m=+991.148863208" observedRunningTime="2025-10-02 01:57:35.142774133 +0000 UTC m=+992.309518173" watchObservedRunningTime="2025-10-02 01:57:35.150460467 +0000 UTC m=+992.317204507" Oct 02 01:57:35 crc kubenswrapper[4775]: E1002 01:57:35.150906 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2" podUID="749bf5d6-b257-47e7-b632-e0edf0321adf" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.152662 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs" event={"ID":"6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f","Type":"ContainerStarted","Data":"980cc01d1d29dfae323f745b75d7181ae3ae21428c2249bf0c0e7f612f7b9a5a"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.153281 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.154742 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn" event={"ID":"faa566e8-9b97-4766-a414-4ed0fccf9a81","Type":"ContainerStarted","Data":"c77912d261f4d7010462377121d1a314881bd56eae11b7863421d11324421c7a"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.155003 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.159546 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" event={"ID":"03a52fee-7a10-42b3-8316-c16a0dcc01aa","Type":"ContainerStarted","Data":"0fd52949e851c9624f8e1353a35bb81dbd080b45c5b179b585ae64d063d8ffa7"} Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.245967 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59" podStartSLOduration=4.070596665 podStartE2EDuration="24.245932706s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.907469438 +0000 UTC m=+971.074213478" lastFinishedPulling="2025-10-02 01:57:34.082805479 +0000 UTC m=+991.249549519" observedRunningTime="2025-10-02 01:57:35.244989663 +0000 UTC m=+992.411733703" watchObservedRunningTime="2025-10-02 01:57:35.245932706 +0000 UTC m=+992.412676746" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.335599 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs" podStartSLOduration=4.225384819 podStartE2EDuration="24.33558268s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.927120375 +0000 UTC m=+971.093864415" lastFinishedPulling="2025-10-02 01:57:34.037318236 +0000 UTC m=+991.204062276" observedRunningTime="2025-10-02 01:57:35.334480082 +0000 UTC m=+992.501224122" watchObservedRunningTime="2025-10-02 01:57:35.33558268 +0000 UTC m=+992.502326720" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.398986 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn" podStartSLOduration=4.16475774 podStartE2EDuration="24.398969473s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.877449352 +0000 UTC m=+971.044193392" lastFinishedPulling="2025-10-02 01:57:34.111661065 +0000 UTC m=+991.278405125" observedRunningTime="2025-10-02 01:57:35.390614863 +0000 UTC m=+992.557358903" watchObservedRunningTime="2025-10-02 01:57:35.398969473 +0000 UTC m=+992.565713513" Oct 02 01:57:35 crc kubenswrapper[4775]: I1002 01:57:35.463749 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" podStartSLOduration=4.301535728 podStartE2EDuration="24.463734401s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.877025721 +0000 UTC m=+971.043769761" lastFinishedPulling="2025-10-02 01:57:34.039224394 +0000 UTC m=+991.205968434" observedRunningTime="2025-10-02 01:57:35.459642798 +0000 UTC m=+992.626386848" watchObservedRunningTime="2025-10-02 01:57:35.463734401 +0000 UTC m=+992.630478441" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.176895 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-5wrms" event={"ID":"c2757728-9f90-4b12-8a40-3a7845c4f461","Type":"ContainerStarted","Data":"0ed0878867ef776bdab93036a7819ce9a0bb2a1af3af1f53a14b89c05ceab099"} Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.177000 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-5wrms" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.180162 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-kv2nm" event={"ID":"35c85a80-497b-47f8-afdd-8ad771e4557d","Type":"ContainerStarted","Data":"301ca2edd3d02de1bf0ae09e70105abdac1794bebeae6080aaf4b6465b0d6e57"} Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.180382 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-kv2nm" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.182381 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-f2khq" event={"ID":"f1ef22ed-d1cd-4bcc-b399-716552b889b8","Type":"ContainerStarted","Data":"4825eacdfa93886072c1b0310f734e99d38a2f035c550a053c1f99ef26f91d2c"} Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.183345 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-f2khq" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.188512 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6" event={"ID":"4175c381-6ab1-4e55-b786-c7b886d69b35","Type":"ContainerStarted","Data":"5228b8963fb58aa736869112eac378ff2dca3b8696f48758e930556596db2ca8"} Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.190790 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2zft5" event={"ID":"600073d0-68ea-45f5-87c2-0775377796b5","Type":"ContainerStarted","Data":"c4a65ebc3141db18e1390486c0455878930c09865913ba484e3c363276c15bae"} Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.192835 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jvrxr" event={"ID":"4dd0da8e-118c-4125-987a-a1122b592002","Type":"ContainerStarted","Data":"8312b5ff83a5fa53e12aa8fa591e4b8e4ea0ab8fed95cdeeb0478da841ead6bf"} Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.192918 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jvrxr" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.194871 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-md78z" event={"ID":"68306a2f-6d51-4de9-9645-d3f0e3bc6e61","Type":"ContainerStarted","Data":"fa4b11a09b1fde0e2f27f6db996773aec9348cd97f1879ec9b5976a14ba01761"} Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.194995 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-md78z" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.197177 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" event={"ID":"03a52fee-7a10-42b3-8316-c16a0dcc01aa","Type":"ContainerStarted","Data":"57b490ca0081a63fadb92d439f895984bae6f1d14809f98181e8ff9f1204b6c3"} Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.197432 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.198790 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-4gjk2" event={"ID":"7e0f47de-d77c-4840-a382-ed6a911f20b1","Type":"ContainerStarted","Data":"7771030c5baaca41bcc884dcd84df23e24ad525aa077a01dc750f2f5c4610638"} Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.198926 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-4gjk2" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.201009 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-mx8pj" event={"ID":"63b4a9a2-62d3-48a8-a2ca-74496dcff908","Type":"ContainerStarted","Data":"fdb282c05e960d786f3968b51bf9ba71026490581fb376b21974961d671ac2e2"} Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.201035 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-mx8pj" Oct 02 01:57:36 crc kubenswrapper[4775]: E1002 01:57:36.203372 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:e1328760310f3bbf4548b8b1268cd711087dd91212b92bb0be287cad1f1b6fe9\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2" podUID="749bf5d6-b257-47e7-b632-e0edf0321adf" Oct 02 01:57:36 crc kubenswrapper[4775]: E1002 01:57:36.203434 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:4cdb30423c14ab48888aeeb699259bd9051284ec9f874ed9bab94c7965f45884\\\"\"" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb" podUID="c066e712-5c8d-48f8-b2fe-daa26e9f53e6" Oct 02 01:57:36 crc kubenswrapper[4775]: E1002 01:57:36.203459 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:09c2f519ea218f6038b7be039b8e6ac33ee93b217b9be0d2d18a5e7f94faae06\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw" podUID="7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62" Oct 02 01:57:36 crc kubenswrapper[4775]: E1002 01:57:36.204033 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:917e6dcc519277c46e42898bc9f0f066790fa7b9633fcde668cc8a68a547c13c\\\"\"" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49" podUID="c815738f-cb2a-4136-9320-f75ceea923c5" Oct 02 01:57:36 crc kubenswrapper[4775]: E1002 01:57:36.204076 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:f5f0d2eb534f763cf6578af513add1c21c1659b2cd75214dfddfedb9eebf6397\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr" podUID="03d2a01f-1564-46f8-9648-e4a826470b44" Oct 02 01:57:36 crc kubenswrapper[4775]: E1002 01:57:36.204248 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:a517abc6427ab73fed93b0bd89a6eb52d0311fbfb0c00752f889baf8ffd5068f\\\"\"" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65" podUID="52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2" Oct 02 01:57:36 crc kubenswrapper[4775]: E1002 01:57:36.205341 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:8fdf377daf05e2fa7346505017078fa81981dd945bf635a64c8022633c68118f\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k" podUID="c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.212791 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-5wrms" podStartSLOduration=5.37208917 podStartE2EDuration="25.212763909s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:12.672013574 +0000 UTC m=+969.838757654" lastFinishedPulling="2025-10-02 01:57:32.512688353 +0000 UTC m=+989.679432393" observedRunningTime="2025-10-02 01:57:36.199630369 +0000 UTC m=+993.366374439" watchObservedRunningTime="2025-10-02 01:57:36.212763909 +0000 UTC m=+993.379507979" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.223416 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-f2khq" podStartSLOduration=6.600423996 podStartE2EDuration="25.223400487s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.34779198 +0000 UTC m=+970.514536010" lastFinishedPulling="2025-10-02 01:57:31.970768471 +0000 UTC m=+989.137512501" observedRunningTime="2025-10-02 01:57:36.217020536 +0000 UTC m=+993.383764576" watchObservedRunningTime="2025-10-02 01:57:36.223400487 +0000 UTC m=+993.390144537" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.236088 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" podStartSLOduration=6.608448531 podStartE2EDuration="25.236074815s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.352754719 +0000 UTC m=+970.519498759" lastFinishedPulling="2025-10-02 01:57:31.980381003 +0000 UTC m=+989.147125043" observedRunningTime="2025-10-02 01:57:36.232962507 +0000 UTC m=+993.399706557" watchObservedRunningTime="2025-10-02 01:57:36.236074815 +0000 UTC m=+993.402818855" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.254658 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-md78z" podStartSLOduration=6.205074089 podStartE2EDuration="25.254639612s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:12.931266741 +0000 UTC m=+970.098010771" lastFinishedPulling="2025-10-02 01:57:31.980832254 +0000 UTC m=+989.147576294" observedRunningTime="2025-10-02 01:57:36.250218641 +0000 UTC m=+993.416962681" watchObservedRunningTime="2025-10-02 01:57:36.254639612 +0000 UTC m=+993.421383662" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.269274 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6" podStartSLOduration=4.063549943 podStartE2EDuration="24.269252429s" podCreationTimestamp="2025-10-02 01:57:12 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.877170385 +0000 UTC m=+971.043914425" lastFinishedPulling="2025-10-02 01:57:34.082872791 +0000 UTC m=+991.249616911" observedRunningTime="2025-10-02 01:57:36.262607712 +0000 UTC m=+993.429351752" watchObservedRunningTime="2025-10-02 01:57:36.269252429 +0000 UTC m=+993.435996489" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.275771 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-kv2nm" podStartSLOduration=7.112028852 podStartE2EDuration="25.275752663s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.816057206 +0000 UTC m=+970.982801246" lastFinishedPulling="2025-10-02 01:57:31.979781017 +0000 UTC m=+989.146525057" observedRunningTime="2025-10-02 01:57:36.274356807 +0000 UTC m=+993.441100877" watchObservedRunningTime="2025-10-02 01:57:36.275752663 +0000 UTC m=+993.442496713" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.291783 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-4gjk2" podStartSLOduration=6.589119107 podStartE2EDuration="25.291763955s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.226140508 +0000 UTC m=+970.392884548" lastFinishedPulling="2025-10-02 01:57:31.928785356 +0000 UTC m=+989.095529396" observedRunningTime="2025-10-02 01:57:36.289272102 +0000 UTC m=+993.456016172" watchObservedRunningTime="2025-10-02 01:57:36.291763955 +0000 UTC m=+993.458507995" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.323542 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jvrxr" podStartSLOduration=7.166118766 podStartE2EDuration="25.323513223s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.363150927 +0000 UTC m=+970.529894967" lastFinishedPulling="2025-10-02 01:57:31.520545384 +0000 UTC m=+988.687289424" observedRunningTime="2025-10-02 01:57:36.314398974 +0000 UTC m=+993.481143054" watchObservedRunningTime="2025-10-02 01:57:36.323513223 +0000 UTC m=+993.490257293" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.335563 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2zft5" podStartSLOduration=6.728996058 podStartE2EDuration="25.335545185s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.363510596 +0000 UTC m=+970.530254656" lastFinishedPulling="2025-10-02 01:57:31.970059743 +0000 UTC m=+989.136803783" observedRunningTime="2025-10-02 01:57:36.329415911 +0000 UTC m=+993.496159951" watchObservedRunningTime="2025-10-02 01:57:36.335545185 +0000 UTC m=+993.502289215" Oct 02 01:57:36 crc kubenswrapper[4775]: I1002 01:57:36.351645 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-mx8pj" podStartSLOduration=6.090386812 podStartE2EDuration="25.35162187s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:12.710008115 +0000 UTC m=+969.876752155" lastFinishedPulling="2025-10-02 01:57:31.971243173 +0000 UTC m=+989.137987213" observedRunningTime="2025-10-02 01:57:36.344326006 +0000 UTC m=+993.511070056" watchObservedRunningTime="2025-10-02 01:57:36.35162187 +0000 UTC m=+993.518365910" Oct 02 01:57:37 crc kubenswrapper[4775]: I1002 01:57:37.215677 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2zft5" Oct 02 01:57:41 crc kubenswrapper[4775]: I1002 01:57:41.731870 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-6ff8b75857-5wrms" Oct 02 01:57:41 crc kubenswrapper[4775]: I1002 01:57:41.794078 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-644bddb6d8-mx8pj" Oct 02 01:57:41 crc kubenswrapper[4775]: I1002 01:57:41.862323 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-84f4f7b77b-4gjk2" Oct 02 01:57:41 crc kubenswrapper[4775]: I1002 01:57:41.971326 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-5bd55b4bff-jvrxr" Oct 02 01:57:42 crc kubenswrapper[4775]: I1002 01:57:42.110265 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84958c4d49-md78z" Oct 02 01:57:42 crc kubenswrapper[4775]: I1002 01:57:42.250666 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5cd4858477-f2khq" Oct 02 01:57:42 crc kubenswrapper[4775]: I1002 01:57:42.324679 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-849d5b9b84-2zft5" Oct 02 01:57:42 crc kubenswrapper[4775]: I1002 01:57:42.359693 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-88c7-qkvrn" Oct 02 01:57:42 crc kubenswrapper[4775]: I1002 01:57:42.449468 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-85777745bb-bpg5g" Oct 02 01:57:42 crc kubenswrapper[4775]: I1002 01:57:42.522626 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-9d6c5db85-jt9hq" Oct 02 01:57:42 crc kubenswrapper[4775]: I1002 01:57:42.617255 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-9976ff44c-xhpvs" Oct 02 01:57:42 crc kubenswrapper[4775]: I1002 01:57:42.625994 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-589c58c6c-kv2nm" Oct 02 01:57:42 crc kubenswrapper[4775]: I1002 01:57:42.643766 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-84d6b4b759-ptq59" Oct 02 01:57:42 crc kubenswrapper[4775]: I1002 01:57:42.891504 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-5869cb545-9qscs" Oct 02 01:57:50 crc kubenswrapper[4775]: I1002 01:57:50.329785 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49" event={"ID":"c815738f-cb2a-4136-9320-f75ceea923c5","Type":"ContainerStarted","Data":"40cf8f8b173b14f04c2eef8c323498bc0befc6bf5fbbc9e8b6969b2ae4163432"} Oct 02 01:57:50 crc kubenswrapper[4775]: I1002 01:57:50.331380 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw" event={"ID":"7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62","Type":"ContainerStarted","Data":"334913ba680894c0b714c426156c5fb4089194cbf1cc7dd2666d55e1fbb7b53e"} Oct 02 01:57:50 crc kubenswrapper[4775]: I1002 01:57:50.332477 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49" Oct 02 01:57:50 crc kubenswrapper[4775]: I1002 01:57:50.332525 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw" Oct 02 01:57:50 crc kubenswrapper[4775]: I1002 01:57:50.334509 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65" event={"ID":"52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2","Type":"ContainerStarted","Data":"0e4824172061f0f62b74b56d713ada31618b95d527f17a9e3472b6c31d6d0ac3"} Oct 02 01:57:50 crc kubenswrapper[4775]: I1002 01:57:50.335051 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65" Oct 02 01:57:50 crc kubenswrapper[4775]: I1002 01:57:50.355804 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw" podStartSLOduration=2.127512489 podStartE2EDuration="38.355779561s" podCreationTimestamp="2025-10-02 01:57:12 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.863906963 +0000 UTC m=+971.030651003" lastFinishedPulling="2025-10-02 01:57:50.092174025 +0000 UTC m=+1007.258918075" observedRunningTime="2025-10-02 01:57:50.346646912 +0000 UTC m=+1007.513390962" watchObservedRunningTime="2025-10-02 01:57:50.355779561 +0000 UTC m=+1007.522523611" Oct 02 01:57:50 crc kubenswrapper[4775]: I1002 01:57:50.366361 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49" podStartSLOduration=2.9489421780000002 podStartE2EDuration="39.366341847s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.156326764 +0000 UTC m=+970.323070804" lastFinishedPulling="2025-10-02 01:57:49.573726393 +0000 UTC m=+1006.740470473" observedRunningTime="2025-10-02 01:57:50.364109341 +0000 UTC m=+1007.530853391" watchObservedRunningTime="2025-10-02 01:57:50.366341847 +0000 UTC m=+1007.533085887" Oct 02 01:57:50 crc kubenswrapper[4775]: I1002 01:57:50.390134 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65" podStartSLOduration=2.978930487 podStartE2EDuration="39.390111754s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.16467844 +0000 UTC m=+970.331422480" lastFinishedPulling="2025-10-02 01:57:49.575859667 +0000 UTC m=+1006.742603747" observedRunningTime="2025-10-02 01:57:50.385123839 +0000 UTC m=+1007.551867919" watchObservedRunningTime="2025-10-02 01:57:50.390111754 +0000 UTC m=+1007.556855794" Oct 02 01:57:51 crc kubenswrapper[4775]: I1002 01:57:51.350917 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2" event={"ID":"749bf5d6-b257-47e7-b632-e0edf0321adf","Type":"ContainerStarted","Data":"4bb26679a4861040a30f0b9e12ac452a8322ac5fb1d7d3cb3d4eeca1029f9ffd"} Oct 02 01:57:51 crc kubenswrapper[4775]: I1002 01:57:51.351449 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2" Oct 02 01:57:51 crc kubenswrapper[4775]: I1002 01:57:51.370616 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2" podStartSLOduration=4.001785222 podStartE2EDuration="40.370596109s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.876308453 +0000 UTC m=+971.043052493" lastFinishedPulling="2025-10-02 01:57:50.24511934 +0000 UTC m=+1007.411863380" observedRunningTime="2025-10-02 01:57:51.365725957 +0000 UTC m=+1008.532470017" watchObservedRunningTime="2025-10-02 01:57:51.370596109 +0000 UTC m=+1008.537340149" Oct 02 01:57:52 crc kubenswrapper[4775]: I1002 01:57:52.359898 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb" event={"ID":"c066e712-5c8d-48f8-b2fe-daa26e9f53e6","Type":"ContainerStarted","Data":"5ba0af635a888a713c52b48c0d36dc31ff49b132417ebf45e12235bde996f22f"} Oct 02 01:57:52 crc kubenswrapper[4775]: I1002 01:57:52.360186 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb" Oct 02 01:57:52 crc kubenswrapper[4775]: I1002 01:57:52.362428 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr" event={"ID":"03d2a01f-1564-46f8-9648-e4a826470b44","Type":"ContainerStarted","Data":"8657a03f753d774852713461e954bffb6eacbbfadf0c5d5c8f570734aa6ed905"} Oct 02 01:57:52 crc kubenswrapper[4775]: I1002 01:57:52.386784 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb" podStartSLOduration=3.515351611 podStartE2EDuration="41.386755972s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.367597102 +0000 UTC m=+970.534341182" lastFinishedPulling="2025-10-02 01:57:51.239001493 +0000 UTC m=+1008.405745543" observedRunningTime="2025-10-02 01:57:52.380414313 +0000 UTC m=+1009.547158363" watchObservedRunningTime="2025-10-02 01:57:52.386755972 +0000 UTC m=+1009.553500022" Oct 02 01:57:52 crc kubenswrapper[4775]: I1002 01:57:52.402941 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr" podStartSLOduration=3.009663749 podStartE2EDuration="41.402920788s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:12.947594043 +0000 UTC m=+970.114338083" lastFinishedPulling="2025-10-02 01:57:51.340851082 +0000 UTC m=+1008.507595122" observedRunningTime="2025-10-02 01:57:52.401325798 +0000 UTC m=+1009.568069848" watchObservedRunningTime="2025-10-02 01:57:52.402920788 +0000 UTC m=+1009.569664828" Oct 02 01:57:53 crc kubenswrapper[4775]: I1002 01:57:53.373292 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k" event={"ID":"c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f","Type":"ContainerStarted","Data":"790bcb8ba9e39fd122bea4546715c8bdf9810fcbf1dd421e9f9967946364d850"} Oct 02 01:57:53 crc kubenswrapper[4775]: I1002 01:57:53.373996 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k" Oct 02 01:57:53 crc kubenswrapper[4775]: I1002 01:57:53.405168 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k" podStartSLOduration=4.017698964 podStartE2EDuration="42.405147531s" podCreationTimestamp="2025-10-02 01:57:11 +0000 UTC" firstStartedPulling="2025-10-02 01:57:13.876686953 +0000 UTC m=+971.043430993" lastFinishedPulling="2025-10-02 01:57:52.26413548 +0000 UTC m=+1009.430879560" observedRunningTime="2025-10-02 01:57:53.398431062 +0000 UTC m=+1010.565175142" watchObservedRunningTime="2025-10-02 01:57:53.405147531 +0000 UTC m=+1010.571891581" Oct 02 01:58:01 crc kubenswrapper[4775]: I1002 01:58:01.886507 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5d889d78cf-5nl49" Oct 02 01:58:01 crc kubenswrapper[4775]: I1002 01:58:01.902668 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr" Oct 02 01:58:01 crc kubenswrapper[4775]: I1002 01:58:01.907507 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-9f4696d94-2mxjr" Oct 02 01:58:02 crc kubenswrapper[4775]: I1002 01:58:02.265785 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-64cd67b5cb-5vt65" Oct 02 01:58:02 crc kubenswrapper[4775]: I1002 01:58:02.331120 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-6d68dbc695-kvhkb" Oct 02 01:58:02 crc kubenswrapper[4775]: I1002 01:58:02.583716 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-7b787867f4-lqrg2" Oct 02 01:58:02 crc kubenswrapper[4775]: I1002 01:58:02.659880 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-b8d54b5d7-vbs7k" Oct 02 01:58:02 crc kubenswrapper[4775]: I1002 01:58:02.776370 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6b9957f54f-mlksw" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.710013 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jhkkr"] Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.712113 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jhkkr" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.716484 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.716874 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.717116 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.725785 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jhkkr"] Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.716707 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-htjwg" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.770333 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rsq8b"] Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.771873 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.778766 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.788275 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rsq8b"] Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.834278 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngrk8\" (UniqueName: \"kubernetes.io/projected/95958cf1-0438-4b2e-bb7b-a5f46188b59b-kube-api-access-ngrk8\") pod \"dnsmasq-dns-675f4bcbfc-jhkkr\" (UID: \"95958cf1-0438-4b2e-bb7b-a5f46188b59b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jhkkr" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.834320 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95958cf1-0438-4b2e-bb7b-a5f46188b59b-config\") pod \"dnsmasq-dns-675f4bcbfc-jhkkr\" (UID: \"95958cf1-0438-4b2e-bb7b-a5f46188b59b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jhkkr" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.935540 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m8jf\" (UniqueName: \"kubernetes.io/projected/26e94503-c561-4fad-8d9d-81dcb08cc1f5-kube-api-access-8m8jf\") pod \"dnsmasq-dns-78dd6ddcc-rsq8b\" (UID: \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.935628 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngrk8\" (UniqueName: \"kubernetes.io/projected/95958cf1-0438-4b2e-bb7b-a5f46188b59b-kube-api-access-ngrk8\") pod \"dnsmasq-dns-675f4bcbfc-jhkkr\" (UID: \"95958cf1-0438-4b2e-bb7b-a5f46188b59b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jhkkr" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.935651 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26e94503-c561-4fad-8d9d-81dcb08cc1f5-config\") pod \"dnsmasq-dns-78dd6ddcc-rsq8b\" (UID: \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.935672 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95958cf1-0438-4b2e-bb7b-a5f46188b59b-config\") pod \"dnsmasq-dns-675f4bcbfc-jhkkr\" (UID: \"95958cf1-0438-4b2e-bb7b-a5f46188b59b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jhkkr" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.935757 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26e94503-c561-4fad-8d9d-81dcb08cc1f5-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rsq8b\" (UID: \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.936713 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95958cf1-0438-4b2e-bb7b-a5f46188b59b-config\") pod \"dnsmasq-dns-675f4bcbfc-jhkkr\" (UID: \"95958cf1-0438-4b2e-bb7b-a5f46188b59b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jhkkr" Oct 02 01:58:16 crc kubenswrapper[4775]: I1002 01:58:16.965536 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngrk8\" (UniqueName: \"kubernetes.io/projected/95958cf1-0438-4b2e-bb7b-a5f46188b59b-kube-api-access-ngrk8\") pod \"dnsmasq-dns-675f4bcbfc-jhkkr\" (UID: \"95958cf1-0438-4b2e-bb7b-a5f46188b59b\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jhkkr" Oct 02 01:58:17 crc kubenswrapper[4775]: I1002 01:58:17.036817 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26e94503-c561-4fad-8d9d-81dcb08cc1f5-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rsq8b\" (UID: \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" Oct 02 01:58:17 crc kubenswrapper[4775]: I1002 01:58:17.036893 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m8jf\" (UniqueName: \"kubernetes.io/projected/26e94503-c561-4fad-8d9d-81dcb08cc1f5-kube-api-access-8m8jf\") pod \"dnsmasq-dns-78dd6ddcc-rsq8b\" (UID: \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" Oct 02 01:58:17 crc kubenswrapper[4775]: I1002 01:58:17.036933 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26e94503-c561-4fad-8d9d-81dcb08cc1f5-config\") pod \"dnsmasq-dns-78dd6ddcc-rsq8b\" (UID: \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" Oct 02 01:58:17 crc kubenswrapper[4775]: I1002 01:58:17.037819 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26e94503-c561-4fad-8d9d-81dcb08cc1f5-config\") pod \"dnsmasq-dns-78dd6ddcc-rsq8b\" (UID: \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" Oct 02 01:58:17 crc kubenswrapper[4775]: I1002 01:58:17.038302 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26e94503-c561-4fad-8d9d-81dcb08cc1f5-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rsq8b\" (UID: \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" Oct 02 01:58:17 crc kubenswrapper[4775]: I1002 01:58:17.054232 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jhkkr" Oct 02 01:58:17 crc kubenswrapper[4775]: I1002 01:58:17.063358 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m8jf\" (UniqueName: \"kubernetes.io/projected/26e94503-c561-4fad-8d9d-81dcb08cc1f5-kube-api-access-8m8jf\") pod \"dnsmasq-dns-78dd6ddcc-rsq8b\" (UID: \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" Oct 02 01:58:17 crc kubenswrapper[4775]: I1002 01:58:17.090920 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" Oct 02 01:58:17 crc kubenswrapper[4775]: I1002 01:58:17.514268 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rsq8b"] Oct 02 01:58:17 crc kubenswrapper[4775]: I1002 01:58:17.523701 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 01:58:17 crc kubenswrapper[4775]: I1002 01:58:17.564925 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jhkkr"] Oct 02 01:58:17 crc kubenswrapper[4775]: W1002 01:58:17.572755 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95958cf1_0438_4b2e_bb7b_a5f46188b59b.slice/crio-0909db59bf30b5abf14237919ca85523a034700ba024abaaebbca29f2e846d8d WatchSource:0}: Error finding container 0909db59bf30b5abf14237919ca85523a034700ba024abaaebbca29f2e846d8d: Status 404 returned error can't find the container with id 0909db59bf30b5abf14237919ca85523a034700ba024abaaebbca29f2e846d8d Oct 02 01:58:17 crc kubenswrapper[4775]: I1002 01:58:17.584717 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-jhkkr" event={"ID":"95958cf1-0438-4b2e-bb7b-a5f46188b59b","Type":"ContainerStarted","Data":"0909db59bf30b5abf14237919ca85523a034700ba024abaaebbca29f2e846d8d"} Oct 02 01:58:17 crc kubenswrapper[4775]: I1002 01:58:17.587652 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" event={"ID":"26e94503-c561-4fad-8d9d-81dcb08cc1f5","Type":"ContainerStarted","Data":"e138f151de84ef99d422e5e3e1a938702212af731197cf747628069f15ce0e51"} Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.319782 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jhkkr"] Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.349142 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nb2rc"] Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.350273 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.365086 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nb2rc"] Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.473622 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-nb2rc\" (UID: \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.473716 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwwth\" (UniqueName: \"kubernetes.io/projected/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-kube-api-access-zwwth\") pod \"dnsmasq-dns-666b6646f7-nb2rc\" (UID: \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.473764 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-config\") pod \"dnsmasq-dns-666b6646f7-nb2rc\" (UID: \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.575225 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-config\") pod \"dnsmasq-dns-666b6646f7-nb2rc\" (UID: \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.575311 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-nb2rc\" (UID: \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.575351 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwwth\" (UniqueName: \"kubernetes.io/projected/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-kube-api-access-zwwth\") pod \"dnsmasq-dns-666b6646f7-nb2rc\" (UID: \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.576886 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-dns-svc\") pod \"dnsmasq-dns-666b6646f7-nb2rc\" (UID: \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.577023 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-config\") pod \"dnsmasq-dns-666b6646f7-nb2rc\" (UID: \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.609398 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwwth\" (UniqueName: \"kubernetes.io/projected/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-kube-api-access-zwwth\") pod \"dnsmasq-dns-666b6646f7-nb2rc\" (UID: \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\") " pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.665016 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rsq8b"] Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.673719 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-b6f52"] Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.679608 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.680852 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-b6f52"] Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.686983 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.780721 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzl2c\" (UniqueName: \"kubernetes.io/projected/0af3416a-b960-4ac6-871c-26c4f0d58368-kube-api-access-dzl2c\") pod \"dnsmasq-dns-57d769cc4f-b6f52\" (UID: \"0af3416a-b960-4ac6-871c-26c4f0d58368\") " pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.780764 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0af3416a-b960-4ac6-871c-26c4f0d58368-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-b6f52\" (UID: \"0af3416a-b960-4ac6-871c-26c4f0d58368\") " pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.780806 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0af3416a-b960-4ac6-871c-26c4f0d58368-config\") pod \"dnsmasq-dns-57d769cc4f-b6f52\" (UID: \"0af3416a-b960-4ac6-871c-26c4f0d58368\") " pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.897046 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzl2c\" (UniqueName: \"kubernetes.io/projected/0af3416a-b960-4ac6-871c-26c4f0d58368-kube-api-access-dzl2c\") pod \"dnsmasq-dns-57d769cc4f-b6f52\" (UID: \"0af3416a-b960-4ac6-871c-26c4f0d58368\") " pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.897080 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0af3416a-b960-4ac6-871c-26c4f0d58368-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-b6f52\" (UID: \"0af3416a-b960-4ac6-871c-26c4f0d58368\") " pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.897192 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0af3416a-b960-4ac6-871c-26c4f0d58368-config\") pod \"dnsmasq-dns-57d769cc4f-b6f52\" (UID: \"0af3416a-b960-4ac6-871c-26c4f0d58368\") " pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.898998 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0af3416a-b960-4ac6-871c-26c4f0d58368-config\") pod \"dnsmasq-dns-57d769cc4f-b6f52\" (UID: \"0af3416a-b960-4ac6-871c-26c4f0d58368\") " pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.899587 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0af3416a-b960-4ac6-871c-26c4f0d58368-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-b6f52\" (UID: \"0af3416a-b960-4ac6-871c-26c4f0d58368\") " pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:58:19 crc kubenswrapper[4775]: I1002 01:58:19.919104 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzl2c\" (UniqueName: \"kubernetes.io/projected/0af3416a-b960-4ac6-871c-26c4f0d58368-kube-api-access-dzl2c\") pod \"dnsmasq-dns-57d769cc4f-b6f52\" (UID: \"0af3416a-b960-4ac6-871c-26c4f0d58368\") " pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.012705 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.239582 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nb2rc"] Oct 02 01:58:20 crc kubenswrapper[4775]: W1002 01:58:20.247818 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06528afb_0cc3_45fc_87d1_e2b0dc88f8e1.slice/crio-81b09fdae5a25645adbf8d337b628e78810306c6fdbd2bb0e8fd5e613c81c535 WatchSource:0}: Error finding container 81b09fdae5a25645adbf8d337b628e78810306c6fdbd2bb0e8fd5e613c81c535: Status 404 returned error can't find the container with id 81b09fdae5a25645adbf8d337b628e78810306c6fdbd2bb0e8fd5e613c81c535 Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.498029 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-b6f52"] Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.508340 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.509739 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.512020 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.512156 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.512178 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.512307 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.512322 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.512419 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.512519 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-qw7sd" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.514509 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 01:58:20 crc kubenswrapper[4775]: W1002 01:58:20.520711 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0af3416a_b960_4ac6_871c_26c4f0d58368.slice/crio-70de01799ae7846bf3ad95e80b78859b7b0b254163fa61d7d936ee2112f5762b WatchSource:0}: Error finding container 70de01799ae7846bf3ad95e80b78859b7b0b254163fa61d7d936ee2112f5762b: Status 404 returned error can't find the container with id 70de01799ae7846bf3ad95e80b78859b7b0b254163fa61d7d936ee2112f5762b Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.616488 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" event={"ID":"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1","Type":"ContainerStarted","Data":"81b09fdae5a25645adbf8d337b628e78810306c6fdbd2bb0e8fd5e613c81c535"} Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.617379 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" event={"ID":"0af3416a-b960-4ac6-871c-26c4f0d58368","Type":"ContainerStarted","Data":"70de01799ae7846bf3ad95e80b78859b7b0b254163fa61d7d936ee2112f5762b"} Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.710856 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.710898 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.710918 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.710995 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.711024 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.711048 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.711073 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6638baa9-b367-40d1-8111-673dca6434ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.711206 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.711275 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6638baa9-b367-40d1-8111-673dca6434ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.711306 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.711326 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkgpt\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-kube-api-access-bkgpt\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.765423 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.766860 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.770104 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.770407 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.770419 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.770442 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.771749 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.772187 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.772313 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-w7jpj" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.786770 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.812143 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6638baa9-b367-40d1-8111-673dca6434ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.812193 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.812213 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkgpt\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-kube-api-access-bkgpt\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.812250 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.812272 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.812289 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.812314 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.812339 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.812360 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.812384 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6638baa9-b367-40d1-8111-673dca6434ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.812409 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.813228 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.813411 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.813602 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.813794 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.814848 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.816512 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.820926 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.821358 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6638baa9-b367-40d1-8111-673dca6434ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.822005 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.822804 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6638baa9-b367-40d1-8111-673dca6434ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.843093 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkgpt\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-kube-api-access-bkgpt\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.876448 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-server-0\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " pod="openstack/rabbitmq-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.920625 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.920682 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.920707 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwbt6\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-kube-api-access-wwbt6\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.920730 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.920752 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.920776 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.920802 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.920827 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.920848 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.920876 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:20 crc kubenswrapper[4775]: I1002 01:58:20.920891 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.022121 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.022162 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.022222 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.022242 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.022269 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwbt6\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-kube-api-access-wwbt6\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.022299 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.022324 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.022352 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.022381 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.022402 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.022422 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.023490 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.023543 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.023709 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.023744 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.023863 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.024334 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.027978 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.028652 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.033975 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.037249 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.044323 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwbt6\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-kube-api-access-wwbt6\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.046903 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.095510 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:58:21 crc kubenswrapper[4775]: I1002 01:58:21.139762 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 01:58:22 crc kubenswrapper[4775]: I1002 01:58:22.928559 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 02 01:58:22 crc kubenswrapper[4775]: I1002 01:58:22.930754 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 01:58:22 crc kubenswrapper[4775]: I1002 01:58:22.933334 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-6q4hq" Oct 02 01:58:22 crc kubenswrapper[4775]: I1002 01:58:22.934541 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 02 01:58:22 crc kubenswrapper[4775]: I1002 01:58:22.934705 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 02 01:58:22 crc kubenswrapper[4775]: I1002 01:58:22.934783 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 02 01:58:22 crc kubenswrapper[4775]: I1002 01:58:22.935433 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 02 01:58:22 crc kubenswrapper[4775]: I1002 01:58:22.944564 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 02 01:58:22 crc kubenswrapper[4775]: I1002 01:58:22.950029 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.059501 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzcxp\" (UniqueName: \"kubernetes.io/projected/002eb223-af34-4c80-9570-894add9b0e3c-kube-api-access-vzcxp\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.059578 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.059603 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.059673 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.059703 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-secrets\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.059740 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.059764 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-config-data-default\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.059788 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-kolla-config\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.059813 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/002eb223-af34-4c80-9570-894add9b0e3c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.161447 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.162235 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-config-data-default\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.162409 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-kolla-config\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.162562 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/002eb223-af34-4c80-9570-894add9b0e3c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.162756 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzcxp\" (UniqueName: \"kubernetes.io/projected/002eb223-af34-4c80-9570-894add9b0e3c-kube-api-access-vzcxp\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.162889 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.163025 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.161993 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.163316 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.163440 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-secrets\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.163489 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/002eb223-af34-4c80-9570-894add9b0e3c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.164376 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-config-data-default\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.165227 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-kolla-config\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.165899 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.167583 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-secrets\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.169292 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.169384 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.183500 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzcxp\" (UniqueName: \"kubernetes.io/projected/002eb223-af34-4c80-9570-894add9b0e3c-kube-api-access-vzcxp\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.184300 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.252109 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.275705 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.277252 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.279748 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-wwnbc" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.280921 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.281077 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.281846 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.287073 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.471494 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.471591 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.471658 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3adc5647-1145-4c44-9ecc-66d1d9a19023-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.471695 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.471754 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.471821 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.472026 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wph9j\" (UniqueName: \"kubernetes.io/projected/3adc5647-1145-4c44-9ecc-66d1d9a19023-kube-api-access-wph9j\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.472081 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.472145 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.573488 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.573741 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.573805 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wph9j\" (UniqueName: \"kubernetes.io/projected/3adc5647-1145-4c44-9ecc-66d1d9a19023-kube-api-access-wph9j\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.573831 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.573865 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.573901 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.573926 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.573976 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3adc5647-1145-4c44-9ecc-66d1d9a19023-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.574022 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.576475 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.576724 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3adc5647-1145-4c44-9ecc-66d1d9a19023-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.576975 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.577277 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.577431 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.584522 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.585558 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.595910 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.596096 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wph9j\" (UniqueName: \"kubernetes.io/projected/3adc5647-1145-4c44-9ecc-66d1d9a19023-kube-api-access-wph9j\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.610751 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.620001 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.786708 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.787583 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.792387 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.792422 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-mx2kq" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.792455 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.803111 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.878935 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406dd24f-d800-44be-b33d-019c532c2feb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.878990 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/406dd24f-d800-44be-b33d-019c532c2feb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.879131 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/406dd24f-d800-44be-b33d-019c532c2feb-kolla-config\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.879224 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/406dd24f-d800-44be-b33d-019c532c2feb-config-data\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.879392 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb4p8\" (UniqueName: \"kubernetes.io/projected/406dd24f-d800-44be-b33d-019c532c2feb-kube-api-access-zb4p8\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.979906 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb4p8\" (UniqueName: \"kubernetes.io/projected/406dd24f-d800-44be-b33d-019c532c2feb-kube-api-access-zb4p8\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.979986 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406dd24f-d800-44be-b33d-019c532c2feb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.980016 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/406dd24f-d800-44be-b33d-019c532c2feb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.980066 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/406dd24f-d800-44be-b33d-019c532c2feb-kolla-config\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.980110 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/406dd24f-d800-44be-b33d-019c532c2feb-config-data\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.980799 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/406dd24f-d800-44be-b33d-019c532c2feb-config-data\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.981004 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/406dd24f-d800-44be-b33d-019c532c2feb-kolla-config\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.984088 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406dd24f-d800-44be-b33d-019c532c2feb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:23 crc kubenswrapper[4775]: I1002 01:58:23.999496 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/406dd24f-d800-44be-b33d-019c532c2feb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:24 crc kubenswrapper[4775]: I1002 01:58:24.016598 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb4p8\" (UniqueName: \"kubernetes.io/projected/406dd24f-d800-44be-b33d-019c532c2feb-kube-api-access-zb4p8\") pod \"memcached-0\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " pod="openstack/memcached-0" Oct 02 01:58:24 crc kubenswrapper[4775]: I1002 01:58:24.115001 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 01:58:25 crc kubenswrapper[4775]: I1002 01:58:25.678725 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 01:58:25 crc kubenswrapper[4775]: I1002 01:58:25.679988 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 01:58:25 crc kubenswrapper[4775]: I1002 01:58:25.681867 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-7v87d" Oct 02 01:58:25 crc kubenswrapper[4775]: I1002 01:58:25.687162 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 01:58:25 crc kubenswrapper[4775]: I1002 01:58:25.715064 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw9m9\" (UniqueName: \"kubernetes.io/projected/6f55acc8-9fa8-496b-a882-d86ea3f28730-kube-api-access-fw9m9\") pod \"kube-state-metrics-0\" (UID: \"6f55acc8-9fa8-496b-a882-d86ea3f28730\") " pod="openstack/kube-state-metrics-0" Oct 02 01:58:25 crc kubenswrapper[4775]: I1002 01:58:25.816400 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw9m9\" (UniqueName: \"kubernetes.io/projected/6f55acc8-9fa8-496b-a882-d86ea3f28730-kube-api-access-fw9m9\") pod \"kube-state-metrics-0\" (UID: \"6f55acc8-9fa8-496b-a882-d86ea3f28730\") " pod="openstack/kube-state-metrics-0" Oct 02 01:58:25 crc kubenswrapper[4775]: I1002 01:58:25.833711 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw9m9\" (UniqueName: \"kubernetes.io/projected/6f55acc8-9fa8-496b-a882-d86ea3f28730-kube-api-access-fw9m9\") pod \"kube-state-metrics-0\" (UID: \"6f55acc8-9fa8-496b-a882-d86ea3f28730\") " pod="openstack/kube-state-metrics-0" Oct 02 01:58:26 crc kubenswrapper[4775]: I1002 01:58:26.045913 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.799507 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.801544 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.803928 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.805047 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.805216 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-x4sl8" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.805366 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.805490 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.808635 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jdbs4"] Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.809661 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.811695 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.811794 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.811870 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-5s6rj" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.829907 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.849469 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jdbs4"] Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.865864 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-xwnrv"] Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.868700 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.874246 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-xwnrv"] Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.981932 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982014 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982039 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-lib\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982057 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982099 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-log-ovn\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982119 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982163 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-etc-ovs\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982188 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6efbe4b3-953d-4927-95c6-9b92708f51eb-scripts\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982283 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-run\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982314 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27jdw\" (UniqueName: \"kubernetes.io/projected/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-kube-api-access-27jdw\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982363 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982385 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfxrf\" (UniqueName: \"kubernetes.io/projected/6efbe4b3-953d-4927-95c6-9b92708f51eb-kube-api-access-vfxrf\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982404 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-run\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982451 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efbe4b3-953d-4927-95c6-9b92708f51eb-combined-ca-bundle\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982474 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-config\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982509 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-run-ovn\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982535 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982558 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-log\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982600 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb7bf\" (UniqueName: \"kubernetes.io/projected/2a54aa55-e3e2-4844-b175-e68442220b92-kube-api-access-cb7bf\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982616 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6efbe4b3-953d-4927-95c6-9b92708f51eb-ovn-controller-tls-certs\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:29 crc kubenswrapper[4775]: I1002 01:58:29.982831 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a54aa55-e3e2-4844-b175-e68442220b92-scripts\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084043 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27jdw\" (UniqueName: \"kubernetes.io/projected/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-kube-api-access-27jdw\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084083 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084110 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfxrf\" (UniqueName: \"kubernetes.io/projected/6efbe4b3-953d-4927-95c6-9b92708f51eb-kube-api-access-vfxrf\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084129 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efbe4b3-953d-4927-95c6-9b92708f51eb-combined-ca-bundle\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084149 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-run\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084168 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-config\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084184 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-run-ovn\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084208 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084236 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-log\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084254 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb7bf\" (UniqueName: \"kubernetes.io/projected/2a54aa55-e3e2-4844-b175-e68442220b92-kube-api-access-cb7bf\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084271 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6efbe4b3-953d-4927-95c6-9b92708f51eb-ovn-controller-tls-certs\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084291 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a54aa55-e3e2-4844-b175-e68442220b92-scripts\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084316 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084336 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084357 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-lib\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084379 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084398 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-log-ovn\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084419 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084436 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-etc-ovs\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084455 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6efbe4b3-953d-4927-95c6-9b92708f51eb-scripts\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084483 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-run\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084679 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-run\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084693 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-run\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084698 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084850 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-log-ovn\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084937 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-log\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.084986 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-run-ovn\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.085077 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.085146 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-lib\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.085204 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-etc-ovs\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.086210 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.088022 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a54aa55-e3e2-4844-b175-e68442220b92-scripts\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.089133 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6efbe4b3-953d-4927-95c6-9b92708f51eb-scripts\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.090152 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-config\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.090816 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efbe4b3-953d-4927-95c6-9b92708f51eb-combined-ca-bundle\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.091075 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6efbe4b3-953d-4927-95c6-9b92708f51eb-ovn-controller-tls-certs\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.110969 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.110976 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.111116 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27jdw\" (UniqueName: \"kubernetes.io/projected/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-kube-api-access-27jdw\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.111578 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfxrf\" (UniqueName: \"kubernetes.io/projected/6efbe4b3-953d-4927-95c6-9b92708f51eb-kube-api-access-vfxrf\") pod \"ovn-controller-jdbs4\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.112272 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb7bf\" (UniqueName: \"kubernetes.io/projected/2a54aa55-e3e2-4844-b175-e68442220b92-kube-api-access-cb7bf\") pod \"ovn-controller-ovs-xwnrv\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.113620 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.130004 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ovsdbserver-nb-0\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.134477 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.188353 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:30 crc kubenswrapper[4775]: I1002 01:58:30.424756 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.093498 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.234362 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.235741 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.237765 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-slgf2" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.238533 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.241060 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.241399 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.258384 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.354490 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mm56r\" (UniqueName: \"kubernetes.io/projected/f7685ff7-607b-416c-9382-2c00edecaa54-kube-api-access-mm56r\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.354555 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7685ff7-607b-416c-9382-2c00edecaa54-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.354585 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7685ff7-607b-416c-9382-2c00edecaa54-config\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.354685 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.354853 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.354895 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.355306 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f7685ff7-607b-416c-9382-2c00edecaa54-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.355371 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.457190 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.457248 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.457295 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f7685ff7-607b-416c-9382-2c00edecaa54-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.457338 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.457406 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mm56r\" (UniqueName: \"kubernetes.io/projected/f7685ff7-607b-416c-9382-2c00edecaa54-kube-api-access-mm56r\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.457441 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7685ff7-607b-416c-9382-2c00edecaa54-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.457464 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7685ff7-607b-416c-9382-2c00edecaa54-config\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.457486 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.457666 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.458909 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f7685ff7-607b-416c-9382-2c00edecaa54-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.461511 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7685ff7-607b-416c-9382-2c00edecaa54-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.461726 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7685ff7-607b-416c-9382-2c00edecaa54-config\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.470360 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.471388 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.479498 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.482867 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mm56r\" (UniqueName: \"kubernetes.io/projected/f7685ff7-607b-416c-9382-2c00edecaa54-kube-api-access-mm56r\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.490130 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: I1002 01:58:33.564647 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:33 crc kubenswrapper[4775]: E1002 01:58:33.810305 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 01:58:33 crc kubenswrapper[4775]: E1002 01:58:33.810480 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ngrk8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-jhkkr_openstack(95958cf1-0438-4b2e-bb7b-a5f46188b59b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 01:58:33 crc kubenswrapper[4775]: E1002 01:58:33.811851 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-jhkkr" podUID="95958cf1-0438-4b2e-bb7b-a5f46188b59b" Oct 02 01:58:34 crc kubenswrapper[4775]: E1002 01:58:34.184911 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 02 01:58:34 crc kubenswrapper[4775]: E1002 01:58:34.185704 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8m8jf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-rsq8b_openstack(26e94503-c561-4fad-8d9d-81dcb08cc1f5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 01:58:34 crc kubenswrapper[4775]: E1002 01:58:34.187040 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" podUID="26e94503-c561-4fad-8d9d-81dcb08cc1f5" Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.252770 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 01:58:34 crc kubenswrapper[4775]: W1002 01:58:34.255290 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6638baa9_b367_40d1_8111_673dca6434ad.slice/crio-dd1198b6383e7906624aea451ad61826c0a3208f8155d62582ed5f31e692daaa WatchSource:0}: Error finding container dd1198b6383e7906624aea451ad61826c0a3208f8155d62582ed5f31e692daaa: Status 404 returned error can't find the container with id dd1198b6383e7906624aea451ad61826c0a3208f8155d62582ed5f31e692daaa Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.414814 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.435762 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.459551 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.532356 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 01:58:34 crc kubenswrapper[4775]: W1002 01:58:34.535837 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5ce93aa_f0e3_48eb_b8a9_c9836edb3e92.slice/crio-e194882b58a7d56c86b9f7a5e4ff2fc880c4254d1368412ee77182aacb63d04d WatchSource:0}: Error finding container e194882b58a7d56c86b9f7a5e4ff2fc880c4254d1368412ee77182aacb63d04d: Status 404 returned error can't find the container with id e194882b58a7d56c86b9f7a5e4ff2fc880c4254d1368412ee77182aacb63d04d Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.593010 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 01:58:34 crc kubenswrapper[4775]: W1002 01:58:34.604274 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod406dd24f_d800_44be_b33d_019c532c2feb.slice/crio-fe62491b1f99d977bbcdd66d6c9f7ce148e13044528e146f34627800b201ae60 WatchSource:0}: Error finding container fe62491b1f99d977bbcdd66d6c9f7ce148e13044528e146f34627800b201ae60: Status 404 returned error can't find the container with id fe62491b1f99d977bbcdd66d6c9f7ce148e13044528e146f34627800b201ae60 Oct 02 01:58:34 crc kubenswrapper[4775]: W1002 01:58:34.605728 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6efbe4b3_953d_4927_95c6_9b92708f51eb.slice/crio-ce846bbdc94bc41919dce048b75479d3b6f39d5af12599c009c079ae055ad4bc WatchSource:0}: Error finding container ce846bbdc94bc41919dce048b75479d3b6f39d5af12599c009c079ae055ad4bc: Status 404 returned error can't find the container with id ce846bbdc94bc41919dce048b75479d3b6f39d5af12599c009c079ae055ad4bc Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.609218 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jdbs4"] Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.744204 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6f55acc8-9fa8-496b-a882-d86ea3f28730","Type":"ContainerStarted","Data":"a7952639931629a679e5654b1879e61531b64cacd1cb10b624464e76d444153e"} Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.745597 4775 generic.go:334] "Generic (PLEG): container finished" podID="0af3416a-b960-4ac6-871c-26c4f0d58368" containerID="4baa3fa4694b068174901ca15c9726a8b3eac40e59412c11c2ff60ca69a4901a" exitCode=0 Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.745641 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" event={"ID":"0af3416a-b960-4ac6-871c-26c4f0d58368","Type":"ContainerDied","Data":"4baa3fa4694b068174901ca15c9726a8b3eac40e59412c11c2ff60ca69a4901a"} Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.747925 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3adc5647-1145-4c44-9ecc-66d1d9a19023","Type":"ContainerStarted","Data":"7fa5b0f45bdda8d329e73a560098dae21985e99f3cfcb9a5ed11b49f0a4f88f7"} Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.749510 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"002eb223-af34-4c80-9570-894add9b0e3c","Type":"ContainerStarted","Data":"6e8fd5394954be2252799b84e7920c64d1173fc31dfcd3a7f3e1daaa735ff5bc"} Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.750599 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"406dd24f-d800-44be-b33d-019c532c2feb","Type":"ContainerStarted","Data":"fe62491b1f99d977bbcdd66d6c9f7ce148e13044528e146f34627800b201ae60"} Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.752744 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6638baa9-b367-40d1-8111-673dca6434ad","Type":"ContainerStarted","Data":"dd1198b6383e7906624aea451ad61826c0a3208f8155d62582ed5f31e692daaa"} Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.754718 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92","Type":"ContainerStarted","Data":"e194882b58a7d56c86b9f7a5e4ff2fc880c4254d1368412ee77182aacb63d04d"} Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.755890 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jdbs4" event={"ID":"6efbe4b3-953d-4927-95c6-9b92708f51eb","Type":"ContainerStarted","Data":"ce846bbdc94bc41919dce048b75479d3b6f39d5af12599c009c079ae055ad4bc"} Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.756753 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9","Type":"ContainerStarted","Data":"1afd2c34b0364a61dfeee78253b8f225d6fa44e8781067eff8c919f7b3f28494"} Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.758464 4775 generic.go:334] "Generic (PLEG): container finished" podID="06528afb-0cc3-45fc-87d1-e2b0dc88f8e1" containerID="ee162fedc8decca1c1e7552c19b226217fed8f4e18a3303849d5bf3fca181269" exitCode=0 Oct 02 01:58:34 crc kubenswrapper[4775]: I1002 01:58:34.760218 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" event={"ID":"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1","Type":"ContainerDied","Data":"ee162fedc8decca1c1e7552c19b226217fed8f4e18a3303849d5bf3fca181269"} Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.081454 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 01:58:35 crc kubenswrapper[4775]: W1002 01:58:35.097931 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7685ff7_607b_416c_9382_2c00edecaa54.slice/crio-e41ea836798b128f59bc00526da81b9cd7e04b857cb248a53de519eb40c2a59e WatchSource:0}: Error finding container e41ea836798b128f59bc00526da81b9cd7e04b857cb248a53de519eb40c2a59e: Status 404 returned error can't find the container with id e41ea836798b128f59bc00526da81b9cd7e04b857cb248a53de519eb40c2a59e Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.153852 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jhkkr" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.184362 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.293474 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26e94503-c561-4fad-8d9d-81dcb08cc1f5-config\") pod \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\" (UID: \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\") " Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.293515 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26e94503-c561-4fad-8d9d-81dcb08cc1f5-dns-svc\") pod \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\" (UID: \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\") " Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.293607 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngrk8\" (UniqueName: \"kubernetes.io/projected/95958cf1-0438-4b2e-bb7b-a5f46188b59b-kube-api-access-ngrk8\") pod \"95958cf1-0438-4b2e-bb7b-a5f46188b59b\" (UID: \"95958cf1-0438-4b2e-bb7b-a5f46188b59b\") " Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.293630 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95958cf1-0438-4b2e-bb7b-a5f46188b59b-config\") pod \"95958cf1-0438-4b2e-bb7b-a5f46188b59b\" (UID: \"95958cf1-0438-4b2e-bb7b-a5f46188b59b\") " Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.293657 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8m8jf\" (UniqueName: \"kubernetes.io/projected/26e94503-c561-4fad-8d9d-81dcb08cc1f5-kube-api-access-8m8jf\") pod \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\" (UID: \"26e94503-c561-4fad-8d9d-81dcb08cc1f5\") " Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.294151 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26e94503-c561-4fad-8d9d-81dcb08cc1f5-config" (OuterVolumeSpecName: "config") pod "26e94503-c561-4fad-8d9d-81dcb08cc1f5" (UID: "26e94503-c561-4fad-8d9d-81dcb08cc1f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.294649 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/26e94503-c561-4fad-8d9d-81dcb08cc1f5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "26e94503-c561-4fad-8d9d-81dcb08cc1f5" (UID: "26e94503-c561-4fad-8d9d-81dcb08cc1f5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.294896 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95958cf1-0438-4b2e-bb7b-a5f46188b59b-config" (OuterVolumeSpecName: "config") pod "95958cf1-0438-4b2e-bb7b-a5f46188b59b" (UID: "95958cf1-0438-4b2e-bb7b-a5f46188b59b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.300705 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26e94503-c561-4fad-8d9d-81dcb08cc1f5-kube-api-access-8m8jf" (OuterVolumeSpecName: "kube-api-access-8m8jf") pod "26e94503-c561-4fad-8d9d-81dcb08cc1f5" (UID: "26e94503-c561-4fad-8d9d-81dcb08cc1f5"). InnerVolumeSpecName "kube-api-access-8m8jf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.303161 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95958cf1-0438-4b2e-bb7b-a5f46188b59b-kube-api-access-ngrk8" (OuterVolumeSpecName: "kube-api-access-ngrk8") pod "95958cf1-0438-4b2e-bb7b-a5f46188b59b" (UID: "95958cf1-0438-4b2e-bb7b-a5f46188b59b"). InnerVolumeSpecName "kube-api-access-ngrk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.356280 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-xwnrv"] Oct 02 01:58:35 crc kubenswrapper[4775]: W1002 01:58:35.358483 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a54aa55_e3e2_4844_b175_e68442220b92.slice/crio-82de9cf13ef9c6d4a716e8c4172b37431c0748f8bd4e24b9c52ea8a28d17c6be WatchSource:0}: Error finding container 82de9cf13ef9c6d4a716e8c4172b37431c0748f8bd4e24b9c52ea8a28d17c6be: Status 404 returned error can't find the container with id 82de9cf13ef9c6d4a716e8c4172b37431c0748f8bd4e24b9c52ea8a28d17c6be Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.395787 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26e94503-c561-4fad-8d9d-81dcb08cc1f5-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.395834 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/26e94503-c561-4fad-8d9d-81dcb08cc1f5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.395846 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngrk8\" (UniqueName: \"kubernetes.io/projected/95958cf1-0438-4b2e-bb7b-a5f46188b59b-kube-api-access-ngrk8\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.395856 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/95958cf1-0438-4b2e-bb7b-a5f46188b59b-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.395865 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8m8jf\" (UniqueName: \"kubernetes.io/projected/26e94503-c561-4fad-8d9d-81dcb08cc1f5-kube-api-access-8m8jf\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.770844 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.777525 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jhkkr" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.782816 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" event={"ID":"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1","Type":"ContainerStarted","Data":"ba2e6c57bd61b6897927e1f72d663b5d16f31b40e42ba6b46a357640082e53f6"} Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.782871 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-rsq8b" event={"ID":"26e94503-c561-4fad-8d9d-81dcb08cc1f5","Type":"ContainerDied","Data":"e138f151de84ef99d422e5e3e1a938702212af731197cf747628069f15ce0e51"} Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.782914 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" event={"ID":"0af3416a-b960-4ac6-871c-26c4f0d58368","Type":"ContainerStarted","Data":"14f9f1517ab73165d8842cdd1989adb91d17d3444e56c926012b8e8a526da7b5"} Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.782940 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.782990 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f7685ff7-607b-416c-9382-2c00edecaa54","Type":"ContainerStarted","Data":"e41ea836798b128f59bc00526da81b9cd7e04b857cb248a53de519eb40c2a59e"} Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.783012 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-jhkkr" event={"ID":"95958cf1-0438-4b2e-bb7b-a5f46188b59b","Type":"ContainerDied","Data":"0909db59bf30b5abf14237919ca85523a034700ba024abaaebbca29f2e846d8d"} Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.783032 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xwnrv" event={"ID":"2a54aa55-e3e2-4844-b175-e68442220b92","Type":"ContainerStarted","Data":"82de9cf13ef9c6d4a716e8c4172b37431c0748f8bd4e24b9c52ea8a28d17c6be"} Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.783054 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.804719 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" podStartSLOduration=2.940125513 podStartE2EDuration="16.804702988s" podCreationTimestamp="2025-10-02 01:58:19 +0000 UTC" firstStartedPulling="2025-10-02 01:58:20.250154144 +0000 UTC m=+1037.416898184" lastFinishedPulling="2025-10-02 01:58:34.114731619 +0000 UTC m=+1051.281475659" observedRunningTime="2025-10-02 01:58:35.792634445 +0000 UTC m=+1052.959378485" watchObservedRunningTime="2025-10-02 01:58:35.804702988 +0000 UTC m=+1052.971447028" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.821902 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" podStartSLOduration=3.306411381 podStartE2EDuration="16.82188368s" podCreationTimestamp="2025-10-02 01:58:19 +0000 UTC" firstStartedPulling="2025-10-02 01:58:20.526663705 +0000 UTC m=+1037.693407745" lastFinishedPulling="2025-10-02 01:58:34.042136004 +0000 UTC m=+1051.208880044" observedRunningTime="2025-10-02 01:58:35.812236798 +0000 UTC m=+1052.978980848" watchObservedRunningTime="2025-10-02 01:58:35.82188368 +0000 UTC m=+1052.988627720" Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.868160 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jhkkr"] Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.879648 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jhkkr"] Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.893864 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rsq8b"] Oct 02 01:58:35 crc kubenswrapper[4775]: I1002 01:58:35.899839 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rsq8b"] Oct 02 01:58:37 crc kubenswrapper[4775]: I1002 01:58:37.781333 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26e94503-c561-4fad-8d9d-81dcb08cc1f5" path="/var/lib/kubelet/pods/26e94503-c561-4fad-8d9d-81dcb08cc1f5/volumes" Oct 02 01:58:37 crc kubenswrapper[4775]: I1002 01:58:37.782442 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95958cf1-0438-4b2e-bb7b-a5f46188b59b" path="/var/lib/kubelet/pods/95958cf1-0438-4b2e-bb7b-a5f46188b59b/volumes" Oct 02 01:58:40 crc kubenswrapper[4775]: I1002 01:58:40.016090 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:58:40 crc kubenswrapper[4775]: I1002 01:58:40.093784 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nb2rc"] Oct 02 01:58:40 crc kubenswrapper[4775]: I1002 01:58:40.094029 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" podUID="06528afb-0cc3-45fc-87d1-e2b0dc88f8e1" containerName="dnsmasq-dns" containerID="cri-o://ba2e6c57bd61b6897927e1f72d663b5d16f31b40e42ba6b46a357640082e53f6" gracePeriod=10 Oct 02 01:58:40 crc kubenswrapper[4775]: I1002 01:58:40.099328 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:40 crc kubenswrapper[4775]: I1002 01:58:40.818780 4775 generic.go:334] "Generic (PLEG): container finished" podID="06528afb-0cc3-45fc-87d1-e2b0dc88f8e1" containerID="ba2e6c57bd61b6897927e1f72d663b5d16f31b40e42ba6b46a357640082e53f6" exitCode=0 Oct 02 01:58:40 crc kubenswrapper[4775]: I1002 01:58:40.818821 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" event={"ID":"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1","Type":"ContainerDied","Data":"ba2e6c57bd61b6897927e1f72d663b5d16f31b40e42ba6b46a357640082e53f6"} Oct 02 01:58:44 crc kubenswrapper[4775]: I1002 01:58:44.815095 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:44 crc kubenswrapper[4775]: I1002 01:58:44.881508 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" event={"ID":"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1","Type":"ContainerDied","Data":"81b09fdae5a25645adbf8d337b628e78810306c6fdbd2bb0e8fd5e613c81c535"} Oct 02 01:58:44 crc kubenswrapper[4775]: I1002 01:58:44.881556 4775 scope.go:117] "RemoveContainer" containerID="ba2e6c57bd61b6897927e1f72d663b5d16f31b40e42ba6b46a357640082e53f6" Oct 02 01:58:44 crc kubenswrapper[4775]: I1002 01:58:44.881670 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" Oct 02 01:58:44 crc kubenswrapper[4775]: I1002 01:58:44.989837 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwwth\" (UniqueName: \"kubernetes.io/projected/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-kube-api-access-zwwth\") pod \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\" (UID: \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\") " Oct 02 01:58:44 crc kubenswrapper[4775]: I1002 01:58:44.989954 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-dns-svc\") pod \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\" (UID: \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\") " Oct 02 01:58:44 crc kubenswrapper[4775]: I1002 01:58:44.990151 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-config\") pod \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\" (UID: \"06528afb-0cc3-45fc-87d1-e2b0dc88f8e1\") " Oct 02 01:58:44 crc kubenswrapper[4775]: I1002 01:58:44.997061 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-kube-api-access-zwwth" (OuterVolumeSpecName: "kube-api-access-zwwth") pod "06528afb-0cc3-45fc-87d1-e2b0dc88f8e1" (UID: "06528afb-0cc3-45fc-87d1-e2b0dc88f8e1"). InnerVolumeSpecName "kube-api-access-zwwth". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:58:45 crc kubenswrapper[4775]: I1002 01:58:45.037603 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-config" (OuterVolumeSpecName: "config") pod "06528afb-0cc3-45fc-87d1-e2b0dc88f8e1" (UID: "06528afb-0cc3-45fc-87d1-e2b0dc88f8e1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:45 crc kubenswrapper[4775]: I1002 01:58:45.039229 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "06528afb-0cc3-45fc-87d1-e2b0dc88f8e1" (UID: "06528afb-0cc3-45fc-87d1-e2b0dc88f8e1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:45 crc kubenswrapper[4775]: I1002 01:58:45.091787 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:45 crc kubenswrapper[4775]: I1002 01:58:45.091820 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:45 crc kubenswrapper[4775]: I1002 01:58:45.091833 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwwth\" (UniqueName: \"kubernetes.io/projected/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1-kube-api-access-zwwth\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:45 crc kubenswrapper[4775]: I1002 01:58:45.223048 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nb2rc"] Oct 02 01:58:45 crc kubenswrapper[4775]: I1002 01:58:45.234787 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-nb2rc"] Oct 02 01:58:45 crc kubenswrapper[4775]: I1002 01:58:45.778291 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06528afb-0cc3-45fc-87d1-e2b0dc88f8e1" path="/var/lib/kubelet/pods/06528afb-0cc3-45fc-87d1-e2b0dc88f8e1/volumes" Oct 02 01:58:47 crc kubenswrapper[4775]: I1002 01:58:47.016701 4775 scope.go:117] "RemoveContainer" containerID="ee162fedc8decca1c1e7552c19b226217fed8f4e18a3303849d5bf3fca181269" Oct 02 01:58:48 crc kubenswrapper[4775]: I1002 01:58:48.927487 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3adc5647-1145-4c44-9ecc-66d1d9a19023","Type":"ContainerStarted","Data":"42d91714451f6055dca4d72c0b7726b16d6c992e3e11e19af8e604d77f38a639"} Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.689172 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-666b6646f7-nb2rc" podUID="06528afb-0cc3-45fc-87d1-e2b0dc88f8e1" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.101:5353: i/o timeout" Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.942762 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jdbs4" event={"ID":"6efbe4b3-953d-4927-95c6-9b92708f51eb","Type":"ContainerStarted","Data":"324549bfccf9b231fa568c7775d9ff418271c352066eb77651c4318e12b8d121"} Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.942908 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-jdbs4" Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.944428 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6638baa9-b367-40d1-8111-673dca6434ad","Type":"ContainerStarted","Data":"21c7d213dfbd2ef3e9ab592c212ff69020698f6ba49a9ebb52c79a43aa64f473"} Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.950386 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9","Type":"ContainerStarted","Data":"ab0021968685a2c8011cf5e84c4a1e4f33be77976a9063ea49b60d6569a853a3"} Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.953762 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"406dd24f-d800-44be-b33d-019c532c2feb","Type":"ContainerStarted","Data":"062b09ca95edb1ddddcb2de55757014ea328d800d380273e8fb5c4e3bb488c11"} Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.953896 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.955156 4775 generic.go:334] "Generic (PLEG): container finished" podID="2a54aa55-e3e2-4844-b175-e68442220b92" containerID="dff1a870836ccb85517741d710815364b5071bb66fd4210ba110dfd834a3ced9" exitCode=0 Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.955240 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xwnrv" event={"ID":"2a54aa55-e3e2-4844-b175-e68442220b92","Type":"ContainerDied","Data":"dff1a870836ccb85517741d710815364b5071bb66fd4210ba110dfd834a3ced9"} Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.956870 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"002eb223-af34-4c80-9570-894add9b0e3c","Type":"ContainerStarted","Data":"2563efa50bd9d9570bf76e9b0e57b1c4fa37d48c0fae0b8b0e1744feefe65559"} Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.960572 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-jdbs4" podStartSLOduration=8.11205876 podStartE2EDuration="20.960551734s" podCreationTimestamp="2025-10-02 01:58:29 +0000 UTC" firstStartedPulling="2025-10-02 01:58:34.608660104 +0000 UTC m=+1051.775404144" lastFinishedPulling="2025-10-02 01:58:47.457153048 +0000 UTC m=+1064.623897118" observedRunningTime="2025-10-02 01:58:49.959273032 +0000 UTC m=+1067.126017072" watchObservedRunningTime="2025-10-02 01:58:49.960551734 +0000 UTC m=+1067.127295804" Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.961253 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92","Type":"ContainerStarted","Data":"0089ce10498396392f25a55b07f3d03f418bbb5e5d51703da8bdb90f1f8cc858"} Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.968020 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6f55acc8-9fa8-496b-a882-d86ea3f28730","Type":"ContainerStarted","Data":"991d076b8c7fdaa82481667a7c1adc27601fb8bff13b965c51aee091cf3511ff"} Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.968136 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 01:58:49 crc kubenswrapper[4775]: I1002 01:58:49.970263 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f7685ff7-607b-416c-9382-2c00edecaa54","Type":"ContainerStarted","Data":"669d1080c1ef87bdc5aa946628176e085a6c1ef96db7059d070d3a1cdd097266"} Oct 02 01:58:50 crc kubenswrapper[4775]: I1002 01:58:50.080072 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.470655675 podStartE2EDuration="27.080052388s" podCreationTimestamp="2025-10-02 01:58:23 +0000 UTC" firstStartedPulling="2025-10-02 01:58:34.607371672 +0000 UTC m=+1051.774115712" lastFinishedPulling="2025-10-02 01:58:47.216768375 +0000 UTC m=+1064.383512425" observedRunningTime="2025-10-02 01:58:50.073499103 +0000 UTC m=+1067.240243143" watchObservedRunningTime="2025-10-02 01:58:50.080052388 +0000 UTC m=+1067.246796448" Oct 02 01:58:50 crc kubenswrapper[4775]: I1002 01:58:50.096150 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.359076543 podStartE2EDuration="25.096126412s" podCreationTimestamp="2025-10-02 01:58:25 +0000 UTC" firstStartedPulling="2025-10-02 01:58:34.413771415 +0000 UTC m=+1051.580515465" lastFinishedPulling="2025-10-02 01:58:48.150821284 +0000 UTC m=+1065.317565334" observedRunningTime="2025-10-02 01:58:50.089488855 +0000 UTC m=+1067.256232895" watchObservedRunningTime="2025-10-02 01:58:50.096126412 +0000 UTC m=+1067.262870452" Oct 02 01:58:50 crc kubenswrapper[4775]: I1002 01:58:50.996205 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xwnrv" event={"ID":"2a54aa55-e3e2-4844-b175-e68442220b92","Type":"ContainerStarted","Data":"2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303"} Oct 02 01:58:50 crc kubenswrapper[4775]: I1002 01:58:50.997214 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xwnrv" event={"ID":"2a54aa55-e3e2-4844-b175-e68442220b92","Type":"ContainerStarted","Data":"19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c"} Oct 02 01:58:51 crc kubenswrapper[4775]: I1002 01:58:51.017500 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-xwnrv" podStartSLOduration=10.682613994 podStartE2EDuration="22.017481931s" podCreationTimestamp="2025-10-02 01:58:29 +0000 UTC" firstStartedPulling="2025-10-02 01:58:35.360234296 +0000 UTC m=+1052.526978336" lastFinishedPulling="2025-10-02 01:58:46.695102233 +0000 UTC m=+1063.861846273" observedRunningTime="2025-10-02 01:58:51.01663056 +0000 UTC m=+1068.183374600" watchObservedRunningTime="2025-10-02 01:58:51.017481931 +0000 UTC m=+1068.184225971" Oct 02 01:58:52 crc kubenswrapper[4775]: I1002 01:58:52.008937 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:52 crc kubenswrapper[4775]: I1002 01:58:52.008990 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.022109 4775 generic.go:334] "Generic (PLEG): container finished" podID="3adc5647-1145-4c44-9ecc-66d1d9a19023" containerID="42d91714451f6055dca4d72c0b7726b16d6c992e3e11e19af8e604d77f38a639" exitCode=0 Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.022180 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3adc5647-1145-4c44-9ecc-66d1d9a19023","Type":"ContainerDied","Data":"42d91714451f6055dca4d72c0b7726b16d6c992e3e11e19af8e604d77f38a639"} Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.025285 4775 generic.go:334] "Generic (PLEG): container finished" podID="002eb223-af34-4c80-9570-894add9b0e3c" containerID="2563efa50bd9d9570bf76e9b0e57b1c4fa37d48c0fae0b8b0e1744feefe65559" exitCode=0 Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.025293 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"002eb223-af34-4c80-9570-894add9b0e3c","Type":"ContainerDied","Data":"2563efa50bd9d9570bf76e9b0e57b1c4fa37d48c0fae0b8b0e1744feefe65559"} Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.029797 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92","Type":"ContainerStarted","Data":"b8d49f290ce08384af0005110e4c784aaece1c7d6ad52d6aaa1e3385f7367c8c"} Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.035417 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f7685ff7-607b-416c-9382-2c00edecaa54","Type":"ContainerStarted","Data":"c0f5e7ac59fadc846940a5bf746aaf3da2e4edb2f52a02d0bbdc57d8847e6271"} Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.105676 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=7.147456537 podStartE2EDuration="25.10565301s" podCreationTimestamp="2025-10-02 01:58:28 +0000 UTC" firstStartedPulling="2025-10-02 01:58:34.53731305 +0000 UTC m=+1051.704057090" lastFinishedPulling="2025-10-02 01:58:52.495509523 +0000 UTC m=+1069.662253563" observedRunningTime="2025-10-02 01:58:53.096859209 +0000 UTC m=+1070.263603269" watchObservedRunningTime="2025-10-02 01:58:53.10565301 +0000 UTC m=+1070.272397070" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.134327 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.715723832 podStartE2EDuration="21.13430588s" podCreationTimestamp="2025-10-02 01:58:32 +0000 UTC" firstStartedPulling="2025-10-02 01:58:35.100971079 +0000 UTC m=+1052.267715109" lastFinishedPulling="2025-10-02 01:58:52.519553107 +0000 UTC m=+1069.686297157" observedRunningTime="2025-10-02 01:58:53.130898645 +0000 UTC m=+1070.297642725" watchObservedRunningTime="2025-10-02 01:58:53.13430588 +0000 UTC m=+1070.301049930" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.245482 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-7szxv"] Oct 02 01:58:53 crc kubenswrapper[4775]: E1002 01:58:53.246881 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06528afb-0cc3-45fc-87d1-e2b0dc88f8e1" containerName="init" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.246911 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="06528afb-0cc3-45fc-87d1-e2b0dc88f8e1" containerName="init" Oct 02 01:58:53 crc kubenswrapper[4775]: E1002 01:58:53.246990 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06528afb-0cc3-45fc-87d1-e2b0dc88f8e1" containerName="dnsmasq-dns" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.247006 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="06528afb-0cc3-45fc-87d1-e2b0dc88f8e1" containerName="dnsmasq-dns" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.247368 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="06528afb-0cc3-45fc-87d1-e2b0dc88f8e1" containerName="dnsmasq-dns" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.248527 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.251927 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.262843 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7szxv"] Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.324177 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-pxgxh"] Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.325495 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.329451 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.339648 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-pxgxh"] Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.355725 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9813c97-677a-42c2-b526-576a4c5f2968-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.355787 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e9813c97-677a-42c2-b526-576a4c5f2968-ovn-rundir\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.355809 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-pxgxh\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.355834 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9813c97-677a-42c2-b526-576a4c5f2968-config\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.355855 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e9813c97-677a-42c2-b526-576a4c5f2968-ovs-rundir\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.355882 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97bcs\" (UniqueName: \"kubernetes.io/projected/e9813c97-677a-42c2-b526-576a4c5f2968-kube-api-access-97bcs\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.355911 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-config\") pod \"dnsmasq-dns-7fd796d7df-pxgxh\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.355926 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-pxgxh\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.355966 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9813c97-677a-42c2-b526-576a4c5f2968-combined-ca-bundle\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.356011 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mls5w\" (UniqueName: \"kubernetes.io/projected/63e32532-7b52-4319-8a7d-8ad3d928d6fb-kube-api-access-mls5w\") pod \"dnsmasq-dns-7fd796d7df-pxgxh\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.405749 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-pxgxh"] Oct 02 01:58:53 crc kubenswrapper[4775]: E1002 01:58:53.406715 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-mls5w ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" podUID="63e32532-7b52-4319-8a7d-8ad3d928d6fb" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.429091 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-gclh6"] Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.430276 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.432026 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.444614 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-gclh6"] Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457233 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mls5w\" (UniqueName: \"kubernetes.io/projected/63e32532-7b52-4319-8a7d-8ad3d928d6fb-kube-api-access-mls5w\") pod \"dnsmasq-dns-7fd796d7df-pxgxh\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457310 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9813c97-677a-42c2-b526-576a4c5f2968-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457346 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e9813c97-677a-42c2-b526-576a4c5f2968-ovn-rundir\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457369 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-pxgxh\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457398 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9813c97-677a-42c2-b526-576a4c5f2968-config\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457424 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457452 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457479 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e9813c97-677a-42c2-b526-576a4c5f2968-ovs-rundir\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457521 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97bcs\" (UniqueName: \"kubernetes.io/projected/e9813c97-677a-42c2-b526-576a4c5f2968-kube-api-access-97bcs\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457543 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-config\") pod \"dnsmasq-dns-7fd796d7df-pxgxh\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457566 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-pxgxh\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457599 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457627 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9813c97-677a-42c2-b526-576a4c5f2968-combined-ca-bundle\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457667 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-config\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.457701 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rfqg\" (UniqueName: \"kubernetes.io/projected/18609394-0332-4e66-886d-6cc6e835e104-kube-api-access-8rfqg\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.458141 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e9813c97-677a-42c2-b526-576a4c5f2968-ovs-rundir\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.458380 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9813c97-677a-42c2-b526-576a4c5f2968-config\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.458728 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e9813c97-677a-42c2-b526-576a4c5f2968-ovn-rundir\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.460811 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-pxgxh\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.461122 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-pxgxh\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.462787 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9813c97-677a-42c2-b526-576a4c5f2968-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.463573 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9813c97-677a-42c2-b526-576a4c5f2968-combined-ca-bundle\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.464576 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-config\") pod \"dnsmasq-dns-7fd796d7df-pxgxh\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.472469 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mls5w\" (UniqueName: \"kubernetes.io/projected/63e32532-7b52-4319-8a7d-8ad3d928d6fb-kube-api-access-mls5w\") pod \"dnsmasq-dns-7fd796d7df-pxgxh\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.474088 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97bcs\" (UniqueName: \"kubernetes.io/projected/e9813c97-677a-42c2-b526-576a4c5f2968-kube-api-access-97bcs\") pod \"ovn-controller-metrics-7szxv\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.559049 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.559100 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-config\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.559126 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rfqg\" (UniqueName: \"kubernetes.io/projected/18609394-0332-4e66-886d-6cc6e835e104-kube-api-access-8rfqg\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.559191 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.559208 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.559992 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-config\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.560039 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.560877 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.563563 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.564971 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.575440 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rfqg\" (UniqueName: \"kubernetes.io/projected/18609394-0332-4e66-886d-6cc6e835e104-kube-api-access-8rfqg\") pod \"dnsmasq-dns-86db49b7ff-gclh6\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.608243 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7szxv" Oct 02 01:58:53 crc kubenswrapper[4775]: I1002 01:58:53.743648 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.047816 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3adc5647-1145-4c44-9ecc-66d1d9a19023","Type":"ContainerStarted","Data":"1209a9ffa5fb6c6952b1b8effd2ebc2e2fb407ab840357e9b3b1f508504a915a"} Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.053770 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.053779 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"002eb223-af34-4c80-9570-894add9b0e3c","Type":"ContainerStarted","Data":"8569575c44b11cca45a36d7a16be245143dd1895ca868b19ec6f5372e6eb3a4d"} Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.064289 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7szxv"] Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.064618 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:54 crc kubenswrapper[4775]: W1002 01:58:54.068363 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9813c97_677a_42c2_b526_576a4c5f2968.slice/crio-824fea2e009c65cbc6e17e23f3849349ab3e9f4be398ec804d6d7dec322e2ce1 WatchSource:0}: Error finding container 824fea2e009c65cbc6e17e23f3849349ab3e9f4be398ec804d6d7dec322e2ce1: Status 404 returned error can't find the container with id 824fea2e009c65cbc6e17e23f3849349ab3e9f4be398ec804d6d7dec322e2ce1 Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.077711 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=19.314448392 podStartE2EDuration="32.077690533s" podCreationTimestamp="2025-10-02 01:58:22 +0000 UTC" firstStartedPulling="2025-10-02 01:58:34.453459743 +0000 UTC m=+1051.620203783" lastFinishedPulling="2025-10-02 01:58:47.216701874 +0000 UTC m=+1064.383445924" observedRunningTime="2025-10-02 01:58:54.073019336 +0000 UTC m=+1071.239763416" watchObservedRunningTime="2025-10-02 01:58:54.077690533 +0000 UTC m=+1071.244434573" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.105182 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=20.512786678 podStartE2EDuration="33.105123883s" podCreationTimestamp="2025-10-02 01:58:21 +0000 UTC" firstStartedPulling="2025-10-02 01:58:34.424285649 +0000 UTC m=+1051.591029689" lastFinishedPulling="2025-10-02 01:58:47.016622834 +0000 UTC m=+1064.183366894" observedRunningTime="2025-10-02 01:58:54.104646041 +0000 UTC m=+1071.271390101" watchObservedRunningTime="2025-10-02 01:58:54.105123883 +0000 UTC m=+1071.271867983" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.116297 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.167306 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mls5w\" (UniqueName: \"kubernetes.io/projected/63e32532-7b52-4319-8a7d-8ad3d928d6fb-kube-api-access-mls5w\") pod \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.167562 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-ovsdbserver-nb\") pod \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.167628 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-dns-svc\") pod \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.167658 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-config\") pod \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\" (UID: \"63e32532-7b52-4319-8a7d-8ad3d928d6fb\") " Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.169214 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "63e32532-7b52-4319-8a7d-8ad3d928d6fb" (UID: "63e32532-7b52-4319-8a7d-8ad3d928d6fb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.169457 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-config" (OuterVolumeSpecName: "config") pod "63e32532-7b52-4319-8a7d-8ad3d928d6fb" (UID: "63e32532-7b52-4319-8a7d-8ad3d928d6fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.169602 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "63e32532-7b52-4319-8a7d-8ad3d928d6fb" (UID: "63e32532-7b52-4319-8a7d-8ad3d928d6fb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.176298 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63e32532-7b52-4319-8a7d-8ad3d928d6fb-kube-api-access-mls5w" (OuterVolumeSpecName: "kube-api-access-mls5w") pod "63e32532-7b52-4319-8a7d-8ad3d928d6fb" (UID: "63e32532-7b52-4319-8a7d-8ad3d928d6fb"). InnerVolumeSpecName "kube-api-access-mls5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.184434 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-gclh6"] Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.270726 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.270750 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.270760 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mls5w\" (UniqueName: \"kubernetes.io/projected/63e32532-7b52-4319-8a7d-8ad3d928d6fb-kube-api-access-mls5w\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.270768 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e32532-7b52-4319-8a7d-8ad3d928d6fb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.425482 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.465984 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.565462 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:54 crc kubenswrapper[4775]: I1002 01:58:54.615439 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.082427 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7szxv" event={"ID":"e9813c97-677a-42c2-b526-576a4c5f2968","Type":"ContainerStarted","Data":"47cbd1a6f145ee46322b61858213b665fd7c5043d5451925bf26dd59dec44238"} Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.082494 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7szxv" event={"ID":"e9813c97-677a-42c2-b526-576a4c5f2968","Type":"ContainerStarted","Data":"824fea2e009c65cbc6e17e23f3849349ab3e9f4be398ec804d6d7dec322e2ce1"} Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.085612 4775 generic.go:334] "Generic (PLEG): container finished" podID="18609394-0332-4e66-886d-6cc6e835e104" containerID="317373c4b1967593aa862614fcfe883198cbfcf9cfddd5b8c7eff73435ed1181" exitCode=0 Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.085838 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" event={"ID":"18609394-0332-4e66-886d-6cc6e835e104","Type":"ContainerDied","Data":"317373c4b1967593aa862614fcfe883198cbfcf9cfddd5b8c7eff73435ed1181"} Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.086040 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" event={"ID":"18609394-0332-4e66-886d-6cc6e835e104","Type":"ContainerStarted","Data":"b6bb140ee056144a0e6a84dcae8fa324be2425328babe29c5272760caa2bb710"} Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.086221 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-pxgxh" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.086991 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.117819 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-7szxv" podStartSLOduration=2.117800188 podStartE2EDuration="2.117800188s" podCreationTimestamp="2025-10-02 01:58:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:58:55.104877273 +0000 UTC m=+1072.271621323" watchObservedRunningTime="2025-10-02 01:58:55.117800188 +0000 UTC m=+1072.284544228" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.195461 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.201480 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.315230 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-pxgxh"] Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.319671 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-pxgxh"] Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.691124 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.692418 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.695970 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-zn9vk" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.696220 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.696395 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.696568 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.698042 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.775370 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63e32532-7b52-4319-8a7d-8ad3d928d6fb" path="/var/lib/kubelet/pods/63e32532-7b52-4319-8a7d-8ad3d928d6fb/volumes" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.795246 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7da01b78-a318-448e-8a4c-7ad3d9e833c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.795311 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7da01b78-a318-448e-8a4c-7ad3d9e833c0-scripts\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.795363 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da01b78-a318-448e-8a4c-7ad3d9e833c0-config\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.795406 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.795436 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.795516 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4n2g\" (UniqueName: \"kubernetes.io/projected/7da01b78-a318-448e-8a4c-7ad3d9e833c0-kube-api-access-q4n2g\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.795541 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.904540 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.904622 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7da01b78-a318-448e-8a4c-7ad3d9e833c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.904662 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7da01b78-a318-448e-8a4c-7ad3d9e833c0-scripts\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.904706 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da01b78-a318-448e-8a4c-7ad3d9e833c0-config\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.904741 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.904766 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.904903 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4n2g\" (UniqueName: \"kubernetes.io/projected/7da01b78-a318-448e-8a4c-7ad3d9e833c0-kube-api-access-q4n2g\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.905177 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7da01b78-a318-448e-8a4c-7ad3d9e833c0-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.906516 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7da01b78-a318-448e-8a4c-7ad3d9e833c0-scripts\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.907072 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da01b78-a318-448e-8a4c-7ad3d9e833c0-config\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.910553 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.911564 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.914909 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-gclh6"] Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.918031 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:55 crc kubenswrapper[4775]: I1002 01:58:55.939000 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4n2g\" (UniqueName: \"kubernetes.io/projected/7da01b78-a318-448e-8a4c-7ad3d9e833c0-kube-api-access-q4n2g\") pod \"ovn-northd-0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " pod="openstack/ovn-northd-0" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.008265 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-gtkrb"] Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.009718 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.015158 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.036323 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-gtkrb"] Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.053599 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.108705 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.108766 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.108810 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktlnv\" (UniqueName: \"kubernetes.io/projected/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-kube-api-access-ktlnv\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.108853 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-config\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.108872 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-dns-svc\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.125994 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" event={"ID":"18609394-0332-4e66-886d-6cc6e835e104","Type":"ContainerStarted","Data":"cf7fdb8ebad51d7a40477f7bb1d1fe90a17924b90f201ba9ce428791d69ad822"} Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.126463 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.149401 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" podStartSLOduration=3.149382498 podStartE2EDuration="3.149382498s" podCreationTimestamp="2025-10-02 01:58:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:58:56.142439003 +0000 UTC m=+1073.309183043" watchObservedRunningTime="2025-10-02 01:58:56.149382498 +0000 UTC m=+1073.316126538" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.211851 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-dns-svc\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.212075 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.212179 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.212365 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktlnv\" (UniqueName: \"kubernetes.io/projected/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-kube-api-access-ktlnv\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.212588 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-config\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.215171 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-dns-svc\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.217859 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.218293 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.218580 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-config\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.238764 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktlnv\" (UniqueName: \"kubernetes.io/projected/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-kube-api-access-ktlnv\") pod \"dnsmasq-dns-698758b865-gtkrb\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.334116 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.486000 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 01:58:56 crc kubenswrapper[4775]: I1002 01:58:56.764177 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-gtkrb"] Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.080350 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.085176 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.086900 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.086904 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-2f4zz" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.089358 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.089931 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.096624 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.137039 4775 generic.go:334] "Generic (PLEG): container finished" podID="ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a" containerID="dd5ab294818540df79b84f0aa04d8b04e35f763a6b0da0ffe3d4f5a627578747" exitCode=0 Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.137096 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-gtkrb" event={"ID":"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a","Type":"ContainerDied","Data":"dd5ab294818540df79b84f0aa04d8b04e35f763a6b0da0ffe3d4f5a627578747"} Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.137121 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-gtkrb" event={"ID":"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a","Type":"ContainerStarted","Data":"bf5ecb046aee777d82ce592f674e794a2acf10d7d3ca5c2657778b43fe8c9011"} Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.139675 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7da01b78-a318-448e-8a4c-7ad3d9e833c0","Type":"ContainerStarted","Data":"480c57752e934fae2adadd95f35d2b6dba80fd0e3972c13c154118141eeb3a0b"} Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.140190 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" podUID="18609394-0332-4e66-886d-6cc6e835e104" containerName="dnsmasq-dns" containerID="cri-o://cf7fdb8ebad51d7a40477f7bb1d1fe90a17924b90f201ba9ce428791d69ad822" gracePeriod=10 Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.225990 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.226151 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f8f38763-2981-4941-944d-b3fdc598be04-cache\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.226184 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.226211 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f8f38763-2981-4941-944d-b3fdc598be04-lock\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.226292 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27k7d\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-kube-api-access-27k7d\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.335096 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f8f38763-2981-4941-944d-b3fdc598be04-cache\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.335166 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.335197 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f8f38763-2981-4941-944d-b3fdc598be04-lock\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.335247 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27k7d\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-kube-api-access-27k7d\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.335475 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.336200 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f8f38763-2981-4941-944d-b3fdc598be04-cache\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.336562 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.337316 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f8f38763-2981-4941-944d-b3fdc598be04-lock\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: E1002 01:58:57.337480 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 01:58:57 crc kubenswrapper[4775]: E1002 01:58:57.337498 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 01:58:57 crc kubenswrapper[4775]: E1002 01:58:57.337546 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift podName:f8f38763-2981-4941-944d-b3fdc598be04 nodeName:}" failed. No retries permitted until 2025-10-02 01:58:57.837530674 +0000 UTC m=+1075.004274714 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift") pod "swift-storage-0" (UID: "f8f38763-2981-4941-944d-b3fdc598be04") : configmap "swift-ring-files" not found Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.353633 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27k7d\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-kube-api-access-27k7d\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.358334 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.507704 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-n65mg"] Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.508740 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.511890 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.511943 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.511996 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.515279 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-n65mg"] Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.535411 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:57 crc kubenswrapper[4775]: E1002 01:58:57.556183 4775 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.47:40408->38.102.83.47:45167: write tcp 38.102.83.47:40408->38.102.83.47:45167: write: broken pipe Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.646628 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rfqg\" (UniqueName: \"kubernetes.io/projected/18609394-0332-4e66-886d-6cc6e835e104-kube-api-access-8rfqg\") pod \"18609394-0332-4e66-886d-6cc6e835e104\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.647544 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-ovsdbserver-nb\") pod \"18609394-0332-4e66-886d-6cc6e835e104\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.647678 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-ovsdbserver-sb\") pod \"18609394-0332-4e66-886d-6cc6e835e104\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.647877 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-config\") pod \"18609394-0332-4e66-886d-6cc6e835e104\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.648126 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-dns-svc\") pod \"18609394-0332-4e66-886d-6cc6e835e104\" (UID: \"18609394-0332-4e66-886d-6cc6e835e104\") " Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.648748 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-combined-ca-bundle\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.649007 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb63c2b-eade-4e1d-a62f-24475df065e2-scripts\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.649068 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fdb63c2b-eade-4e1d-a62f-24475df065e2-ring-data-devices\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.649143 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwjgt\" (UniqueName: \"kubernetes.io/projected/fdb63c2b-eade-4e1d-a62f-24475df065e2-kube-api-access-pwjgt\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.649285 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-swiftconf\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.649446 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-dispersionconf\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.649501 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fdb63c2b-eade-4e1d-a62f-24475df065e2-etc-swift\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.653902 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18609394-0332-4e66-886d-6cc6e835e104-kube-api-access-8rfqg" (OuterVolumeSpecName: "kube-api-access-8rfqg") pod "18609394-0332-4e66-886d-6cc6e835e104" (UID: "18609394-0332-4e66-886d-6cc6e835e104"). InnerVolumeSpecName "kube-api-access-8rfqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.703045 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-config" (OuterVolumeSpecName: "config") pod "18609394-0332-4e66-886d-6cc6e835e104" (UID: "18609394-0332-4e66-886d-6cc6e835e104"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.703373 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "18609394-0332-4e66-886d-6cc6e835e104" (UID: "18609394-0332-4e66-886d-6cc6e835e104"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.714469 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "18609394-0332-4e66-886d-6cc6e835e104" (UID: "18609394-0332-4e66-886d-6cc6e835e104"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.715004 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "18609394-0332-4e66-886d-6cc6e835e104" (UID: "18609394-0332-4e66-886d-6cc6e835e104"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.750628 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb63c2b-eade-4e1d-a62f-24475df065e2-scripts\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.750682 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fdb63c2b-eade-4e1d-a62f-24475df065e2-ring-data-devices\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.750707 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwjgt\" (UniqueName: \"kubernetes.io/projected/fdb63c2b-eade-4e1d-a62f-24475df065e2-kube-api-access-pwjgt\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.750743 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-swiftconf\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.750766 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-dispersionconf\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.750783 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fdb63c2b-eade-4e1d-a62f-24475df065e2-etc-swift\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.750849 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-combined-ca-bundle\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.750895 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.750905 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.750914 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rfqg\" (UniqueName: \"kubernetes.io/projected/18609394-0332-4e66-886d-6cc6e835e104-kube-api-access-8rfqg\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.750924 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.750933 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/18609394-0332-4e66-886d-6cc6e835e104-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.751433 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb63c2b-eade-4e1d-a62f-24475df065e2-scripts\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.751583 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fdb63c2b-eade-4e1d-a62f-24475df065e2-ring-data-devices\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.752005 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fdb63c2b-eade-4e1d-a62f-24475df065e2-etc-swift\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.754054 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-dispersionconf\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.754216 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-combined-ca-bundle\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.754821 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-swiftconf\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.769267 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwjgt\" (UniqueName: \"kubernetes.io/projected/fdb63c2b-eade-4e1d-a62f-24475df065e2-kube-api-access-pwjgt\") pod \"swift-ring-rebalance-n65mg\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.834764 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:58:57 crc kubenswrapper[4775]: I1002 01:58:57.852203 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:57 crc kubenswrapper[4775]: E1002 01:58:57.852349 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 01:58:57 crc kubenswrapper[4775]: E1002 01:58:57.852370 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 01:58:57 crc kubenswrapper[4775]: E1002 01:58:57.852428 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift podName:f8f38763-2981-4941-944d-b3fdc598be04 nodeName:}" failed. No retries permitted until 2025-10-02 01:58:58.852409336 +0000 UTC m=+1076.019153376 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift") pod "swift-storage-0" (UID: "f8f38763-2981-4941-944d-b3fdc598be04") : configmap "swift-ring-files" not found Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.172877 4775 generic.go:334] "Generic (PLEG): container finished" podID="18609394-0332-4e66-886d-6cc6e835e104" containerID="cf7fdb8ebad51d7a40477f7bb1d1fe90a17924b90f201ba9ce428791d69ad822" exitCode=0 Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.172935 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.172966 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" event={"ID":"18609394-0332-4e66-886d-6cc6e835e104","Type":"ContainerDied","Data":"cf7fdb8ebad51d7a40477f7bb1d1fe90a17924b90f201ba9ce428791d69ad822"} Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.173433 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-gclh6" event={"ID":"18609394-0332-4e66-886d-6cc6e835e104","Type":"ContainerDied","Data":"b6bb140ee056144a0e6a84dcae8fa324be2425328babe29c5272760caa2bb710"} Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.173450 4775 scope.go:117] "RemoveContainer" containerID="cf7fdb8ebad51d7a40477f7bb1d1fe90a17924b90f201ba9ce428791d69ad822" Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.176770 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-gtkrb" event={"ID":"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a","Type":"ContainerStarted","Data":"cc7ff92e14efddf8b284f5a7099a1ce780c7a0e2fa56dfd1b699868a8fd238d8"} Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.177485 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.197506 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-gtkrb" podStartSLOduration=3.19749118 podStartE2EDuration="3.19749118s" podCreationTimestamp="2025-10-02 01:58:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:58:58.196437184 +0000 UTC m=+1075.363181224" watchObservedRunningTime="2025-10-02 01:58:58.19749118 +0000 UTC m=+1075.364235220" Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.201820 4775 scope.go:117] "RemoveContainer" containerID="317373c4b1967593aa862614fcfe883198cbfcf9cfddd5b8c7eff73435ed1181" Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.214358 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-gclh6"] Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.220260 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-gclh6"] Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.224490 4775 scope.go:117] "RemoveContainer" containerID="cf7fdb8ebad51d7a40477f7bb1d1fe90a17924b90f201ba9ce428791d69ad822" Oct 02 01:58:58 crc kubenswrapper[4775]: E1002 01:58:58.225915 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf7fdb8ebad51d7a40477f7bb1d1fe90a17924b90f201ba9ce428791d69ad822\": container with ID starting with cf7fdb8ebad51d7a40477f7bb1d1fe90a17924b90f201ba9ce428791d69ad822 not found: ID does not exist" containerID="cf7fdb8ebad51d7a40477f7bb1d1fe90a17924b90f201ba9ce428791d69ad822" Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.225984 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf7fdb8ebad51d7a40477f7bb1d1fe90a17924b90f201ba9ce428791d69ad822"} err="failed to get container status \"cf7fdb8ebad51d7a40477f7bb1d1fe90a17924b90f201ba9ce428791d69ad822\": rpc error: code = NotFound desc = could not find container \"cf7fdb8ebad51d7a40477f7bb1d1fe90a17924b90f201ba9ce428791d69ad822\": container with ID starting with cf7fdb8ebad51d7a40477f7bb1d1fe90a17924b90f201ba9ce428791d69ad822 not found: ID does not exist" Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.226025 4775 scope.go:117] "RemoveContainer" containerID="317373c4b1967593aa862614fcfe883198cbfcf9cfddd5b8c7eff73435ed1181" Oct 02 01:58:58 crc kubenswrapper[4775]: E1002 01:58:58.226369 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"317373c4b1967593aa862614fcfe883198cbfcf9cfddd5b8c7eff73435ed1181\": container with ID starting with 317373c4b1967593aa862614fcfe883198cbfcf9cfddd5b8c7eff73435ed1181 not found: ID does not exist" containerID="317373c4b1967593aa862614fcfe883198cbfcf9cfddd5b8c7eff73435ed1181" Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.226409 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"317373c4b1967593aa862614fcfe883198cbfcf9cfddd5b8c7eff73435ed1181"} err="failed to get container status \"317373c4b1967593aa862614fcfe883198cbfcf9cfddd5b8c7eff73435ed1181\": rpc error: code = NotFound desc = could not find container \"317373c4b1967593aa862614fcfe883198cbfcf9cfddd5b8c7eff73435ed1181\": container with ID starting with 317373c4b1967593aa862614fcfe883198cbfcf9cfddd5b8c7eff73435ed1181 not found: ID does not exist" Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.499582 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-n65mg"] Oct 02 01:58:58 crc kubenswrapper[4775]: W1002 01:58:58.507067 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdb63c2b_eade_4e1d_a62f_24475df065e2.slice/crio-c856d203cc28f9fe7aa2010c078524592b99e0906d8e2e77e1c7e1430de5da90 WatchSource:0}: Error finding container c856d203cc28f9fe7aa2010c078524592b99e0906d8e2e77e1c7e1430de5da90: Status 404 returned error can't find the container with id c856d203cc28f9fe7aa2010c078524592b99e0906d8e2e77e1c7e1430de5da90 Oct 02 01:58:58 crc kubenswrapper[4775]: I1002 01:58:58.878557 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:58:58 crc kubenswrapper[4775]: E1002 01:58:58.878815 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 01:58:58 crc kubenswrapper[4775]: E1002 01:58:58.878854 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 01:58:58 crc kubenswrapper[4775]: E1002 01:58:58.878920 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift podName:f8f38763-2981-4941-944d-b3fdc598be04 nodeName:}" failed. No retries permitted until 2025-10-02 01:59:00.878901238 +0000 UTC m=+1078.045645278 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift") pod "swift-storage-0" (UID: "f8f38763-2981-4941-944d-b3fdc598be04") : configmap "swift-ring-files" not found Oct 02 01:58:59 crc kubenswrapper[4775]: I1002 01:58:59.188631 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-n65mg" event={"ID":"fdb63c2b-eade-4e1d-a62f-24475df065e2","Type":"ContainerStarted","Data":"c856d203cc28f9fe7aa2010c078524592b99e0906d8e2e77e1c7e1430de5da90"} Oct 02 01:58:59 crc kubenswrapper[4775]: I1002 01:58:59.190633 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7da01b78-a318-448e-8a4c-7ad3d9e833c0","Type":"ContainerStarted","Data":"5b7338db5965862d0736735b7d766b4ed845f3e4cf8b4b81acd87174941ecfa6"} Oct 02 01:58:59 crc kubenswrapper[4775]: I1002 01:58:59.190665 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7da01b78-a318-448e-8a4c-7ad3d9e833c0","Type":"ContainerStarted","Data":"39f92c334c8a479d69a76f6ee477dc9e38b2f2fb13b731fddbffda7d7dac3511"} Oct 02 01:58:59 crc kubenswrapper[4775]: I1002 01:58:59.191677 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 02 01:58:59 crc kubenswrapper[4775]: I1002 01:58:59.232075 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.655656999 podStartE2EDuration="4.232055085s" podCreationTimestamp="2025-10-02 01:58:55 +0000 UTC" firstStartedPulling="2025-10-02 01:58:56.490110452 +0000 UTC m=+1073.656854492" lastFinishedPulling="2025-10-02 01:58:58.066508538 +0000 UTC m=+1075.233252578" observedRunningTime="2025-10-02 01:58:59.229276306 +0000 UTC m=+1076.396020346" watchObservedRunningTime="2025-10-02 01:58:59.232055085 +0000 UTC m=+1076.398799125" Oct 02 01:58:59 crc kubenswrapper[4775]: I1002 01:58:59.778280 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18609394-0332-4e66-886d-6cc6e835e104" path="/var/lib/kubelet/pods/18609394-0332-4e66-886d-6cc6e835e104/volumes" Oct 02 01:59:00 crc kubenswrapper[4775]: I1002 01:59:00.913067 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:59:00 crc kubenswrapper[4775]: E1002 01:59:00.913614 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 01:59:00 crc kubenswrapper[4775]: E1002 01:59:00.913639 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 01:59:00 crc kubenswrapper[4775]: E1002 01:59:00.913710 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift podName:f8f38763-2981-4941-944d-b3fdc598be04 nodeName:}" failed. No retries permitted until 2025-10-02 01:59:04.913685875 +0000 UTC m=+1082.080429955 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift") pod "swift-storage-0" (UID: "f8f38763-2981-4941-944d-b3fdc598be04") : configmap "swift-ring-files" not found Oct 02 01:59:03 crc kubenswrapper[4775]: I1002 01:59:03.244182 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-n65mg" event={"ID":"fdb63c2b-eade-4e1d-a62f-24475df065e2","Type":"ContainerStarted","Data":"7de9c008d839d22eacc4887e2eeff05b010de483778c5ec8d189d41a801d7a62"} Oct 02 01:59:03 crc kubenswrapper[4775]: I1002 01:59:03.252762 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 02 01:59:03 crc kubenswrapper[4775]: I1002 01:59:03.252835 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 02 01:59:03 crc kubenswrapper[4775]: I1002 01:59:03.268382 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-n65mg" podStartSLOduration=2.733660533 podStartE2EDuration="6.268357492s" podCreationTimestamp="2025-10-02 01:58:57 +0000 UTC" firstStartedPulling="2025-10-02 01:58:58.509559474 +0000 UTC m=+1075.676303514" lastFinishedPulling="2025-10-02 01:59:02.044256433 +0000 UTC m=+1079.211000473" observedRunningTime="2025-10-02 01:59:03.262916035 +0000 UTC m=+1080.429660115" watchObservedRunningTime="2025-10-02 01:59:03.268357492 +0000 UTC m=+1080.435101562" Oct 02 01:59:03 crc kubenswrapper[4775]: I1002 01:59:03.308112 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 02 01:59:03 crc kubenswrapper[4775]: I1002 01:59:03.620876 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 02 01:59:03 crc kubenswrapper[4775]: I1002 01:59:03.621485 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 02 01:59:03 crc kubenswrapper[4775]: I1002 01:59:03.682027 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 02 01:59:04 crc kubenswrapper[4775]: I1002 01:59:04.330113 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 02 01:59:04 crc kubenswrapper[4775]: I1002 01:59:04.333917 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 02 01:59:04 crc kubenswrapper[4775]: I1002 01:59:04.921172 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:59:04 crc kubenswrapper[4775]: E1002 01:59:04.921370 4775 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 02 01:59:04 crc kubenswrapper[4775]: E1002 01:59:04.921384 4775 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 02 01:59:04 crc kubenswrapper[4775]: E1002 01:59:04.921439 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift podName:f8f38763-2981-4941-944d-b3fdc598be04 nodeName:}" failed. No retries permitted until 2025-10-02 01:59:12.921422814 +0000 UTC m=+1090.088166854 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift") pod "swift-storage-0" (UID: "f8f38763-2981-4941-944d-b3fdc598be04") : configmap "swift-ring-files" not found Oct 02 01:59:06 crc kubenswrapper[4775]: I1002 01:59:06.336200 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:59:06 crc kubenswrapper[4775]: I1002 01:59:06.414357 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-b6f52"] Oct 02 01:59:06 crc kubenswrapper[4775]: I1002 01:59:06.415194 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" podUID="0af3416a-b960-4ac6-871c-26c4f0d58368" containerName="dnsmasq-dns" containerID="cri-o://14f9f1517ab73165d8842cdd1989adb91d17d3444e56c926012b8e8a526da7b5" gracePeriod=10 Oct 02 01:59:06 crc kubenswrapper[4775]: I1002 01:59:06.854297 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:59:06 crc kubenswrapper[4775]: I1002 01:59:06.958073 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0af3416a-b960-4ac6-871c-26c4f0d58368-dns-svc\") pod \"0af3416a-b960-4ac6-871c-26c4f0d58368\" (UID: \"0af3416a-b960-4ac6-871c-26c4f0d58368\") " Oct 02 01:59:06 crc kubenswrapper[4775]: I1002 01:59:06.958302 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dzl2c\" (UniqueName: \"kubernetes.io/projected/0af3416a-b960-4ac6-871c-26c4f0d58368-kube-api-access-dzl2c\") pod \"0af3416a-b960-4ac6-871c-26c4f0d58368\" (UID: \"0af3416a-b960-4ac6-871c-26c4f0d58368\") " Oct 02 01:59:06 crc kubenswrapper[4775]: I1002 01:59:06.958337 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0af3416a-b960-4ac6-871c-26c4f0d58368-config\") pod \"0af3416a-b960-4ac6-871c-26c4f0d58368\" (UID: \"0af3416a-b960-4ac6-871c-26c4f0d58368\") " Oct 02 01:59:06 crc kubenswrapper[4775]: I1002 01:59:06.965125 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0af3416a-b960-4ac6-871c-26c4f0d58368-kube-api-access-dzl2c" (OuterVolumeSpecName: "kube-api-access-dzl2c") pod "0af3416a-b960-4ac6-871c-26c4f0d58368" (UID: "0af3416a-b960-4ac6-871c-26c4f0d58368"). InnerVolumeSpecName "kube-api-access-dzl2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:06 crc kubenswrapper[4775]: I1002 01:59:06.996229 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0af3416a-b960-4ac6-871c-26c4f0d58368-config" (OuterVolumeSpecName: "config") pod "0af3416a-b960-4ac6-871c-26c4f0d58368" (UID: "0af3416a-b960-4ac6-871c-26c4f0d58368"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.004811 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0af3416a-b960-4ac6-871c-26c4f0d58368-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0af3416a-b960-4ac6-871c-26c4f0d58368" (UID: "0af3416a-b960-4ac6-871c-26c4f0d58368"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.061224 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dzl2c\" (UniqueName: \"kubernetes.io/projected/0af3416a-b960-4ac6-871c-26c4f0d58368-kube-api-access-dzl2c\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.061265 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0af3416a-b960-4ac6-871c-26c4f0d58368-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.061279 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0af3416a-b960-4ac6-871c-26c4f0d58368-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.234281 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.234335 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.290303 4775 generic.go:334] "Generic (PLEG): container finished" podID="0af3416a-b960-4ac6-871c-26c4f0d58368" containerID="14f9f1517ab73165d8842cdd1989adb91d17d3444e56c926012b8e8a526da7b5" exitCode=0 Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.290345 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" event={"ID":"0af3416a-b960-4ac6-871c-26c4f0d58368","Type":"ContainerDied","Data":"14f9f1517ab73165d8842cdd1989adb91d17d3444e56c926012b8e8a526da7b5"} Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.290401 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" event={"ID":"0af3416a-b960-4ac6-871c-26c4f0d58368","Type":"ContainerDied","Data":"70de01799ae7846bf3ad95e80b78859b7b0b254163fa61d7d936ee2112f5762b"} Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.290395 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-b6f52" Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.290474 4775 scope.go:117] "RemoveContainer" containerID="14f9f1517ab73165d8842cdd1989adb91d17d3444e56c926012b8e8a526da7b5" Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.328266 4775 scope.go:117] "RemoveContainer" containerID="4baa3fa4694b068174901ca15c9726a8b3eac40e59412c11c2ff60ca69a4901a" Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.332541 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-b6f52"] Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.339780 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-b6f52"] Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.365920 4775 scope.go:117] "RemoveContainer" containerID="14f9f1517ab73165d8842cdd1989adb91d17d3444e56c926012b8e8a526da7b5" Oct 02 01:59:07 crc kubenswrapper[4775]: E1002 01:59:07.366477 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14f9f1517ab73165d8842cdd1989adb91d17d3444e56c926012b8e8a526da7b5\": container with ID starting with 14f9f1517ab73165d8842cdd1989adb91d17d3444e56c926012b8e8a526da7b5 not found: ID does not exist" containerID="14f9f1517ab73165d8842cdd1989adb91d17d3444e56c926012b8e8a526da7b5" Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.366512 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14f9f1517ab73165d8842cdd1989adb91d17d3444e56c926012b8e8a526da7b5"} err="failed to get container status \"14f9f1517ab73165d8842cdd1989adb91d17d3444e56c926012b8e8a526da7b5\": rpc error: code = NotFound desc = could not find container \"14f9f1517ab73165d8842cdd1989adb91d17d3444e56c926012b8e8a526da7b5\": container with ID starting with 14f9f1517ab73165d8842cdd1989adb91d17d3444e56c926012b8e8a526da7b5 not found: ID does not exist" Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.366539 4775 scope.go:117] "RemoveContainer" containerID="4baa3fa4694b068174901ca15c9726a8b3eac40e59412c11c2ff60ca69a4901a" Oct 02 01:59:07 crc kubenswrapper[4775]: E1002 01:59:07.367175 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4baa3fa4694b068174901ca15c9726a8b3eac40e59412c11c2ff60ca69a4901a\": container with ID starting with 4baa3fa4694b068174901ca15c9726a8b3eac40e59412c11c2ff60ca69a4901a not found: ID does not exist" containerID="4baa3fa4694b068174901ca15c9726a8b3eac40e59412c11c2ff60ca69a4901a" Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.367218 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4baa3fa4694b068174901ca15c9726a8b3eac40e59412c11c2ff60ca69a4901a"} err="failed to get container status \"4baa3fa4694b068174901ca15c9726a8b3eac40e59412c11c2ff60ca69a4901a\": rpc error: code = NotFound desc = could not find container \"4baa3fa4694b068174901ca15c9726a8b3eac40e59412c11c2ff60ca69a4901a\": container with ID starting with 4baa3fa4694b068174901ca15c9726a8b3eac40e59412c11c2ff60ca69a4901a not found: ID does not exist" Oct 02 01:59:07 crc kubenswrapper[4775]: I1002 01:59:07.778066 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0af3416a-b960-4ac6-871c-26c4f0d58368" path="/var/lib/kubelet/pods/0af3416a-b960-4ac6-871c-26c4f0d58368/volumes" Oct 02 01:59:10 crc kubenswrapper[4775]: I1002 01:59:10.346095 4775 generic.go:334] "Generic (PLEG): container finished" podID="fdb63c2b-eade-4e1d-a62f-24475df065e2" containerID="7de9c008d839d22eacc4887e2eeff05b010de483778c5ec8d189d41a801d7a62" exitCode=0 Oct 02 01:59:10 crc kubenswrapper[4775]: I1002 01:59:10.346197 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-n65mg" event={"ID":"fdb63c2b-eade-4e1d-a62f-24475df065e2","Type":"ContainerDied","Data":"7de9c008d839d22eacc4887e2eeff05b010de483778c5ec8d189d41a801d7a62"} Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.115263 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.684568 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.761472 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fdb63c2b-eade-4e1d-a62f-24475df065e2-ring-data-devices\") pod \"fdb63c2b-eade-4e1d-a62f-24475df065e2\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.761604 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb63c2b-eade-4e1d-a62f-24475df065e2-scripts\") pod \"fdb63c2b-eade-4e1d-a62f-24475df065e2\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.761679 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-combined-ca-bundle\") pod \"fdb63c2b-eade-4e1d-a62f-24475df065e2\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.762231 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdb63c2b-eade-4e1d-a62f-24475df065e2-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "fdb63c2b-eade-4e1d-a62f-24475df065e2" (UID: "fdb63c2b-eade-4e1d-a62f-24475df065e2"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.761753 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-swiftconf\") pod \"fdb63c2b-eade-4e1d-a62f-24475df065e2\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.762680 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-dispersionconf\") pod \"fdb63c2b-eade-4e1d-a62f-24475df065e2\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.762718 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwjgt\" (UniqueName: \"kubernetes.io/projected/fdb63c2b-eade-4e1d-a62f-24475df065e2-kube-api-access-pwjgt\") pod \"fdb63c2b-eade-4e1d-a62f-24475df065e2\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.762806 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fdb63c2b-eade-4e1d-a62f-24475df065e2-etc-swift\") pod \"fdb63c2b-eade-4e1d-a62f-24475df065e2\" (UID: \"fdb63c2b-eade-4e1d-a62f-24475df065e2\") " Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.763412 4775 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/fdb63c2b-eade-4e1d-a62f-24475df065e2-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.764738 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdb63c2b-eade-4e1d-a62f-24475df065e2-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "fdb63c2b-eade-4e1d-a62f-24475df065e2" (UID: "fdb63c2b-eade-4e1d-a62f-24475df065e2"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.771661 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdb63c2b-eade-4e1d-a62f-24475df065e2-kube-api-access-pwjgt" (OuterVolumeSpecName: "kube-api-access-pwjgt") pod "fdb63c2b-eade-4e1d-a62f-24475df065e2" (UID: "fdb63c2b-eade-4e1d-a62f-24475df065e2"). InnerVolumeSpecName "kube-api-access-pwjgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.773250 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "fdb63c2b-eade-4e1d-a62f-24475df065e2" (UID: "fdb63c2b-eade-4e1d-a62f-24475df065e2"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.784750 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdb63c2b-eade-4e1d-a62f-24475df065e2-scripts" (OuterVolumeSpecName: "scripts") pod "fdb63c2b-eade-4e1d-a62f-24475df065e2" (UID: "fdb63c2b-eade-4e1d-a62f-24475df065e2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.787472 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fdb63c2b-eade-4e1d-a62f-24475df065e2" (UID: "fdb63c2b-eade-4e1d-a62f-24475df065e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.793259 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "fdb63c2b-eade-4e1d-a62f-24475df065e2" (UID: "fdb63c2b-eade-4e1d-a62f-24475df065e2"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.864612 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fdb63c2b-eade-4e1d-a62f-24475df065e2-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.864647 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.864661 4775 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.864673 4775 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/fdb63c2b-eade-4e1d-a62f-24475df065e2-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.864684 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwjgt\" (UniqueName: \"kubernetes.io/projected/fdb63c2b-eade-4e1d-a62f-24475df065e2-kube-api-access-pwjgt\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:11 crc kubenswrapper[4775]: I1002 01:59:11.864696 4775 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/fdb63c2b-eade-4e1d-a62f-24475df065e2-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:12 crc kubenswrapper[4775]: I1002 01:59:12.370567 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-n65mg" event={"ID":"fdb63c2b-eade-4e1d-a62f-24475df065e2","Type":"ContainerDied","Data":"c856d203cc28f9fe7aa2010c078524592b99e0906d8e2e77e1c7e1430de5da90"} Oct 02 01:59:12 crc kubenswrapper[4775]: I1002 01:59:12.370629 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c856d203cc28f9fe7aa2010c078524592b99e0906d8e2e77e1c7e1430de5da90" Oct 02 01:59:12 crc kubenswrapper[4775]: I1002 01:59:12.370636 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-n65mg" Oct 02 01:59:12 crc kubenswrapper[4775]: I1002 01:59:12.986823 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:59:12 crc kubenswrapper[4775]: I1002 01:59:12.996247 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift\") pod \"swift-storage-0\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " pod="openstack/swift-storage-0" Oct 02 01:59:13 crc kubenswrapper[4775]: I1002 01:59:13.016288 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 01:59:13 crc kubenswrapper[4775]: I1002 01:59:13.422154 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 02 01:59:13 crc kubenswrapper[4775]: I1002 01:59:13.806892 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-j4qvx"] Oct 02 01:59:13 crc kubenswrapper[4775]: E1002 01:59:13.807354 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18609394-0332-4e66-886d-6cc6e835e104" containerName="dnsmasq-dns" Oct 02 01:59:13 crc kubenswrapper[4775]: I1002 01:59:13.807383 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="18609394-0332-4e66-886d-6cc6e835e104" containerName="dnsmasq-dns" Oct 02 01:59:13 crc kubenswrapper[4775]: E1002 01:59:13.807403 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18609394-0332-4e66-886d-6cc6e835e104" containerName="init" Oct 02 01:59:13 crc kubenswrapper[4775]: I1002 01:59:13.807416 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="18609394-0332-4e66-886d-6cc6e835e104" containerName="init" Oct 02 01:59:13 crc kubenswrapper[4775]: E1002 01:59:13.807441 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb63c2b-eade-4e1d-a62f-24475df065e2" containerName="swift-ring-rebalance" Oct 02 01:59:13 crc kubenswrapper[4775]: I1002 01:59:13.807453 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb63c2b-eade-4e1d-a62f-24475df065e2" containerName="swift-ring-rebalance" Oct 02 01:59:13 crc kubenswrapper[4775]: E1002 01:59:13.807477 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0af3416a-b960-4ac6-871c-26c4f0d58368" containerName="init" Oct 02 01:59:13 crc kubenswrapper[4775]: I1002 01:59:13.807487 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0af3416a-b960-4ac6-871c-26c4f0d58368" containerName="init" Oct 02 01:59:13 crc kubenswrapper[4775]: E1002 01:59:13.807501 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0af3416a-b960-4ac6-871c-26c4f0d58368" containerName="dnsmasq-dns" Oct 02 01:59:13 crc kubenswrapper[4775]: I1002 01:59:13.807511 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0af3416a-b960-4ac6-871c-26c4f0d58368" containerName="dnsmasq-dns" Oct 02 01:59:13 crc kubenswrapper[4775]: I1002 01:59:13.807734 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0af3416a-b960-4ac6-871c-26c4f0d58368" containerName="dnsmasq-dns" Oct 02 01:59:13 crc kubenswrapper[4775]: I1002 01:59:13.807756 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="18609394-0332-4e66-886d-6cc6e835e104" containerName="dnsmasq-dns" Oct 02 01:59:13 crc kubenswrapper[4775]: I1002 01:59:13.807777 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdb63c2b-eade-4e1d-a62f-24475df065e2" containerName="swift-ring-rebalance" Oct 02 01:59:13 crc kubenswrapper[4775]: I1002 01:59:13.808610 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j4qvx" Oct 02 01:59:13 crc kubenswrapper[4775]: I1002 01:59:13.817518 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-j4qvx"] Oct 02 01:59:13 crc kubenswrapper[4775]: I1002 01:59:13.902578 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgd86\" (UniqueName: \"kubernetes.io/projected/8a95c042-cb96-42be-8192-2af9ad63d43d-kube-api-access-wgd86\") pod \"keystone-db-create-j4qvx\" (UID: \"8a95c042-cb96-42be-8192-2af9ad63d43d\") " pod="openstack/keystone-db-create-j4qvx" Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.004600 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgd86\" (UniqueName: \"kubernetes.io/projected/8a95c042-cb96-42be-8192-2af9ad63d43d-kube-api-access-wgd86\") pod \"keystone-db-create-j4qvx\" (UID: \"8a95c042-cb96-42be-8192-2af9ad63d43d\") " pod="openstack/keystone-db-create-j4qvx" Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.054313 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-582w8"] Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.055455 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-582w8" Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.058059 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgd86\" (UniqueName: \"kubernetes.io/projected/8a95c042-cb96-42be-8192-2af9ad63d43d-kube-api-access-wgd86\") pod \"keystone-db-create-j4qvx\" (UID: \"8a95c042-cb96-42be-8192-2af9ad63d43d\") " pod="openstack/keystone-db-create-j4qvx" Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.061322 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-582w8"] Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.135494 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j4qvx" Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.209634 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sk8r\" (UniqueName: \"kubernetes.io/projected/d01b3c76-8f88-4fc6-be13-4f8150a14949-kube-api-access-9sk8r\") pod \"placement-db-create-582w8\" (UID: \"d01b3c76-8f88-4fc6-be13-4f8150a14949\") " pod="openstack/placement-db-create-582w8" Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.299837 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-m576z"] Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.301393 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-m576z" Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.307611 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-m576z"] Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.311268 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sk8r\" (UniqueName: \"kubernetes.io/projected/d01b3c76-8f88-4fc6-be13-4f8150a14949-kube-api-access-9sk8r\") pod \"placement-db-create-582w8\" (UID: \"d01b3c76-8f88-4fc6-be13-4f8150a14949\") " pod="openstack/placement-db-create-582w8" Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.329706 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sk8r\" (UniqueName: \"kubernetes.io/projected/d01b3c76-8f88-4fc6-be13-4f8150a14949-kube-api-access-9sk8r\") pod \"placement-db-create-582w8\" (UID: \"d01b3c76-8f88-4fc6-be13-4f8150a14949\") " pod="openstack/placement-db-create-582w8" Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.387186 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"50c7ff7d8ff529121dc875f3de87c4eaed5a9d4222abbe043b5147a5e835b4b1"} Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.404556 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-582w8" Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.412674 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbgjg\" (UniqueName: \"kubernetes.io/projected/c2862399-bef9-4b91-9376-3ceff3377234-kube-api-access-jbgjg\") pod \"glance-db-create-m576z\" (UID: \"c2862399-bef9-4b91-9376-3ceff3377234\") " pod="openstack/glance-db-create-m576z" Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.514388 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbgjg\" (UniqueName: \"kubernetes.io/projected/c2862399-bef9-4b91-9376-3ceff3377234-kube-api-access-jbgjg\") pod \"glance-db-create-m576z\" (UID: \"c2862399-bef9-4b91-9376-3ceff3377234\") " pod="openstack/glance-db-create-m576z" Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.534226 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbgjg\" (UniqueName: \"kubernetes.io/projected/c2862399-bef9-4b91-9376-3ceff3377234-kube-api-access-jbgjg\") pod \"glance-db-create-m576z\" (UID: \"c2862399-bef9-4b91-9376-3ceff3377234\") " pod="openstack/glance-db-create-m576z" Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.620212 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-j4qvx"] Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.621075 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-m576z" Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.854789 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-m576z"] Oct 02 01:59:14 crc kubenswrapper[4775]: W1002 01:59:14.856090 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2862399_bef9_4b91_9376_3ceff3377234.slice/crio-0e989e7d5d55fdb1b68af51a8573cfcceb6749a30d753322c1c850f8ab439ed5 WatchSource:0}: Error finding container 0e989e7d5d55fdb1b68af51a8573cfcceb6749a30d753322c1c850f8ab439ed5: Status 404 returned error can't find the container with id 0e989e7d5d55fdb1b68af51a8573cfcceb6749a30d753322c1c850f8ab439ed5 Oct 02 01:59:14 crc kubenswrapper[4775]: W1002 01:59:14.860571 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd01b3c76_8f88_4fc6_be13_4f8150a14949.slice/crio-2935032b031ffe8d8df0a0953b5e0a8ace6f579e7704a521b17e00235809d3cf WatchSource:0}: Error finding container 2935032b031ffe8d8df0a0953b5e0a8ace6f579e7704a521b17e00235809d3cf: Status 404 returned error can't find the container with id 2935032b031ffe8d8df0a0953b5e0a8ace6f579e7704a521b17e00235809d3cf Oct 02 01:59:14 crc kubenswrapper[4775]: I1002 01:59:14.864509 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-582w8"] Oct 02 01:59:15 crc kubenswrapper[4775]: I1002 01:59:15.407425 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-j4qvx" event={"ID":"8a95c042-cb96-42be-8192-2af9ad63d43d","Type":"ContainerStarted","Data":"4af49828720104976bf7c6bb5fc2da2e9b2da748200c0332a9ce42070048ad17"} Oct 02 01:59:15 crc kubenswrapper[4775]: I1002 01:59:15.408115 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-j4qvx" event={"ID":"8a95c042-cb96-42be-8192-2af9ad63d43d","Type":"ContainerStarted","Data":"231070eed26d2013432aa5c8edeed3dffb71b9e11bd34e2b4249ebe56b118236"} Oct 02 01:59:15 crc kubenswrapper[4775]: I1002 01:59:15.410265 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-582w8" event={"ID":"d01b3c76-8f88-4fc6-be13-4f8150a14949","Type":"ContainerStarted","Data":"2935032b031ffe8d8df0a0953b5e0a8ace6f579e7704a521b17e00235809d3cf"} Oct 02 01:59:15 crc kubenswrapper[4775]: I1002 01:59:15.412491 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-m576z" event={"ID":"c2862399-bef9-4b91-9376-3ceff3377234","Type":"ContainerStarted","Data":"0e989e7d5d55fdb1b68af51a8573cfcceb6749a30d753322c1c850f8ab439ed5"} Oct 02 01:59:16 crc kubenswrapper[4775]: I1002 01:59:16.424244 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-582w8" event={"ID":"d01b3c76-8f88-4fc6-be13-4f8150a14949","Type":"ContainerStarted","Data":"b4c3d90e94f7841e53ddb7eb14f21806ea980605924b570fe8fd49f97bbe5369"} Oct 02 01:59:16 crc kubenswrapper[4775]: I1002 01:59:16.427027 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-m576z" event={"ID":"c2862399-bef9-4b91-9376-3ceff3377234","Type":"ContainerStarted","Data":"4e8acae44e5a153d024f7a04cbd26852d78255397c1796bbd4fd45591adafa27"} Oct 02 01:59:16 crc kubenswrapper[4775]: I1002 01:59:16.459661 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-582w8" podStartSLOduration=2.457980701 podStartE2EDuration="2.457980701s" podCreationTimestamp="2025-10-02 01:59:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:59:16.445692092 +0000 UTC m=+1093.612436142" watchObservedRunningTime="2025-10-02 01:59:16.457980701 +0000 UTC m=+1093.624724801" Oct 02 01:59:16 crc kubenswrapper[4775]: I1002 01:59:16.488288 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-j4qvx" podStartSLOduration=3.488259342 podStartE2EDuration="3.488259342s" podCreationTimestamp="2025-10-02 01:59:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:59:16.460577456 +0000 UTC m=+1093.627321506" watchObservedRunningTime="2025-10-02 01:59:16.488259342 +0000 UTC m=+1093.655003402" Oct 02 01:59:16 crc kubenswrapper[4775]: I1002 01:59:16.494177 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-m576z" podStartSLOduration=2.49416585 podStartE2EDuration="2.49416585s" podCreationTimestamp="2025-10-02 01:59:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:59:16.477339157 +0000 UTC m=+1093.644083217" watchObservedRunningTime="2025-10-02 01:59:16.49416585 +0000 UTC m=+1093.660909900" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.215607 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jdbs4" podUID="6efbe4b3-953d-4927-95c6-9b92708f51eb" containerName="ovn-controller" probeResult="failure" output=< Oct 02 01:59:20 crc kubenswrapper[4775]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 02 01:59:20 crc kubenswrapper[4775]: > Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.243195 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.256453 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.527776 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jdbs4-config-pl22p"] Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.529610 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.532893 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.538639 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jdbs4-config-pl22p"] Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.635354 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-log-ovn\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.635440 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b45614-cad3-4f7c-984f-ac1367f67f1b-additional-scripts\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.635486 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nclgn\" (UniqueName: \"kubernetes.io/projected/d1b45614-cad3-4f7c-984f-ac1367f67f1b-kube-api-access-nclgn\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.635823 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1b45614-cad3-4f7c-984f-ac1367f67f1b-scripts\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.635899 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-run-ovn\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.635950 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-run\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.737803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nclgn\" (UniqueName: \"kubernetes.io/projected/d1b45614-cad3-4f7c-984f-ac1367f67f1b-kube-api-access-nclgn\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.737985 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1b45614-cad3-4f7c-984f-ac1367f67f1b-scripts\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.738040 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-run-ovn\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.738101 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-run\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.738251 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-log-ovn\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.738315 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b45614-cad3-4f7c-984f-ac1367f67f1b-additional-scripts\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.738427 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-run-ovn\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.738471 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-run\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.738432 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-log-ovn\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.739546 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b45614-cad3-4f7c-984f-ac1367f67f1b-additional-scripts\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.741000 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1b45614-cad3-4f7c-984f-ac1367f67f1b-scripts\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.758743 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nclgn\" (UniqueName: \"kubernetes.io/projected/d1b45614-cad3-4f7c-984f-ac1367f67f1b-kube-api-access-nclgn\") pod \"ovn-controller-jdbs4-config-pl22p\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:20 crc kubenswrapper[4775]: I1002 01:59:20.897357 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:21 crc kubenswrapper[4775]: I1002 01:59:21.419019 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jdbs4-config-pl22p"] Oct 02 01:59:21 crc kubenswrapper[4775]: I1002 01:59:21.486856 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jdbs4-config-pl22p" event={"ID":"d1b45614-cad3-4f7c-984f-ac1367f67f1b","Type":"ContainerStarted","Data":"9f83b81fa8b1da629b1131628bf952b46c959bd280fa094cc0f4bee68b4b4f64"} Oct 02 01:59:22 crc kubenswrapper[4775]: I1002 01:59:22.495749 4775 generic.go:334] "Generic (PLEG): container finished" podID="c2862399-bef9-4b91-9376-3ceff3377234" containerID="4e8acae44e5a153d024f7a04cbd26852d78255397c1796bbd4fd45591adafa27" exitCode=0 Oct 02 01:59:22 crc kubenswrapper[4775]: I1002 01:59:22.495909 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-m576z" event={"ID":"c2862399-bef9-4b91-9376-3ceff3377234","Type":"ContainerDied","Data":"4e8acae44e5a153d024f7a04cbd26852d78255397c1796bbd4fd45591adafa27"} Oct 02 01:59:22 crc kubenswrapper[4775]: I1002 01:59:22.498536 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jdbs4-config-pl22p" event={"ID":"d1b45614-cad3-4f7c-984f-ac1367f67f1b","Type":"ContainerStarted","Data":"5b6fefa0d5b41b1d53bf1ad854a62fb1e03570cf6981e37e33270368af706fc8"} Oct 02 01:59:22 crc kubenswrapper[4775]: I1002 01:59:22.501008 4775 generic.go:334] "Generic (PLEG): container finished" podID="6638baa9-b367-40d1-8111-673dca6434ad" containerID="21c7d213dfbd2ef3e9ab592c212ff69020698f6ba49a9ebb52c79a43aa64f473" exitCode=0 Oct 02 01:59:22 crc kubenswrapper[4775]: I1002 01:59:22.501085 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6638baa9-b367-40d1-8111-673dca6434ad","Type":"ContainerDied","Data":"21c7d213dfbd2ef3e9ab592c212ff69020698f6ba49a9ebb52c79a43aa64f473"} Oct 02 01:59:22 crc kubenswrapper[4775]: I1002 01:59:22.503123 4775 generic.go:334] "Generic (PLEG): container finished" podID="8a95c042-cb96-42be-8192-2af9ad63d43d" containerID="4af49828720104976bf7c6bb5fc2da2e9b2da748200c0332a9ce42070048ad17" exitCode=0 Oct 02 01:59:22 crc kubenswrapper[4775]: I1002 01:59:22.503195 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-j4qvx" event={"ID":"8a95c042-cb96-42be-8192-2af9ad63d43d","Type":"ContainerDied","Data":"4af49828720104976bf7c6bb5fc2da2e9b2da748200c0332a9ce42070048ad17"} Oct 02 01:59:22 crc kubenswrapper[4775]: I1002 01:59:22.506324 4775 generic.go:334] "Generic (PLEG): container finished" podID="d01b3c76-8f88-4fc6-be13-4f8150a14949" containerID="b4c3d90e94f7841e53ddb7eb14f21806ea980605924b570fe8fd49f97bbe5369" exitCode=0 Oct 02 01:59:22 crc kubenswrapper[4775]: I1002 01:59:22.506417 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-582w8" event={"ID":"d01b3c76-8f88-4fc6-be13-4f8150a14949","Type":"ContainerDied","Data":"b4c3d90e94f7841e53ddb7eb14f21806ea980605924b570fe8fd49f97bbe5369"} Oct 02 01:59:22 crc kubenswrapper[4775]: I1002 01:59:22.507939 4775 generic.go:334] "Generic (PLEG): container finished" podID="fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" containerID="ab0021968685a2c8011cf5e84c4a1e4f33be77976a9063ea49b60d6569a853a3" exitCode=0 Oct 02 01:59:22 crc kubenswrapper[4775]: I1002 01:59:22.507997 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9","Type":"ContainerDied","Data":"ab0021968685a2c8011cf5e84c4a1e4f33be77976a9063ea49b60d6569a853a3"} Oct 02 01:59:22 crc kubenswrapper[4775]: I1002 01:59:22.603055 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-jdbs4-config-pl22p" podStartSLOduration=2.6030376840000002 podStartE2EDuration="2.603037684s" podCreationTimestamp="2025-10-02 01:59:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:59:22.601946667 +0000 UTC m=+1099.768690717" watchObservedRunningTime="2025-10-02 01:59:22.603037684 +0000 UTC m=+1099.769781724" Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.525812 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9","Type":"ContainerStarted","Data":"c4f574fe362f1c98b06c39ae135acf138f8c156d889aa930a36b192af34be971"} Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.526852 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.529226 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"35e88601383e15eea377e15e1d17c552b0acdc4bf1bb0f64c98b395f45c898cf"} Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.529276 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"1ad3d27e30ff5a6104b653354bc8348baa64801d396c0e15ef454a96004528cf"} Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.529288 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"93013dec09b53cd82ac1d950f89f6d21acde0be93880dc71f932f03315d26749"} Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.529296 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"e1f6027f6175a22e8ea34c7faa9698a5ad269b7fc4b557121c200890ba72ee33"} Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.531041 4775 generic.go:334] "Generic (PLEG): container finished" podID="d1b45614-cad3-4f7c-984f-ac1367f67f1b" containerID="5b6fefa0d5b41b1d53bf1ad854a62fb1e03570cf6981e37e33270368af706fc8" exitCode=0 Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.531108 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jdbs4-config-pl22p" event={"ID":"d1b45614-cad3-4f7c-984f-ac1367f67f1b","Type":"ContainerDied","Data":"5b6fefa0d5b41b1d53bf1ad854a62fb1e03570cf6981e37e33270368af706fc8"} Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.533121 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6638baa9-b367-40d1-8111-673dca6434ad","Type":"ContainerStarted","Data":"836e86acc5faf2a3e9656e8d8d2d493dd928f3335bd576fc6d293e8ae72cc25c"} Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.533980 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.562993 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=51.0882608 podStartE2EDuration="1m4.562972614s" podCreationTimestamp="2025-10-02 01:58:19 +0000 UTC" firstStartedPulling="2025-10-02 01:58:33.742225256 +0000 UTC m=+1050.908969316" lastFinishedPulling="2025-10-02 01:58:47.21693709 +0000 UTC m=+1064.383681130" observedRunningTime="2025-10-02 01:59:23.554391658 +0000 UTC m=+1100.721135698" watchObservedRunningTime="2025-10-02 01:59:23.562972614 +0000 UTC m=+1100.729716654" Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.587452 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=52.247105969 podStartE2EDuration="1m4.587437389s" podCreationTimestamp="2025-10-02 01:58:19 +0000 UTC" firstStartedPulling="2025-10-02 01:58:34.257894037 +0000 UTC m=+1051.424638077" lastFinishedPulling="2025-10-02 01:58:46.598225467 +0000 UTC m=+1063.764969497" observedRunningTime="2025-10-02 01:59:23.58429601 +0000 UTC m=+1100.751040050" watchObservedRunningTime="2025-10-02 01:59:23.587437389 +0000 UTC m=+1100.754181429" Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.905559 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-m576z" Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.972317 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j4qvx" Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.977249 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-582w8" Oct 02 01:59:23 crc kubenswrapper[4775]: I1002 01:59:23.995725 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbgjg\" (UniqueName: \"kubernetes.io/projected/c2862399-bef9-4b91-9376-3ceff3377234-kube-api-access-jbgjg\") pod \"c2862399-bef9-4b91-9376-3ceff3377234\" (UID: \"c2862399-bef9-4b91-9376-3ceff3377234\") " Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.002614 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2862399-bef9-4b91-9376-3ceff3377234-kube-api-access-jbgjg" (OuterVolumeSpecName: "kube-api-access-jbgjg") pod "c2862399-bef9-4b91-9376-3ceff3377234" (UID: "c2862399-bef9-4b91-9376-3ceff3377234"). InnerVolumeSpecName "kube-api-access-jbgjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.097121 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sk8r\" (UniqueName: \"kubernetes.io/projected/d01b3c76-8f88-4fc6-be13-4f8150a14949-kube-api-access-9sk8r\") pod \"d01b3c76-8f88-4fc6-be13-4f8150a14949\" (UID: \"d01b3c76-8f88-4fc6-be13-4f8150a14949\") " Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.097302 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgd86\" (UniqueName: \"kubernetes.io/projected/8a95c042-cb96-42be-8192-2af9ad63d43d-kube-api-access-wgd86\") pod \"8a95c042-cb96-42be-8192-2af9ad63d43d\" (UID: \"8a95c042-cb96-42be-8192-2af9ad63d43d\") " Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.097822 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbgjg\" (UniqueName: \"kubernetes.io/projected/c2862399-bef9-4b91-9376-3ceff3377234-kube-api-access-jbgjg\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.100039 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d01b3c76-8f88-4fc6-be13-4f8150a14949-kube-api-access-9sk8r" (OuterVolumeSpecName: "kube-api-access-9sk8r") pod "d01b3c76-8f88-4fc6-be13-4f8150a14949" (UID: "d01b3c76-8f88-4fc6-be13-4f8150a14949"). InnerVolumeSpecName "kube-api-access-9sk8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.101642 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a95c042-cb96-42be-8192-2af9ad63d43d-kube-api-access-wgd86" (OuterVolumeSpecName: "kube-api-access-wgd86") pod "8a95c042-cb96-42be-8192-2af9ad63d43d" (UID: "8a95c042-cb96-42be-8192-2af9ad63d43d"). InnerVolumeSpecName "kube-api-access-wgd86". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.199550 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sk8r\" (UniqueName: \"kubernetes.io/projected/d01b3c76-8f88-4fc6-be13-4f8150a14949-kube-api-access-9sk8r\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.199581 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgd86\" (UniqueName: \"kubernetes.io/projected/8a95c042-cb96-42be-8192-2af9ad63d43d-kube-api-access-wgd86\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.544798 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-582w8" event={"ID":"d01b3c76-8f88-4fc6-be13-4f8150a14949","Type":"ContainerDied","Data":"2935032b031ffe8d8df0a0953b5e0a8ace6f579e7704a521b17e00235809d3cf"} Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.544878 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2935032b031ffe8d8df0a0953b5e0a8ace6f579e7704a521b17e00235809d3cf" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.544824 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-582w8" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.547601 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-m576z" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.547612 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-m576z" event={"ID":"c2862399-bef9-4b91-9376-3ceff3377234","Type":"ContainerDied","Data":"0e989e7d5d55fdb1b68af51a8573cfcceb6749a30d753322c1c850f8ab439ed5"} Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.547740 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e989e7d5d55fdb1b68af51a8573cfcceb6749a30d753322c1c850f8ab439ed5" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.550056 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-j4qvx" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.553081 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-j4qvx" event={"ID":"8a95c042-cb96-42be-8192-2af9ad63d43d","Type":"ContainerDied","Data":"231070eed26d2013432aa5c8edeed3dffb71b9e11bd34e2b4249ebe56b118236"} Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.553148 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="231070eed26d2013432aa5c8edeed3dffb71b9e11bd34e2b4249ebe56b118236" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.868449 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.914756 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-run-ovn\") pod \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.914822 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b45614-cad3-4f7c-984f-ac1367f67f1b-additional-scripts\") pod \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.914875 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1b45614-cad3-4f7c-984f-ac1367f67f1b-scripts\") pod \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.914897 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nclgn\" (UniqueName: \"kubernetes.io/projected/d1b45614-cad3-4f7c-984f-ac1367f67f1b-kube-api-access-nclgn\") pod \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.914904 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "d1b45614-cad3-4f7c-984f-ac1367f67f1b" (UID: "d1b45614-cad3-4f7c-984f-ac1367f67f1b"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.914926 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-run\") pod \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.914978 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-run" (OuterVolumeSpecName: "var-run") pod "d1b45614-cad3-4f7c-984f-ac1367f67f1b" (UID: "d1b45614-cad3-4f7c-984f-ac1367f67f1b"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.915051 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-log-ovn\") pod \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\" (UID: \"d1b45614-cad3-4f7c-984f-ac1367f67f1b\") " Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.915637 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b45614-cad3-4f7c-984f-ac1367f67f1b-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "d1b45614-cad3-4f7c-984f-ac1367f67f1b" (UID: "d1b45614-cad3-4f7c-984f-ac1367f67f1b"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.915666 4775 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.915691 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "d1b45614-cad3-4f7c-984f-ac1367f67f1b" (UID: "d1b45614-cad3-4f7c-984f-ac1367f67f1b"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.915706 4775 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.915837 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b45614-cad3-4f7c-984f-ac1367f67f1b-scripts" (OuterVolumeSpecName: "scripts") pod "d1b45614-cad3-4f7c-984f-ac1367f67f1b" (UID: "d1b45614-cad3-4f7c-984f-ac1367f67f1b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:24 crc kubenswrapper[4775]: I1002 01:59:24.919791 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1b45614-cad3-4f7c-984f-ac1367f67f1b-kube-api-access-nclgn" (OuterVolumeSpecName: "kube-api-access-nclgn") pod "d1b45614-cad3-4f7c-984f-ac1367f67f1b" (UID: "d1b45614-cad3-4f7c-984f-ac1367f67f1b"). InnerVolumeSpecName "kube-api-access-nclgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:25 crc kubenswrapper[4775]: I1002 01:59:25.016316 4775 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d1b45614-cad3-4f7c-984f-ac1367f67f1b-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:25 crc kubenswrapper[4775]: I1002 01:59:25.016341 4775 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b45614-cad3-4f7c-984f-ac1367f67f1b-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:25 crc kubenswrapper[4775]: I1002 01:59:25.016351 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d1b45614-cad3-4f7c-984f-ac1367f67f1b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:25 crc kubenswrapper[4775]: I1002 01:59:25.016361 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nclgn\" (UniqueName: \"kubernetes.io/projected/d1b45614-cad3-4f7c-984f-ac1367f67f1b-kube-api-access-nclgn\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:25 crc kubenswrapper[4775]: I1002 01:59:25.172152 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-jdbs4" Oct 02 01:59:25 crc kubenswrapper[4775]: I1002 01:59:25.559697 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jdbs4-config-pl22p" event={"ID":"d1b45614-cad3-4f7c-984f-ac1367f67f1b","Type":"ContainerDied","Data":"9f83b81fa8b1da629b1131628bf952b46c959bd280fa094cc0f4bee68b4b4f64"} Oct 02 01:59:25 crc kubenswrapper[4775]: I1002 01:59:25.559732 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f83b81fa8b1da629b1131628bf952b46c959bd280fa094cc0f4bee68b4b4f64" Oct 02 01:59:25 crc kubenswrapper[4775]: I1002 01:59:25.559782 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jdbs4-config-pl22p" Oct 02 01:59:25 crc kubenswrapper[4775]: I1002 01:59:25.995187 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jdbs4-config-pl22p"] Oct 02 01:59:26 crc kubenswrapper[4775]: I1002 01:59:26.002939 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jdbs4-config-pl22p"] Oct 02 01:59:26 crc kubenswrapper[4775]: I1002 01:59:26.571001 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"783ca618d590525d6b1848438fe648b59eddb9e00181a39a7cabe0cd1d417d06"} Oct 02 01:59:26 crc kubenswrapper[4775]: I1002 01:59:26.571064 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"05475c086b8f31273f54dfe829e118fed85817c9d16d5585b851871fc39726bb"} Oct 02 01:59:26 crc kubenswrapper[4775]: I1002 01:59:26.571077 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"24cb4406a78e877a5465e016c64de5577c4354d83b07c019e2d33609646a1f9a"} Oct 02 01:59:26 crc kubenswrapper[4775]: I1002 01:59:26.571086 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"a994b8ddd7e1ce4fa5946caa21707a15ffb193c4b308890c750b2ed3745c88d6"} Oct 02 01:59:27 crc kubenswrapper[4775]: I1002 01:59:27.585465 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"ad68cb16ec4f02e54fb4a47b88804b69e1f0edd04ff7e5ec3f86cdfe3bdfc45d"} Oct 02 01:59:27 crc kubenswrapper[4775]: I1002 01:59:27.585825 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"b14587fcd25cde3d5786c84529e1eac9820c2a094fc6d465d49b957d0ca1f117"} Oct 02 01:59:27 crc kubenswrapper[4775]: I1002 01:59:27.778575 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1b45614-cad3-4f7c-984f-ac1367f67f1b" path="/var/lib/kubelet/pods/d1b45614-cad3-4f7c-984f-ac1367f67f1b/volumes" Oct 02 01:59:28 crc kubenswrapper[4775]: I1002 01:59:28.602268 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"e43595485c4ef19d17323cd2f15cc8c6c01cfd72d63f9027002a311666221ea2"} Oct 02 01:59:28 crc kubenswrapper[4775]: I1002 01:59:28.602576 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"a9fa0f88f8c28ceafea5f7d19fe1fa35b3b7c2e9511b51b17451d5c799b2adc6"} Oct 02 01:59:28 crc kubenswrapper[4775]: I1002 01:59:28.602588 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"3b361b7fc7887886b0530a00fe1e6d2dd5a5e58cc96a870064fbcbeb95903b7b"} Oct 02 01:59:28 crc kubenswrapper[4775]: I1002 01:59:28.602597 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"6772443aac50d3396138519e7b7ffbb69baa9434eb913eb475d376b8df66e3ec"} Oct 02 01:59:29 crc kubenswrapper[4775]: I1002 01:59:29.615042 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerStarted","Data":"9a1e1ae6edc64baff01fb577c5fdff389a505836706e4e8e364c628542190596"} Oct 02 01:59:29 crc kubenswrapper[4775]: I1002 01:59:29.651927 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.969359107 podStartE2EDuration="33.651909605s" podCreationTimestamp="2025-10-02 01:58:56 +0000 UTC" firstStartedPulling="2025-10-02 01:59:13.426259205 +0000 UTC m=+1090.593003255" lastFinishedPulling="2025-10-02 01:59:27.108809713 +0000 UTC m=+1104.275553753" observedRunningTime="2025-10-02 01:59:29.643796926 +0000 UTC m=+1106.810540966" watchObservedRunningTime="2025-10-02 01:59:29.651909605 +0000 UTC m=+1106.818653645" Oct 02 01:59:29 crc kubenswrapper[4775]: I1002 01:59:29.938086 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-nx6tz"] Oct 02 01:59:29 crc kubenswrapper[4775]: E1002 01:59:29.938532 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a95c042-cb96-42be-8192-2af9ad63d43d" containerName="mariadb-database-create" Oct 02 01:59:29 crc kubenswrapper[4775]: I1002 01:59:29.938556 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a95c042-cb96-42be-8192-2af9ad63d43d" containerName="mariadb-database-create" Oct 02 01:59:29 crc kubenswrapper[4775]: E1002 01:59:29.938603 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d01b3c76-8f88-4fc6-be13-4f8150a14949" containerName="mariadb-database-create" Oct 02 01:59:29 crc kubenswrapper[4775]: I1002 01:59:29.938613 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d01b3c76-8f88-4fc6-be13-4f8150a14949" containerName="mariadb-database-create" Oct 02 01:59:29 crc kubenswrapper[4775]: E1002 01:59:29.938631 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2862399-bef9-4b91-9376-3ceff3377234" containerName="mariadb-database-create" Oct 02 01:59:29 crc kubenswrapper[4775]: I1002 01:59:29.938639 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2862399-bef9-4b91-9376-3ceff3377234" containerName="mariadb-database-create" Oct 02 01:59:29 crc kubenswrapper[4775]: E1002 01:59:29.938650 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b45614-cad3-4f7c-984f-ac1367f67f1b" containerName="ovn-config" Oct 02 01:59:29 crc kubenswrapper[4775]: I1002 01:59:29.938658 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b45614-cad3-4f7c-984f-ac1367f67f1b" containerName="ovn-config" Oct 02 01:59:29 crc kubenswrapper[4775]: I1002 01:59:29.938839 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a95c042-cb96-42be-8192-2af9ad63d43d" containerName="mariadb-database-create" Oct 02 01:59:29 crc kubenswrapper[4775]: I1002 01:59:29.938874 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2862399-bef9-4b91-9376-3ceff3377234" containerName="mariadb-database-create" Oct 02 01:59:29 crc kubenswrapper[4775]: I1002 01:59:29.938888 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d01b3c76-8f88-4fc6-be13-4f8150a14949" containerName="mariadb-database-create" Oct 02 01:59:29 crc kubenswrapper[4775]: I1002 01:59:29.938908 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b45614-cad3-4f7c-984f-ac1367f67f1b" containerName="ovn-config" Oct 02 01:59:29 crc kubenswrapper[4775]: I1002 01:59:29.940134 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:29 crc kubenswrapper[4775]: I1002 01:59:29.941641 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 02 01:59:29 crc kubenswrapper[4775]: I1002 01:59:29.953620 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-nx6tz"] Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.002021 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-config\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.002070 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.002096 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.002116 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.002180 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.002205 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj4cb\" (UniqueName: \"kubernetes.io/projected/11558ec3-d684-403c-9afe-f08cb76fb18a-kube-api-access-tj4cb\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.103368 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.103457 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.103487 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj4cb\" (UniqueName: \"kubernetes.io/projected/11558ec3-d684-403c-9afe-f08cb76fb18a-kube-api-access-tj4cb\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.103551 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-config\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.103572 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.103593 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.104503 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-ovsdbserver-sb\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.104503 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-dns-swift-storage-0\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.104565 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-dns-svc\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.104890 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-ovsdbserver-nb\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.105042 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-config\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.130234 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj4cb\" (UniqueName: \"kubernetes.io/projected/11558ec3-d684-403c-9afe-f08cb76fb18a-kube-api-access-tj4cb\") pod \"dnsmasq-dns-77585f5f8c-nx6tz\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.258781 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:30 crc kubenswrapper[4775]: I1002 01:59:30.733290 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-nx6tz"] Oct 02 01:59:30 crc kubenswrapper[4775]: W1002 01:59:30.738654 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod11558ec3_d684_403c_9afe_f08cb76fb18a.slice/crio-c479cd7a690e0c2e3e22dfa9f63c6404211aa4de14826c477648d0508f77fab6 WatchSource:0}: Error finding container c479cd7a690e0c2e3e22dfa9f63c6404211aa4de14826c477648d0508f77fab6: Status 404 returned error can't find the container with id c479cd7a690e0c2e3e22dfa9f63c6404211aa4de14826c477648d0508f77fab6 Oct 02 01:59:31 crc kubenswrapper[4775]: I1002 01:59:31.628040 4775 generic.go:334] "Generic (PLEG): container finished" podID="11558ec3-d684-403c-9afe-f08cb76fb18a" containerID="a62f05a1cf2565d94fb4a6d347e31465959d96929cca83a05d2567b5bcd0704c" exitCode=0 Oct 02 01:59:31 crc kubenswrapper[4775]: I1002 01:59:31.628117 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" event={"ID":"11558ec3-d684-403c-9afe-f08cb76fb18a","Type":"ContainerDied","Data":"a62f05a1cf2565d94fb4a6d347e31465959d96929cca83a05d2567b5bcd0704c"} Oct 02 01:59:31 crc kubenswrapper[4775]: I1002 01:59:31.628308 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" event={"ID":"11558ec3-d684-403c-9afe-f08cb76fb18a","Type":"ContainerStarted","Data":"c479cd7a690e0c2e3e22dfa9f63c6404211aa4de14826c477648d0508f77fab6"} Oct 02 01:59:32 crc kubenswrapper[4775]: I1002 01:59:32.645608 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" event={"ID":"11558ec3-d684-403c-9afe-f08cb76fb18a","Type":"ContainerStarted","Data":"372a057f47923d18519533578d61e7d8f697805ca65253b9dde7c9953336d300"} Oct 02 01:59:32 crc kubenswrapper[4775]: I1002 01:59:32.645942 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:32 crc kubenswrapper[4775]: I1002 01:59:32.677431 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" podStartSLOduration=3.677403477 podStartE2EDuration="3.677403477s" podCreationTimestamp="2025-10-02 01:59:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 01:59:32.664507495 +0000 UTC m=+1109.831251595" watchObservedRunningTime="2025-10-02 01:59:32.677403477 +0000 UTC m=+1109.844147547" Oct 02 01:59:33 crc kubenswrapper[4775]: I1002 01:59:33.843916 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7e50-account-create-nlkzk"] Oct 02 01:59:33 crc kubenswrapper[4775]: I1002 01:59:33.847010 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7e50-account-create-nlkzk" Oct 02 01:59:33 crc kubenswrapper[4775]: I1002 01:59:33.850089 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 02 01:59:33 crc kubenswrapper[4775]: I1002 01:59:33.858935 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt2jc\" (UniqueName: \"kubernetes.io/projected/32229132-5589-46e1-a172-228eaa41bfce-kube-api-access-qt2jc\") pod \"keystone-7e50-account-create-nlkzk\" (UID: \"32229132-5589-46e1-a172-228eaa41bfce\") " pod="openstack/keystone-7e50-account-create-nlkzk" Oct 02 01:59:33 crc kubenswrapper[4775]: I1002 01:59:33.871020 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7e50-account-create-nlkzk"] Oct 02 01:59:33 crc kubenswrapper[4775]: I1002 01:59:33.960741 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt2jc\" (UniqueName: \"kubernetes.io/projected/32229132-5589-46e1-a172-228eaa41bfce-kube-api-access-qt2jc\") pod \"keystone-7e50-account-create-nlkzk\" (UID: \"32229132-5589-46e1-a172-228eaa41bfce\") " pod="openstack/keystone-7e50-account-create-nlkzk" Oct 02 01:59:33 crc kubenswrapper[4775]: I1002 01:59:33.990069 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt2jc\" (UniqueName: \"kubernetes.io/projected/32229132-5589-46e1-a172-228eaa41bfce-kube-api-access-qt2jc\") pod \"keystone-7e50-account-create-nlkzk\" (UID: \"32229132-5589-46e1-a172-228eaa41bfce\") " pod="openstack/keystone-7e50-account-create-nlkzk" Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.141545 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-f75b-account-create-6t2rc"] Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.143331 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f75b-account-create-6t2rc" Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.147447 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.156517 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f75b-account-create-6t2rc"] Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.164890 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v24dk\" (UniqueName: \"kubernetes.io/projected/b7f333d7-e144-4ad1-be2f-aae61372eab5-kube-api-access-v24dk\") pod \"placement-f75b-account-create-6t2rc\" (UID: \"b7f333d7-e144-4ad1-be2f-aae61372eab5\") " pod="openstack/placement-f75b-account-create-6t2rc" Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.171527 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7e50-account-create-nlkzk" Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.266778 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v24dk\" (UniqueName: \"kubernetes.io/projected/b7f333d7-e144-4ad1-be2f-aae61372eab5-kube-api-access-v24dk\") pod \"placement-f75b-account-create-6t2rc\" (UID: \"b7f333d7-e144-4ad1-be2f-aae61372eab5\") " pod="openstack/placement-f75b-account-create-6t2rc" Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.299983 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v24dk\" (UniqueName: \"kubernetes.io/projected/b7f333d7-e144-4ad1-be2f-aae61372eab5-kube-api-access-v24dk\") pod \"placement-f75b-account-create-6t2rc\" (UID: \"b7f333d7-e144-4ad1-be2f-aae61372eab5\") " pod="openstack/placement-f75b-account-create-6t2rc" Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.383541 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7e50-account-create-nlkzk"] Oct 02 01:59:34 crc kubenswrapper[4775]: W1002 01:59:34.385267 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod32229132_5589_46e1_a172_228eaa41bfce.slice/crio-28f81ad70e7b6dbe936e7135acad0d0a891b379c23636a2504d739329a0c2533 WatchSource:0}: Error finding container 28f81ad70e7b6dbe936e7135acad0d0a891b379c23636a2504d739329a0c2533: Status 404 returned error can't find the container with id 28f81ad70e7b6dbe936e7135acad0d0a891b379c23636a2504d739329a0c2533 Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.462673 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f75b-account-create-6t2rc" Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.462750 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-95b1-account-create-57xcl"] Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.464190 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-95b1-account-create-57xcl" Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.466343 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.487454 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-95b1-account-create-57xcl"] Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.574328 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vpbb8\" (UniqueName: \"kubernetes.io/projected/bcde2ae4-f3e6-4a25-a9d9-ede53f848a44-kube-api-access-vpbb8\") pod \"glance-95b1-account-create-57xcl\" (UID: \"bcde2ae4-f3e6-4a25-a9d9-ede53f848a44\") " pod="openstack/glance-95b1-account-create-57xcl" Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.669138 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7e50-account-create-nlkzk" event={"ID":"32229132-5589-46e1-a172-228eaa41bfce","Type":"ContainerStarted","Data":"28f81ad70e7b6dbe936e7135acad0d0a891b379c23636a2504d739329a0c2533"} Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.676349 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vpbb8\" (UniqueName: \"kubernetes.io/projected/bcde2ae4-f3e6-4a25-a9d9-ede53f848a44-kube-api-access-vpbb8\") pod \"glance-95b1-account-create-57xcl\" (UID: \"bcde2ae4-f3e6-4a25-a9d9-ede53f848a44\") " pod="openstack/glance-95b1-account-create-57xcl" Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.695113 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vpbb8\" (UniqueName: \"kubernetes.io/projected/bcde2ae4-f3e6-4a25-a9d9-ede53f848a44-kube-api-access-vpbb8\") pod \"glance-95b1-account-create-57xcl\" (UID: \"bcde2ae4-f3e6-4a25-a9d9-ede53f848a44\") " pod="openstack/glance-95b1-account-create-57xcl" Oct 02 01:59:34 crc kubenswrapper[4775]: I1002 01:59:34.804879 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-95b1-account-create-57xcl" Oct 02 01:59:35 crc kubenswrapper[4775]: I1002 01:59:35.005212 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-f75b-account-create-6t2rc"] Oct 02 01:59:35 crc kubenswrapper[4775]: I1002 01:59:35.377649 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-95b1-account-create-57xcl"] Oct 02 01:59:35 crc kubenswrapper[4775]: W1002 01:59:35.383539 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcde2ae4_f3e6_4a25_a9d9_ede53f848a44.slice/crio-95224c82798010bca9e7b9f8d5939ceb697bf6ac32c9c3fd8d4d3032116bed02 WatchSource:0}: Error finding container 95224c82798010bca9e7b9f8d5939ceb697bf6ac32c9c3fd8d4d3032116bed02: Status 404 returned error can't find the container with id 95224c82798010bca9e7b9f8d5939ceb697bf6ac32c9c3fd8d4d3032116bed02 Oct 02 01:59:35 crc kubenswrapper[4775]: I1002 01:59:35.683669 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-95b1-account-create-57xcl" event={"ID":"bcde2ae4-f3e6-4a25-a9d9-ede53f848a44","Type":"ContainerStarted","Data":"95224c82798010bca9e7b9f8d5939ceb697bf6ac32c9c3fd8d4d3032116bed02"} Oct 02 01:59:35 crc kubenswrapper[4775]: I1002 01:59:35.686730 4775 generic.go:334] "Generic (PLEG): container finished" podID="b7f333d7-e144-4ad1-be2f-aae61372eab5" containerID="dd169edaae10c781fde2aef6637ad4843abc69b4d5b73927c6dc220be9781a04" exitCode=0 Oct 02 01:59:35 crc kubenswrapper[4775]: I1002 01:59:35.686808 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f75b-account-create-6t2rc" event={"ID":"b7f333d7-e144-4ad1-be2f-aae61372eab5","Type":"ContainerDied","Data":"dd169edaae10c781fde2aef6637ad4843abc69b4d5b73927c6dc220be9781a04"} Oct 02 01:59:35 crc kubenswrapper[4775]: I1002 01:59:35.686840 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f75b-account-create-6t2rc" event={"ID":"b7f333d7-e144-4ad1-be2f-aae61372eab5","Type":"ContainerStarted","Data":"a224f495bd6739640861da20df55064a4643db16c499184480b0d077f0355ab8"} Oct 02 01:59:35 crc kubenswrapper[4775]: I1002 01:59:35.689904 4775 generic.go:334] "Generic (PLEG): container finished" podID="32229132-5589-46e1-a172-228eaa41bfce" containerID="63549f219f1ba666ecd985751af31b24d4b8335c07e064c09cf580d3c02e98c2" exitCode=0 Oct 02 01:59:35 crc kubenswrapper[4775]: I1002 01:59:35.689988 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7e50-account-create-nlkzk" event={"ID":"32229132-5589-46e1-a172-228eaa41bfce","Type":"ContainerDied","Data":"63549f219f1ba666ecd985751af31b24d4b8335c07e064c09cf580d3c02e98c2"} Oct 02 01:59:36 crc kubenswrapper[4775]: I1002 01:59:36.703482 4775 generic.go:334] "Generic (PLEG): container finished" podID="bcde2ae4-f3e6-4a25-a9d9-ede53f848a44" containerID="af48ca44cb3f8a0a6d273249e42c44fcc94dcbeba408fe59a65dad59c211b74b" exitCode=0 Oct 02 01:59:36 crc kubenswrapper[4775]: I1002 01:59:36.703601 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-95b1-account-create-57xcl" event={"ID":"bcde2ae4-f3e6-4a25-a9d9-ede53f848a44","Type":"ContainerDied","Data":"af48ca44cb3f8a0a6d273249e42c44fcc94dcbeba408fe59a65dad59c211b74b"} Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.116980 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7e50-account-create-nlkzk" Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.125330 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f75b-account-create-6t2rc" Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.220812 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qt2jc\" (UniqueName: \"kubernetes.io/projected/32229132-5589-46e1-a172-228eaa41bfce-kube-api-access-qt2jc\") pod \"32229132-5589-46e1-a172-228eaa41bfce\" (UID: \"32229132-5589-46e1-a172-228eaa41bfce\") " Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.220973 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v24dk\" (UniqueName: \"kubernetes.io/projected/b7f333d7-e144-4ad1-be2f-aae61372eab5-kube-api-access-v24dk\") pod \"b7f333d7-e144-4ad1-be2f-aae61372eab5\" (UID: \"b7f333d7-e144-4ad1-be2f-aae61372eab5\") " Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.229058 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32229132-5589-46e1-a172-228eaa41bfce-kube-api-access-qt2jc" (OuterVolumeSpecName: "kube-api-access-qt2jc") pod "32229132-5589-46e1-a172-228eaa41bfce" (UID: "32229132-5589-46e1-a172-228eaa41bfce"). InnerVolumeSpecName "kube-api-access-qt2jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.229123 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7f333d7-e144-4ad1-be2f-aae61372eab5-kube-api-access-v24dk" (OuterVolumeSpecName: "kube-api-access-v24dk") pod "b7f333d7-e144-4ad1-be2f-aae61372eab5" (UID: "b7f333d7-e144-4ad1-be2f-aae61372eab5"). InnerVolumeSpecName "kube-api-access-v24dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.233513 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.233566 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.323924 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qt2jc\" (UniqueName: \"kubernetes.io/projected/32229132-5589-46e1-a172-228eaa41bfce-kube-api-access-qt2jc\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.324003 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v24dk\" (UniqueName: \"kubernetes.io/projected/b7f333d7-e144-4ad1-be2f-aae61372eab5-kube-api-access-v24dk\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.720095 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-f75b-account-create-6t2rc" event={"ID":"b7f333d7-e144-4ad1-be2f-aae61372eab5","Type":"ContainerDied","Data":"a224f495bd6739640861da20df55064a4643db16c499184480b0d077f0355ab8"} Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.720133 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a224f495bd6739640861da20df55064a4643db16c499184480b0d077f0355ab8" Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.720170 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-f75b-account-create-6t2rc" Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.722221 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7e50-account-create-nlkzk" event={"ID":"32229132-5589-46e1-a172-228eaa41bfce","Type":"ContainerDied","Data":"28f81ad70e7b6dbe936e7135acad0d0a891b379c23636a2504d739329a0c2533"} Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.722457 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28f81ad70e7b6dbe936e7135acad0d0a891b379c23636a2504d739329a0c2533" Oct 02 01:59:37 crc kubenswrapper[4775]: I1002 01:59:37.722340 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7e50-account-create-nlkzk" Oct 02 01:59:38 crc kubenswrapper[4775]: I1002 01:59:38.129405 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-95b1-account-create-57xcl" Oct 02 01:59:38 crc kubenswrapper[4775]: I1002 01:59:38.139051 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vpbb8\" (UniqueName: \"kubernetes.io/projected/bcde2ae4-f3e6-4a25-a9d9-ede53f848a44-kube-api-access-vpbb8\") pod \"bcde2ae4-f3e6-4a25-a9d9-ede53f848a44\" (UID: \"bcde2ae4-f3e6-4a25-a9d9-ede53f848a44\") " Oct 02 01:59:38 crc kubenswrapper[4775]: I1002 01:59:38.147099 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcde2ae4-f3e6-4a25-a9d9-ede53f848a44-kube-api-access-vpbb8" (OuterVolumeSpecName: "kube-api-access-vpbb8") pod "bcde2ae4-f3e6-4a25-a9d9-ede53f848a44" (UID: "bcde2ae4-f3e6-4a25-a9d9-ede53f848a44"). InnerVolumeSpecName "kube-api-access-vpbb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:38 crc kubenswrapper[4775]: I1002 01:59:38.241720 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vpbb8\" (UniqueName: \"kubernetes.io/projected/bcde2ae4-f3e6-4a25-a9d9-ede53f848a44-kube-api-access-vpbb8\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:38 crc kubenswrapper[4775]: I1002 01:59:38.735472 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-95b1-account-create-57xcl" event={"ID":"bcde2ae4-f3e6-4a25-a9d9-ede53f848a44","Type":"ContainerDied","Data":"95224c82798010bca9e7b9f8d5939ceb697bf6ac32c9c3fd8d4d3032116bed02"} Oct 02 01:59:38 crc kubenswrapper[4775]: I1002 01:59:38.735839 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95224c82798010bca9e7b9f8d5939ceb697bf6ac32c9c3fd8d4d3032116bed02" Oct 02 01:59:38 crc kubenswrapper[4775]: I1002 01:59:38.735534 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-95b1-account-create-57xcl" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.615465 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-q2bd5"] Oct 02 01:59:39 crc kubenswrapper[4775]: E1002 01:59:39.616166 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7f333d7-e144-4ad1-be2f-aae61372eab5" containerName="mariadb-account-create" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.616189 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7f333d7-e144-4ad1-be2f-aae61372eab5" containerName="mariadb-account-create" Oct 02 01:59:39 crc kubenswrapper[4775]: E1002 01:59:39.616208 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcde2ae4-f3e6-4a25-a9d9-ede53f848a44" containerName="mariadb-account-create" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.616217 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcde2ae4-f3e6-4a25-a9d9-ede53f848a44" containerName="mariadb-account-create" Oct 02 01:59:39 crc kubenswrapper[4775]: E1002 01:59:39.616235 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32229132-5589-46e1-a172-228eaa41bfce" containerName="mariadb-account-create" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.616244 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="32229132-5589-46e1-a172-228eaa41bfce" containerName="mariadb-account-create" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.616485 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcde2ae4-f3e6-4a25-a9d9-ede53f848a44" containerName="mariadb-account-create" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.616502 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="32229132-5589-46e1-a172-228eaa41bfce" containerName="mariadb-account-create" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.616529 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7f333d7-e144-4ad1-be2f-aae61372eab5" containerName="mariadb-account-create" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.617201 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-q2bd5" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.621916 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5shrf" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.622168 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.624478 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-q2bd5"] Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.669533 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-combined-ca-bundle\") pod \"glance-db-sync-q2bd5\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " pod="openstack/glance-db-sync-q2bd5" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.669663 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-config-data\") pod \"glance-db-sync-q2bd5\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " pod="openstack/glance-db-sync-q2bd5" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.669751 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-db-sync-config-data\") pod \"glance-db-sync-q2bd5\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " pod="openstack/glance-db-sync-q2bd5" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.669986 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn9sd\" (UniqueName: \"kubernetes.io/projected/8df5f08f-cb36-4414-adb4-9bd68db8a41c-kube-api-access-sn9sd\") pod \"glance-db-sync-q2bd5\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " pod="openstack/glance-db-sync-q2bd5" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.771300 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-config-data\") pod \"glance-db-sync-q2bd5\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " pod="openstack/glance-db-sync-q2bd5" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.771409 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-db-sync-config-data\") pod \"glance-db-sync-q2bd5\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " pod="openstack/glance-db-sync-q2bd5" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.772577 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn9sd\" (UniqueName: \"kubernetes.io/projected/8df5f08f-cb36-4414-adb4-9bd68db8a41c-kube-api-access-sn9sd\") pod \"glance-db-sync-q2bd5\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " pod="openstack/glance-db-sync-q2bd5" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.772672 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-combined-ca-bundle\") pod \"glance-db-sync-q2bd5\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " pod="openstack/glance-db-sync-q2bd5" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.778290 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-combined-ca-bundle\") pod \"glance-db-sync-q2bd5\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " pod="openstack/glance-db-sync-q2bd5" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.782120 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-config-data\") pod \"glance-db-sync-q2bd5\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " pod="openstack/glance-db-sync-q2bd5" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.787858 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn9sd\" (UniqueName: \"kubernetes.io/projected/8df5f08f-cb36-4414-adb4-9bd68db8a41c-kube-api-access-sn9sd\") pod \"glance-db-sync-q2bd5\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " pod="openstack/glance-db-sync-q2bd5" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.788234 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-db-sync-config-data\") pod \"glance-db-sync-q2bd5\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " pod="openstack/glance-db-sync-q2bd5" Oct 02 01:59:39 crc kubenswrapper[4775]: I1002 01:59:39.971373 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-q2bd5" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.263146 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.318372 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-gtkrb"] Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.318689 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-gtkrb" podUID="ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a" containerName="dnsmasq-dns" containerID="cri-o://cc7ff92e14efddf8b284f5a7099a1ce780c7a0e2fa56dfd1b699868a8fd238d8" gracePeriod=10 Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.564206 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-q2bd5"] Oct 02 01:59:40 crc kubenswrapper[4775]: W1002 01:59:40.571450 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8df5f08f_cb36_4414_adb4_9bd68db8a41c.slice/crio-7bfbc9e9f62e78d6dbed72ae8e234ccdf7983c37e2518f17f4a28e4542086dcc WatchSource:0}: Error finding container 7bfbc9e9f62e78d6dbed72ae8e234ccdf7983c37e2518f17f4a28e4542086dcc: Status 404 returned error can't find the container with id 7bfbc9e9f62e78d6dbed72ae8e234ccdf7983c37e2518f17f4a28e4542086dcc Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.710394 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.750973 4775 generic.go:334] "Generic (PLEG): container finished" podID="ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a" containerID="cc7ff92e14efddf8b284f5a7099a1ce780c7a0e2fa56dfd1b699868a8fd238d8" exitCode=0 Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.751023 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-gtkrb" event={"ID":"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a","Type":"ContainerDied","Data":"cc7ff92e14efddf8b284f5a7099a1ce780c7a0e2fa56dfd1b699868a8fd238d8"} Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.751047 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-gtkrb" event={"ID":"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a","Type":"ContainerDied","Data":"bf5ecb046aee777d82ce592f674e794a2acf10d7d3ca5c2657778b43fe8c9011"} Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.751062 4775 scope.go:117] "RemoveContainer" containerID="cc7ff92e14efddf8b284f5a7099a1ce780c7a0e2fa56dfd1b699868a8fd238d8" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.751146 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-gtkrb" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.753784 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-q2bd5" event={"ID":"8df5f08f-cb36-4414-adb4-9bd68db8a41c","Type":"ContainerStarted","Data":"7bfbc9e9f62e78d6dbed72ae8e234ccdf7983c37e2518f17f4a28e4542086dcc"} Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.772009 4775 scope.go:117] "RemoveContainer" containerID="dd5ab294818540df79b84f0aa04d8b04e35f763a6b0da0ffe3d4f5a627578747" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.790794 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktlnv\" (UniqueName: \"kubernetes.io/projected/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-kube-api-access-ktlnv\") pod \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.790835 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-dns-svc\") pod \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.791022 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-ovsdbserver-nb\") pod \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.791058 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-ovsdbserver-sb\") pod \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.791154 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-config\") pod \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\" (UID: \"ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a\") " Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.796440 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-kube-api-access-ktlnv" (OuterVolumeSpecName: "kube-api-access-ktlnv") pod "ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a" (UID: "ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a"). InnerVolumeSpecName "kube-api-access-ktlnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.828293 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a" (UID: "ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.832108 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a" (UID: "ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.836974 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a" (UID: "ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.838171 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-config" (OuterVolumeSpecName: "config") pod "ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a" (UID: "ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.894273 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.894472 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.894484 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-config\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.894494 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktlnv\" (UniqueName: \"kubernetes.io/projected/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-kube-api-access-ktlnv\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.894504 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.917716 4775 scope.go:117] "RemoveContainer" containerID="cc7ff92e14efddf8b284f5a7099a1ce780c7a0e2fa56dfd1b699868a8fd238d8" Oct 02 01:59:40 crc kubenswrapper[4775]: E1002 01:59:40.918230 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc7ff92e14efddf8b284f5a7099a1ce780c7a0e2fa56dfd1b699868a8fd238d8\": container with ID starting with cc7ff92e14efddf8b284f5a7099a1ce780c7a0e2fa56dfd1b699868a8fd238d8 not found: ID does not exist" containerID="cc7ff92e14efddf8b284f5a7099a1ce780c7a0e2fa56dfd1b699868a8fd238d8" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.918271 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc7ff92e14efddf8b284f5a7099a1ce780c7a0e2fa56dfd1b699868a8fd238d8"} err="failed to get container status \"cc7ff92e14efddf8b284f5a7099a1ce780c7a0e2fa56dfd1b699868a8fd238d8\": rpc error: code = NotFound desc = could not find container \"cc7ff92e14efddf8b284f5a7099a1ce780c7a0e2fa56dfd1b699868a8fd238d8\": container with ID starting with cc7ff92e14efddf8b284f5a7099a1ce780c7a0e2fa56dfd1b699868a8fd238d8 not found: ID does not exist" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.918297 4775 scope.go:117] "RemoveContainer" containerID="dd5ab294818540df79b84f0aa04d8b04e35f763a6b0da0ffe3d4f5a627578747" Oct 02 01:59:40 crc kubenswrapper[4775]: E1002 01:59:40.918687 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd5ab294818540df79b84f0aa04d8b04e35f763a6b0da0ffe3d4f5a627578747\": container with ID starting with dd5ab294818540df79b84f0aa04d8b04e35f763a6b0da0ffe3d4f5a627578747 not found: ID does not exist" containerID="dd5ab294818540df79b84f0aa04d8b04e35f763a6b0da0ffe3d4f5a627578747" Oct 02 01:59:40 crc kubenswrapper[4775]: I1002 01:59:40.918714 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd5ab294818540df79b84f0aa04d8b04e35f763a6b0da0ffe3d4f5a627578747"} err="failed to get container status \"dd5ab294818540df79b84f0aa04d8b04e35f763a6b0da0ffe3d4f5a627578747\": rpc error: code = NotFound desc = could not find container \"dd5ab294818540df79b84f0aa04d8b04e35f763a6b0da0ffe3d4f5a627578747\": container with ID starting with dd5ab294818540df79b84f0aa04d8b04e35f763a6b0da0ffe3d4f5a627578747 not found: ID does not exist" Oct 02 01:59:41 crc kubenswrapper[4775]: I1002 01:59:41.088188 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-gtkrb"] Oct 02 01:59:41 crc kubenswrapper[4775]: I1002 01:59:41.099226 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 01:59:41 crc kubenswrapper[4775]: I1002 01:59:41.099795 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-gtkrb"] Oct 02 01:59:41 crc kubenswrapper[4775]: I1002 01:59:41.142180 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 01:59:41 crc kubenswrapper[4775]: I1002 01:59:41.776860 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a" path="/var/lib/kubelet/pods/ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a/volumes" Oct 02 01:59:42 crc kubenswrapper[4775]: I1002 01:59:42.799773 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-t6skn"] Oct 02 01:59:42 crc kubenswrapper[4775]: E1002 01:59:42.801223 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a" containerName="init" Oct 02 01:59:42 crc kubenswrapper[4775]: I1002 01:59:42.801311 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a" containerName="init" Oct 02 01:59:42 crc kubenswrapper[4775]: E1002 01:59:42.801404 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a" containerName="dnsmasq-dns" Oct 02 01:59:42 crc kubenswrapper[4775]: I1002 01:59:42.801478 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a" containerName="dnsmasq-dns" Oct 02 01:59:42 crc kubenswrapper[4775]: I1002 01:59:42.801755 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab7102fc-6f9b-4ca6-be20-4d2bed3ab89a" containerName="dnsmasq-dns" Oct 02 01:59:42 crc kubenswrapper[4775]: I1002 01:59:42.802470 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t6skn" Oct 02 01:59:42 crc kubenswrapper[4775]: I1002 01:59:42.826773 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jw58\" (UniqueName: \"kubernetes.io/projected/f14a7a78-6b40-4c82-8dc6-64cc069cb1fd-kube-api-access-4jw58\") pod \"cinder-db-create-t6skn\" (UID: \"f14a7a78-6b40-4c82-8dc6-64cc069cb1fd\") " pod="openstack/cinder-db-create-t6skn" Oct 02 01:59:42 crc kubenswrapper[4775]: I1002 01:59:42.826889 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-t6skn"] Oct 02 01:59:42 crc kubenswrapper[4775]: I1002 01:59:42.899916 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-pfvt8"] Oct 02 01:59:42 crc kubenswrapper[4775]: I1002 01:59:42.901075 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pfvt8" Oct 02 01:59:42 crc kubenswrapper[4775]: I1002 01:59:42.906800 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-pfvt8"] Oct 02 01:59:42 crc kubenswrapper[4775]: I1002 01:59:42.927749 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jw58\" (UniqueName: \"kubernetes.io/projected/f14a7a78-6b40-4c82-8dc6-64cc069cb1fd-kube-api-access-4jw58\") pod \"cinder-db-create-t6skn\" (UID: \"f14a7a78-6b40-4c82-8dc6-64cc069cb1fd\") " pod="openstack/cinder-db-create-t6skn" Oct 02 01:59:42 crc kubenswrapper[4775]: I1002 01:59:42.948115 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jw58\" (UniqueName: \"kubernetes.io/projected/f14a7a78-6b40-4c82-8dc6-64cc069cb1fd-kube-api-access-4jw58\") pod \"cinder-db-create-t6skn\" (UID: \"f14a7a78-6b40-4c82-8dc6-64cc069cb1fd\") " pod="openstack/cinder-db-create-t6skn" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.029025 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjpcg\" (UniqueName: \"kubernetes.io/projected/77750932-5356-469a-ad3d-640973d830d9-kube-api-access-sjpcg\") pod \"barbican-db-create-pfvt8\" (UID: \"77750932-5356-469a-ad3d-640973d830d9\") " pod="openstack/barbican-db-create-pfvt8" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.123556 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-x8t6d"] Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.124128 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t6skn" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.124822 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x8t6d" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.135242 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjpcg\" (UniqueName: \"kubernetes.io/projected/77750932-5356-469a-ad3d-640973d830d9-kube-api-access-sjpcg\") pod \"barbican-db-create-pfvt8\" (UID: \"77750932-5356-469a-ad3d-640973d830d9\") " pod="openstack/barbican-db-create-pfvt8" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.135316 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjxxz\" (UniqueName: \"kubernetes.io/projected/cdaf4602-ca11-419d-846c-425d3d5bc766-kube-api-access-tjxxz\") pod \"neutron-db-create-x8t6d\" (UID: \"cdaf4602-ca11-419d-846c-425d3d5bc766\") " pod="openstack/neutron-db-create-x8t6d" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.138857 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-x8t6d"] Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.170880 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-vqhvp"] Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.171860 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vqhvp" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.173600 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.174143 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjpcg\" (UniqueName: \"kubernetes.io/projected/77750932-5356-469a-ad3d-640973d830d9-kube-api-access-sjpcg\") pod \"barbican-db-create-pfvt8\" (UID: \"77750932-5356-469a-ad3d-640973d830d9\") " pod="openstack/barbican-db-create-pfvt8" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.175706 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.175842 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.175939 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-shncl" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.180853 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-vqhvp"] Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.214315 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pfvt8" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.235751 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66k8b\" (UniqueName: \"kubernetes.io/projected/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-kube-api-access-66k8b\") pod \"keystone-db-sync-vqhvp\" (UID: \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\") " pod="openstack/keystone-db-sync-vqhvp" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.235807 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-config-data\") pod \"keystone-db-sync-vqhvp\" (UID: \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\") " pod="openstack/keystone-db-sync-vqhvp" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.235853 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjxxz\" (UniqueName: \"kubernetes.io/projected/cdaf4602-ca11-419d-846c-425d3d5bc766-kube-api-access-tjxxz\") pod \"neutron-db-create-x8t6d\" (UID: \"cdaf4602-ca11-419d-846c-425d3d5bc766\") " pod="openstack/neutron-db-create-x8t6d" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.235873 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-combined-ca-bundle\") pod \"keystone-db-sync-vqhvp\" (UID: \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\") " pod="openstack/keystone-db-sync-vqhvp" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.251016 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjxxz\" (UniqueName: \"kubernetes.io/projected/cdaf4602-ca11-419d-846c-425d3d5bc766-kube-api-access-tjxxz\") pod \"neutron-db-create-x8t6d\" (UID: \"cdaf4602-ca11-419d-846c-425d3d5bc766\") " pod="openstack/neutron-db-create-x8t6d" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.336886 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66k8b\" (UniqueName: \"kubernetes.io/projected/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-kube-api-access-66k8b\") pod \"keystone-db-sync-vqhvp\" (UID: \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\") " pod="openstack/keystone-db-sync-vqhvp" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.337215 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-config-data\") pod \"keystone-db-sync-vqhvp\" (UID: \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\") " pod="openstack/keystone-db-sync-vqhvp" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.337258 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-combined-ca-bundle\") pod \"keystone-db-sync-vqhvp\" (UID: \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\") " pod="openstack/keystone-db-sync-vqhvp" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.342465 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-config-data\") pod \"keystone-db-sync-vqhvp\" (UID: \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\") " pod="openstack/keystone-db-sync-vqhvp" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.351589 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-combined-ca-bundle\") pod \"keystone-db-sync-vqhvp\" (UID: \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\") " pod="openstack/keystone-db-sync-vqhvp" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.354370 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66k8b\" (UniqueName: \"kubernetes.io/projected/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-kube-api-access-66k8b\") pod \"keystone-db-sync-vqhvp\" (UID: \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\") " pod="openstack/keystone-db-sync-vqhvp" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.452393 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x8t6d" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.604679 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vqhvp" Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.671601 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-t6skn"] Oct 02 01:59:43 crc kubenswrapper[4775]: W1002 01:59:43.677164 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf14a7a78_6b40_4c82_8dc6_64cc069cb1fd.slice/crio-355953705afe08b1e5dfa8f11f62db10ae0443b44e41fc69daa208cda77f6e18 WatchSource:0}: Error finding container 355953705afe08b1e5dfa8f11f62db10ae0443b44e41fc69daa208cda77f6e18: Status 404 returned error can't find the container with id 355953705afe08b1e5dfa8f11f62db10ae0443b44e41fc69daa208cda77f6e18 Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.722138 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-pfvt8"] Oct 02 01:59:43 crc kubenswrapper[4775]: W1002 01:59:43.724406 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77750932_5356_469a_ad3d_640973d830d9.slice/crio-2fc2be6015a74cc8efe9fa4237d6a9b638ff3d1f3406b6fb8c2fd72aa610da2b WatchSource:0}: Error finding container 2fc2be6015a74cc8efe9fa4237d6a9b638ff3d1f3406b6fb8c2fd72aa610da2b: Status 404 returned error can't find the container with id 2fc2be6015a74cc8efe9fa4237d6a9b638ff3d1f3406b6fb8c2fd72aa610da2b Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.781748 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-pfvt8" event={"ID":"77750932-5356-469a-ad3d-640973d830d9","Type":"ContainerStarted","Data":"2fc2be6015a74cc8efe9fa4237d6a9b638ff3d1f3406b6fb8c2fd72aa610da2b"} Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.787287 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-t6skn" event={"ID":"f14a7a78-6b40-4c82-8dc6-64cc069cb1fd","Type":"ContainerStarted","Data":"355953705afe08b1e5dfa8f11f62db10ae0443b44e41fc69daa208cda77f6e18"} Oct 02 01:59:43 crc kubenswrapper[4775]: I1002 01:59:43.974268 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-x8t6d"] Oct 02 01:59:43 crc kubenswrapper[4775]: W1002 01:59:43.996188 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcdaf4602_ca11_419d_846c_425d3d5bc766.slice/crio-0f16f8eb2a52981c0e35842625fd9df7945294f082547394dea8d00ff66a4ca9 WatchSource:0}: Error finding container 0f16f8eb2a52981c0e35842625fd9df7945294f082547394dea8d00ff66a4ca9: Status 404 returned error can't find the container with id 0f16f8eb2a52981c0e35842625fd9df7945294f082547394dea8d00ff66a4ca9 Oct 02 01:59:44 crc kubenswrapper[4775]: I1002 01:59:44.093820 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-vqhvp"] Oct 02 01:59:44 crc kubenswrapper[4775]: W1002 01:59:44.107095 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d9d080a_bb9e_4394_947f_5fb5df9a52ec.slice/crio-232ac4d08332857f4917f6952793af032897f2b017e4317d19f47a258f791891 WatchSource:0}: Error finding container 232ac4d08332857f4917f6952793af032897f2b017e4317d19f47a258f791891: Status 404 returned error can't find the container with id 232ac4d08332857f4917f6952793af032897f2b017e4317d19f47a258f791891 Oct 02 01:59:44 crc kubenswrapper[4775]: I1002 01:59:44.797614 4775 generic.go:334] "Generic (PLEG): container finished" podID="77750932-5356-469a-ad3d-640973d830d9" containerID="fd0824af89285e773df2653c2d0e7d72035cffd31ec8d3fbdbb045018aca4264" exitCode=0 Oct 02 01:59:44 crc kubenswrapper[4775]: I1002 01:59:44.797800 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-pfvt8" event={"ID":"77750932-5356-469a-ad3d-640973d830d9","Type":"ContainerDied","Data":"fd0824af89285e773df2653c2d0e7d72035cffd31ec8d3fbdbb045018aca4264"} Oct 02 01:59:44 crc kubenswrapper[4775]: I1002 01:59:44.799597 4775 generic.go:334] "Generic (PLEG): container finished" podID="cdaf4602-ca11-419d-846c-425d3d5bc766" containerID="fb50998c7136bd8340419cbe461f93a9ee8bc40a5577fb31338f8fea2f87aead" exitCode=0 Oct 02 01:59:44 crc kubenswrapper[4775]: I1002 01:59:44.799663 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x8t6d" event={"ID":"cdaf4602-ca11-419d-846c-425d3d5bc766","Type":"ContainerDied","Data":"fb50998c7136bd8340419cbe461f93a9ee8bc40a5577fb31338f8fea2f87aead"} Oct 02 01:59:44 crc kubenswrapper[4775]: I1002 01:59:44.799682 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x8t6d" event={"ID":"cdaf4602-ca11-419d-846c-425d3d5bc766","Type":"ContainerStarted","Data":"0f16f8eb2a52981c0e35842625fd9df7945294f082547394dea8d00ff66a4ca9"} Oct 02 01:59:44 crc kubenswrapper[4775]: I1002 01:59:44.801619 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vqhvp" event={"ID":"4d9d080a-bb9e-4394-947f-5fb5df9a52ec","Type":"ContainerStarted","Data":"232ac4d08332857f4917f6952793af032897f2b017e4317d19f47a258f791891"} Oct 02 01:59:44 crc kubenswrapper[4775]: I1002 01:59:44.807552 4775 generic.go:334] "Generic (PLEG): container finished" podID="f14a7a78-6b40-4c82-8dc6-64cc069cb1fd" containerID="83c116e723946985568da8014ced2542488c38c32b369e695abb19f1ab5c6a58" exitCode=0 Oct 02 01:59:44 crc kubenswrapper[4775]: I1002 01:59:44.807575 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-t6skn" event={"ID":"f14a7a78-6b40-4c82-8dc6-64cc069cb1fd","Type":"ContainerDied","Data":"83c116e723946985568da8014ced2542488c38c32b369e695abb19f1ab5c6a58"} Oct 02 01:59:46 crc kubenswrapper[4775]: I1002 01:59:46.073224 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t6skn" Oct 02 01:59:46 crc kubenswrapper[4775]: I1002 01:59:46.085939 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jw58\" (UniqueName: \"kubernetes.io/projected/f14a7a78-6b40-4c82-8dc6-64cc069cb1fd-kube-api-access-4jw58\") pod \"f14a7a78-6b40-4c82-8dc6-64cc069cb1fd\" (UID: \"f14a7a78-6b40-4c82-8dc6-64cc069cb1fd\") " Oct 02 01:59:46 crc kubenswrapper[4775]: I1002 01:59:46.098940 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f14a7a78-6b40-4c82-8dc6-64cc069cb1fd-kube-api-access-4jw58" (OuterVolumeSpecName: "kube-api-access-4jw58") pod "f14a7a78-6b40-4c82-8dc6-64cc069cb1fd" (UID: "f14a7a78-6b40-4c82-8dc6-64cc069cb1fd"). InnerVolumeSpecName "kube-api-access-4jw58". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:46 crc kubenswrapper[4775]: I1002 01:59:46.188716 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jw58\" (UniqueName: \"kubernetes.io/projected/f14a7a78-6b40-4c82-8dc6-64cc069cb1fd-kube-api-access-4jw58\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:46 crc kubenswrapper[4775]: I1002 01:59:46.827484 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-t6skn" event={"ID":"f14a7a78-6b40-4c82-8dc6-64cc069cb1fd","Type":"ContainerDied","Data":"355953705afe08b1e5dfa8f11f62db10ae0443b44e41fc69daa208cda77f6e18"} Oct 02 01:59:46 crc kubenswrapper[4775]: I1002 01:59:46.827551 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="355953705afe08b1e5dfa8f11f62db10ae0443b44e41fc69daa208cda77f6e18" Oct 02 01:59:46 crc kubenswrapper[4775]: I1002 01:59:46.827624 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-t6skn" Oct 02 01:59:52 crc kubenswrapper[4775]: I1002 01:59:52.974896 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-5592-account-create-f4hz6"] Oct 02 01:59:52 crc kubenswrapper[4775]: E1002 01:59:52.976058 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f14a7a78-6b40-4c82-8dc6-64cc069cb1fd" containerName="mariadb-database-create" Oct 02 01:59:52 crc kubenswrapper[4775]: I1002 01:59:52.976080 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f14a7a78-6b40-4c82-8dc6-64cc069cb1fd" containerName="mariadb-database-create" Oct 02 01:59:52 crc kubenswrapper[4775]: I1002 01:59:52.976419 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f14a7a78-6b40-4c82-8dc6-64cc069cb1fd" containerName="mariadb-database-create" Oct 02 01:59:52 crc kubenswrapper[4775]: I1002 01:59:52.977390 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5592-account-create-f4hz6" Oct 02 01:59:52 crc kubenswrapper[4775]: I1002 01:59:52.979615 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 02 01:59:52 crc kubenswrapper[4775]: I1002 01:59:52.989573 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5592-account-create-f4hz6"] Oct 02 01:59:53 crc kubenswrapper[4775]: I1002 01:59:53.005336 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6tl9\" (UniqueName: \"kubernetes.io/projected/a34ad439-41ea-4ee1-a7b3-4075d9923915-kube-api-access-x6tl9\") pod \"cinder-5592-account-create-f4hz6\" (UID: \"a34ad439-41ea-4ee1-a7b3-4075d9923915\") " pod="openstack/cinder-5592-account-create-f4hz6" Oct 02 01:59:53 crc kubenswrapper[4775]: I1002 01:59:53.106831 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6tl9\" (UniqueName: \"kubernetes.io/projected/a34ad439-41ea-4ee1-a7b3-4075d9923915-kube-api-access-x6tl9\") pod \"cinder-5592-account-create-f4hz6\" (UID: \"a34ad439-41ea-4ee1-a7b3-4075d9923915\") " pod="openstack/cinder-5592-account-create-f4hz6" Oct 02 01:59:53 crc kubenswrapper[4775]: I1002 01:59:53.132883 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6tl9\" (UniqueName: \"kubernetes.io/projected/a34ad439-41ea-4ee1-a7b3-4075d9923915-kube-api-access-x6tl9\") pod \"cinder-5592-account-create-f4hz6\" (UID: \"a34ad439-41ea-4ee1-a7b3-4075d9923915\") " pod="openstack/cinder-5592-account-create-f4hz6" Oct 02 01:59:53 crc kubenswrapper[4775]: I1002 01:59:53.303114 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5592-account-create-f4hz6" Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.329321 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pfvt8" Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.334190 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x8t6d" Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.369057 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjpcg\" (UniqueName: \"kubernetes.io/projected/77750932-5356-469a-ad3d-640973d830d9-kube-api-access-sjpcg\") pod \"77750932-5356-469a-ad3d-640973d830d9\" (UID: \"77750932-5356-469a-ad3d-640973d830d9\") " Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.369101 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjxxz\" (UniqueName: \"kubernetes.io/projected/cdaf4602-ca11-419d-846c-425d3d5bc766-kube-api-access-tjxxz\") pod \"cdaf4602-ca11-419d-846c-425d3d5bc766\" (UID: \"cdaf4602-ca11-419d-846c-425d3d5bc766\") " Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.374069 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdaf4602-ca11-419d-846c-425d3d5bc766-kube-api-access-tjxxz" (OuterVolumeSpecName: "kube-api-access-tjxxz") pod "cdaf4602-ca11-419d-846c-425d3d5bc766" (UID: "cdaf4602-ca11-419d-846c-425d3d5bc766"). InnerVolumeSpecName "kube-api-access-tjxxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.374340 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77750932-5356-469a-ad3d-640973d830d9-kube-api-access-sjpcg" (OuterVolumeSpecName: "kube-api-access-sjpcg") pod "77750932-5356-469a-ad3d-640973d830d9" (UID: "77750932-5356-469a-ad3d-640973d830d9"). InnerVolumeSpecName "kube-api-access-sjpcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.470679 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjpcg\" (UniqueName: \"kubernetes.io/projected/77750932-5356-469a-ad3d-640973d830d9-kube-api-access-sjpcg\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.470704 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjxxz\" (UniqueName: \"kubernetes.io/projected/cdaf4602-ca11-419d-846c-425d3d5bc766-kube-api-access-tjxxz\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.875075 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-5592-account-create-f4hz6"] Oct 02 01:59:56 crc kubenswrapper[4775]: W1002 01:59:56.886045 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda34ad439_41ea_4ee1_a7b3_4075d9923915.slice/crio-752015abb90ad831caadcbc15d948aa55af35c94bee1aa33ad85af36f5037087 WatchSource:0}: Error finding container 752015abb90ad831caadcbc15d948aa55af35c94bee1aa33ad85af36f5037087: Status 404 returned error can't find the container with id 752015abb90ad831caadcbc15d948aa55af35c94bee1aa33ad85af36f5037087 Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.921213 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vqhvp" event={"ID":"4d9d080a-bb9e-4394-947f-5fb5df9a52ec","Type":"ContainerStarted","Data":"b8c221ce14f71e8617a931e96a1dc0043afe3b845f98c89cf4bbb25eb5600820"} Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.924756 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-pfvt8" event={"ID":"77750932-5356-469a-ad3d-640973d830d9","Type":"ContainerDied","Data":"2fc2be6015a74cc8efe9fa4237d6a9b638ff3d1f3406b6fb8c2fd72aa610da2b"} Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.924784 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fc2be6015a74cc8efe9fa4237d6a9b638ff3d1f3406b6fb8c2fd72aa610da2b" Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.924834 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-pfvt8" Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.926524 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-x8t6d" event={"ID":"cdaf4602-ca11-419d-846c-425d3d5bc766","Type":"ContainerDied","Data":"0f16f8eb2a52981c0e35842625fd9df7945294f082547394dea8d00ff66a4ca9"} Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.926547 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f16f8eb2a52981c0e35842625fd9df7945294f082547394dea8d00ff66a4ca9" Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.926552 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-x8t6d" Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.927386 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5592-account-create-f4hz6" event={"ID":"a34ad439-41ea-4ee1-a7b3-4075d9923915","Type":"ContainerStarted","Data":"752015abb90ad831caadcbc15d948aa55af35c94bee1aa33ad85af36f5037087"} Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.928450 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-q2bd5" event={"ID":"8df5f08f-cb36-4414-adb4-9bd68db8a41c","Type":"ContainerStarted","Data":"8b262fc92c0f625440f236f2622f5d81981f9d0a4057e92dd3a337e627891bd7"} Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.945423 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-vqhvp" podStartSLOduration=2.001884107 podStartE2EDuration="13.945374317s" podCreationTimestamp="2025-10-02 01:59:43 +0000 UTC" firstStartedPulling="2025-10-02 01:59:44.109890171 +0000 UTC m=+1121.276634221" lastFinishedPulling="2025-10-02 01:59:56.053380381 +0000 UTC m=+1133.220124431" observedRunningTime="2025-10-02 01:59:56.937762741 +0000 UTC m=+1134.104506811" watchObservedRunningTime="2025-10-02 01:59:56.945374317 +0000 UTC m=+1134.112118367" Oct 02 01:59:56 crc kubenswrapper[4775]: I1002 01:59:56.957262 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-q2bd5" podStartSLOduration=2.445813996 podStartE2EDuration="17.957242683s" podCreationTimestamp="2025-10-02 01:59:39 +0000 UTC" firstStartedPulling="2025-10-02 01:59:40.582564269 +0000 UTC m=+1117.749308309" lastFinishedPulling="2025-10-02 01:59:56.093992956 +0000 UTC m=+1133.260736996" observedRunningTime="2025-10-02 01:59:56.953799684 +0000 UTC m=+1134.120543754" watchObservedRunningTime="2025-10-02 01:59:56.957242683 +0000 UTC m=+1134.123986723" Oct 02 01:59:57 crc kubenswrapper[4775]: I1002 01:59:57.960080 4775 generic.go:334] "Generic (PLEG): container finished" podID="a34ad439-41ea-4ee1-a7b3-4075d9923915" containerID="ea5cddb115e84b5b26c272b09e40f8bf1b5253c0f27dee2955ed905b7c9f3a4f" exitCode=0 Oct 02 01:59:57 crc kubenswrapper[4775]: I1002 01:59:57.960223 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5592-account-create-f4hz6" event={"ID":"a34ad439-41ea-4ee1-a7b3-4075d9923915","Type":"ContainerDied","Data":"ea5cddb115e84b5b26c272b09e40f8bf1b5253c0f27dee2955ed905b7c9f3a4f"} Oct 02 01:59:59 crc kubenswrapper[4775]: I1002 01:59:59.332551 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5592-account-create-f4hz6" Oct 02 01:59:59 crc kubenswrapper[4775]: I1002 01:59:59.426563 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6tl9\" (UniqueName: \"kubernetes.io/projected/a34ad439-41ea-4ee1-a7b3-4075d9923915-kube-api-access-x6tl9\") pod \"a34ad439-41ea-4ee1-a7b3-4075d9923915\" (UID: \"a34ad439-41ea-4ee1-a7b3-4075d9923915\") " Oct 02 01:59:59 crc kubenswrapper[4775]: I1002 01:59:59.446875 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a34ad439-41ea-4ee1-a7b3-4075d9923915-kube-api-access-x6tl9" (OuterVolumeSpecName: "kube-api-access-x6tl9") pod "a34ad439-41ea-4ee1-a7b3-4075d9923915" (UID: "a34ad439-41ea-4ee1-a7b3-4075d9923915"). InnerVolumeSpecName "kube-api-access-x6tl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 01:59:59 crc kubenswrapper[4775]: I1002 01:59:59.529250 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6tl9\" (UniqueName: \"kubernetes.io/projected/a34ad439-41ea-4ee1-a7b3-4075d9923915-kube-api-access-x6tl9\") on node \"crc\" DevicePath \"\"" Oct 02 01:59:59 crc kubenswrapper[4775]: I1002 01:59:59.982853 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-5592-account-create-f4hz6" event={"ID":"a34ad439-41ea-4ee1-a7b3-4075d9923915","Type":"ContainerDied","Data":"752015abb90ad831caadcbc15d948aa55af35c94bee1aa33ad85af36f5037087"} Oct 02 01:59:59 crc kubenswrapper[4775]: I1002 01:59:59.983200 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="752015abb90ad831caadcbc15d948aa55af35c94bee1aa33ad85af36f5037087" Oct 02 01:59:59 crc kubenswrapper[4775]: I1002 01:59:59.982873 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-5592-account-create-f4hz6" Oct 02 01:59:59 crc kubenswrapper[4775]: I1002 01:59:59.985275 4775 generic.go:334] "Generic (PLEG): container finished" podID="4d9d080a-bb9e-4394-947f-5fb5df9a52ec" containerID="b8c221ce14f71e8617a931e96a1dc0043afe3b845f98c89cf4bbb25eb5600820" exitCode=0 Oct 02 01:59:59 crc kubenswrapper[4775]: I1002 01:59:59.985328 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vqhvp" event={"ID":"4d9d080a-bb9e-4394-947f-5fb5df9a52ec","Type":"ContainerDied","Data":"b8c221ce14f71e8617a931e96a1dc0043afe3b845f98c89cf4bbb25eb5600820"} Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.144380 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7"] Oct 02 02:00:00 crc kubenswrapper[4775]: E1002 02:00:00.144765 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdaf4602-ca11-419d-846c-425d3d5bc766" containerName="mariadb-database-create" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.144786 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdaf4602-ca11-419d-846c-425d3d5bc766" containerName="mariadb-database-create" Oct 02 02:00:00 crc kubenswrapper[4775]: E1002 02:00:00.144832 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a34ad439-41ea-4ee1-a7b3-4075d9923915" containerName="mariadb-account-create" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.144841 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a34ad439-41ea-4ee1-a7b3-4075d9923915" containerName="mariadb-account-create" Oct 02 02:00:00 crc kubenswrapper[4775]: E1002 02:00:00.144862 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77750932-5356-469a-ad3d-640973d830d9" containerName="mariadb-database-create" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.144870 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="77750932-5356-469a-ad3d-640973d830d9" containerName="mariadb-database-create" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.145329 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="77750932-5356-469a-ad3d-640973d830d9" containerName="mariadb-database-create" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.145356 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a34ad439-41ea-4ee1-a7b3-4075d9923915" containerName="mariadb-account-create" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.145385 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdaf4602-ca11-419d-846c-425d3d5bc766" containerName="mariadb-database-create" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.148036 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.151353 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.151642 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.165506 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7"] Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.242870 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adbfa780-e6fe-402d-a040-175271a4b1e1-secret-volume\") pod \"collect-profiles-29322840-w98q7\" (UID: \"adbfa780-e6fe-402d-a040-175271a4b1e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.242944 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adbfa780-e6fe-402d-a040-175271a4b1e1-config-volume\") pod \"collect-profiles-29322840-w98q7\" (UID: \"adbfa780-e6fe-402d-a040-175271a4b1e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.243149 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdq9s\" (UniqueName: \"kubernetes.io/projected/adbfa780-e6fe-402d-a040-175271a4b1e1-kube-api-access-vdq9s\") pod \"collect-profiles-29322840-w98q7\" (UID: \"adbfa780-e6fe-402d-a040-175271a4b1e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.345304 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adbfa780-e6fe-402d-a040-175271a4b1e1-config-volume\") pod \"collect-profiles-29322840-w98q7\" (UID: \"adbfa780-e6fe-402d-a040-175271a4b1e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.345457 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdq9s\" (UniqueName: \"kubernetes.io/projected/adbfa780-e6fe-402d-a040-175271a4b1e1-kube-api-access-vdq9s\") pod \"collect-profiles-29322840-w98q7\" (UID: \"adbfa780-e6fe-402d-a040-175271a4b1e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.345587 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adbfa780-e6fe-402d-a040-175271a4b1e1-secret-volume\") pod \"collect-profiles-29322840-w98q7\" (UID: \"adbfa780-e6fe-402d-a040-175271a4b1e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.347435 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adbfa780-e6fe-402d-a040-175271a4b1e1-config-volume\") pod \"collect-profiles-29322840-w98q7\" (UID: \"adbfa780-e6fe-402d-a040-175271a4b1e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.358098 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adbfa780-e6fe-402d-a040-175271a4b1e1-secret-volume\") pod \"collect-profiles-29322840-w98q7\" (UID: \"adbfa780-e6fe-402d-a040-175271a4b1e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.367834 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdq9s\" (UniqueName: \"kubernetes.io/projected/adbfa780-e6fe-402d-a040-175271a4b1e1-kube-api-access-vdq9s\") pod \"collect-profiles-29322840-w98q7\" (UID: \"adbfa780-e6fe-402d-a040-175271a4b1e1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.477701 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.751903 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7"] Oct 02 02:00:00 crc kubenswrapper[4775]: W1002 02:00:00.758486 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadbfa780_e6fe_402d_a040_175271a4b1e1.slice/crio-263fff84a532a44e19563392ed03cb8b541fa92080ac12c9325fcbb282581618 WatchSource:0}: Error finding container 263fff84a532a44e19563392ed03cb8b541fa92080ac12c9325fcbb282581618: Status 404 returned error can't find the container with id 263fff84a532a44e19563392ed03cb8b541fa92080ac12c9325fcbb282581618 Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.997481 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" event={"ID":"adbfa780-e6fe-402d-a040-175271a4b1e1","Type":"ContainerStarted","Data":"b296aa4012103f2996652ab29adbbb7c34e7f48034623ae89d178928e0a69556"} Oct 02 02:00:00 crc kubenswrapper[4775]: I1002 02:00:00.997876 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" event={"ID":"adbfa780-e6fe-402d-a040-175271a4b1e1","Type":"ContainerStarted","Data":"263fff84a532a44e19563392ed03cb8b541fa92080ac12c9325fcbb282581618"} Oct 02 02:00:01 crc kubenswrapper[4775]: I1002 02:00:01.028437 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" podStartSLOduration=1.028415355 podStartE2EDuration="1.028415355s" podCreationTimestamp="2025-10-02 02:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:01.021790985 +0000 UTC m=+1138.188535025" watchObservedRunningTime="2025-10-02 02:00:01.028415355 +0000 UTC m=+1138.195159395" Oct 02 02:00:01 crc kubenswrapper[4775]: E1002 02:00:01.231106 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadbfa780_e6fe_402d_a040_175271a4b1e1.slice/crio-b296aa4012103f2996652ab29adbbb7c34e7f48034623ae89d178928e0a69556.scope\": RecentStats: unable to find data in memory cache]" Oct 02 02:00:01 crc kubenswrapper[4775]: I1002 02:00:01.269540 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vqhvp" Oct 02 02:00:01 crc kubenswrapper[4775]: I1002 02:00:01.362662 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66k8b\" (UniqueName: \"kubernetes.io/projected/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-kube-api-access-66k8b\") pod \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\" (UID: \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\") " Oct 02 02:00:01 crc kubenswrapper[4775]: I1002 02:00:01.362782 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-config-data\") pod \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\" (UID: \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\") " Oct 02 02:00:01 crc kubenswrapper[4775]: I1002 02:00:01.362927 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-combined-ca-bundle\") pod \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\" (UID: \"4d9d080a-bb9e-4394-947f-5fb5df9a52ec\") " Oct 02 02:00:01 crc kubenswrapper[4775]: I1002 02:00:01.368626 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-kube-api-access-66k8b" (OuterVolumeSpecName: "kube-api-access-66k8b") pod "4d9d080a-bb9e-4394-947f-5fb5df9a52ec" (UID: "4d9d080a-bb9e-4394-947f-5fb5df9a52ec"). InnerVolumeSpecName "kube-api-access-66k8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:01 crc kubenswrapper[4775]: I1002 02:00:01.396112 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d9d080a-bb9e-4394-947f-5fb5df9a52ec" (UID: "4d9d080a-bb9e-4394-947f-5fb5df9a52ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:01 crc kubenswrapper[4775]: I1002 02:00:01.411134 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-config-data" (OuterVolumeSpecName: "config-data") pod "4d9d080a-bb9e-4394-947f-5fb5df9a52ec" (UID: "4d9d080a-bb9e-4394-947f-5fb5df9a52ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:01 crc kubenswrapper[4775]: I1002 02:00:01.465429 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66k8b\" (UniqueName: \"kubernetes.io/projected/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-kube-api-access-66k8b\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:01 crc kubenswrapper[4775]: I1002 02:00:01.465487 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:01 crc kubenswrapper[4775]: I1002 02:00:01.465507 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9d080a-bb9e-4394-947f-5fb5df9a52ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.010509 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vqhvp" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.011031 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vqhvp" event={"ID":"4d9d080a-bb9e-4394-947f-5fb5df9a52ec","Type":"ContainerDied","Data":"232ac4d08332857f4917f6952793af032897f2b017e4317d19f47a258f791891"} Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.011111 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="232ac4d08332857f4917f6952793af032897f2b017e4317d19f47a258f791891" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.013194 4775 generic.go:334] "Generic (PLEG): container finished" podID="adbfa780-e6fe-402d-a040-175271a4b1e1" containerID="b296aa4012103f2996652ab29adbbb7c34e7f48034623ae89d178928e0a69556" exitCode=0 Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.013232 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" event={"ID":"adbfa780-e6fe-402d-a040-175271a4b1e1","Type":"ContainerDied","Data":"b296aa4012103f2996652ab29adbbb7c34e7f48034623ae89d178928e0a69556"} Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.260449 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-xrvrd"] Oct 02 02:00:02 crc kubenswrapper[4775]: E1002 02:00:02.260826 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9d080a-bb9e-4394-947f-5fb5df9a52ec" containerName="keystone-db-sync" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.260840 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9d080a-bb9e-4394-947f-5fb5df9a52ec" containerName="keystone-db-sync" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.261098 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9d080a-bb9e-4394-947f-5fb5df9a52ec" containerName="keystone-db-sync" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.262071 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.275885 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-xrvrd"] Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.312030 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-hsh5j"] Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.312978 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.315693 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-shncl" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.315935 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.316088 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.316350 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.324704 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hsh5j"] Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.385742 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-credential-keys\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.386578 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwv27\" (UniqueName: \"kubernetes.io/projected/8f436799-ffa1-45dc-9564-3fbb6b64fffe-kube-api-access-fwv27\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.386667 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-scripts\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.386734 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ntbg\" (UniqueName: \"kubernetes.io/projected/683f8a41-9042-4a9d-9fda-03784d72e926-kube-api-access-2ntbg\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.386846 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-dns-svc\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.386909 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.386999 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-fernet-keys\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.387076 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.387149 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-config\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.387309 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-config-data\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.387431 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-combined-ca-bundle\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.387495 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.447788 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.449624 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.451338 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.454619 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.490039 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-combined-ca-bundle\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.490249 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.490353 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.490414 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-config-data\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.490529 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-credential-keys\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.490580 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwv27\" (UniqueName: \"kubernetes.io/projected/8f436799-ffa1-45dc-9564-3fbb6b64fffe-kube-api-access-fwv27\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.490613 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-scripts\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.490632 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ntbg\" (UniqueName: \"kubernetes.io/projected/683f8a41-9042-4a9d-9fda-03784d72e926-kube-api-access-2ntbg\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.491122 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-run-httpd\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.491146 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-dns-svc\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.491160 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.491184 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-fernet-keys\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.491206 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-scripts\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.491230 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.491245 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-config\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.491260 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.491260 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-dns-swift-storage-0\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.491283 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-log-httpd\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.491531 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpdgg\" (UniqueName: \"kubernetes.io/projected/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-kube-api-access-wpdgg\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.491572 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-config-data\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.491789 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-dns-svc\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.492317 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-ovsdbserver-nb\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.492846 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-ovsdbserver-sb\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.493103 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-config\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.495641 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-config-data\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.495944 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-credential-keys\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.496624 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-combined-ca-bundle\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.504254 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-fernet-keys\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.504856 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-scripts\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.507688 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ntbg\" (UniqueName: \"kubernetes.io/projected/683f8a41-9042-4a9d-9fda-03784d72e926-kube-api-access-2ntbg\") pod \"keystone-bootstrap-hsh5j\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.510040 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.546982 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwv27\" (UniqueName: \"kubernetes.io/projected/8f436799-ffa1-45dc-9564-3fbb6b64fffe-kube-api-access-fwv27\") pod \"dnsmasq-dns-55fff446b9-xrvrd\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.586780 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.593365 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.593420 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-log-httpd\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.593450 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpdgg\" (UniqueName: \"kubernetes.io/projected/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-kube-api-access-wpdgg\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.593504 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.593520 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-config-data\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.593609 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-run-httpd\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.593643 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-scripts\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.593982 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-log-httpd\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.594136 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-run-httpd\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.599134 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-config-data\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.601459 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-scripts\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.607286 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.607522 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.624022 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpdgg\" (UniqueName: \"kubernetes.io/projected/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-kube-api-access-wpdgg\") pod \"ceilometer-0\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.635294 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.676262 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-xrvrd"] Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.697873 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-hdtb5"] Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.699731 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.705167 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.705430 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.705584 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-5smvp" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.730212 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-skdqz"] Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.731501 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.742276 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hdtb5"] Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.764038 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.792047 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-skdqz"] Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.800124 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-config\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.800217 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njjkh\" (UniqueName: \"kubernetes.io/projected/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-kube-api-access-njjkh\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.800256 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-combined-ca-bundle\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.800286 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.800308 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-scripts\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.800331 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-logs\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.800356 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.800375 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.800394 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-config-data\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.800431 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfqg5\" (UniqueName: \"kubernetes.io/projected/babe38d7-4bf5-47f7-acb3-22161a9eac50-kube-api-access-hfqg5\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.800471 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.902373 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-config\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.902975 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njjkh\" (UniqueName: \"kubernetes.io/projected/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-kube-api-access-njjkh\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.903022 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-combined-ca-bundle\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.903051 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.903081 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-scripts\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.903123 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-logs\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.903159 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.903181 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.903202 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-config-data\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.903256 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfqg5\" (UniqueName: \"kubernetes.io/projected/babe38d7-4bf5-47f7-acb3-22161a9eac50-kube-api-access-hfqg5\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.903299 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.903630 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-config\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.904580 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-dns-swift-storage-0\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.904645 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-dns-svc\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.904884 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-logs\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.905421 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-ovsdbserver-nb\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.905900 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-ovsdbserver-sb\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.909024 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-combined-ca-bundle\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.911351 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-config-data\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.911571 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-scripts\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.919256 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njjkh\" (UniqueName: \"kubernetes.io/projected/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-kube-api-access-njjkh\") pod \"placement-db-sync-hdtb5\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.919482 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfqg5\" (UniqueName: \"kubernetes.io/projected/babe38d7-4bf5-47f7-acb3-22161a9eac50-kube-api-access-hfqg5\") pod \"dnsmasq-dns-76fcf4b695-skdqz\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.959994 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-b963-account-create-zbgmn"] Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.961012 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b963-account-create-zbgmn" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.962947 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 02 02:00:02 crc kubenswrapper[4775]: I1002 02:00:02.966454 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b963-account-create-zbgmn"] Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.004560 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbhgk\" (UniqueName: \"kubernetes.io/projected/0b46397b-0c28-46ab-8f20-1426466202b3-kube-api-access-fbhgk\") pod \"barbican-b963-account-create-zbgmn\" (UID: \"0b46397b-0c28-46ab-8f20-1426466202b3\") " pod="openstack/barbican-b963-account-create-zbgmn" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.053030 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.059725 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.105785 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbhgk\" (UniqueName: \"kubernetes.io/projected/0b46397b-0c28-46ab-8f20-1426466202b3-kube-api-access-fbhgk\") pod \"barbican-b963-account-create-zbgmn\" (UID: \"0b46397b-0c28-46ab-8f20-1426466202b3\") " pod="openstack/barbican-b963-account-create-zbgmn" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.140031 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbhgk\" (UniqueName: \"kubernetes.io/projected/0b46397b-0c28-46ab-8f20-1426466202b3-kube-api-access-fbhgk\") pod \"barbican-b963-account-create-zbgmn\" (UID: \"0b46397b-0c28-46ab-8f20-1426466202b3\") " pod="openstack/barbican-b963-account-create-zbgmn" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.165122 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-19ef-account-create-n7j6v"] Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.166152 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-19ef-account-create-n7j6v" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.193217 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.193809 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-xrvrd"] Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.211648 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbqkv\" (UniqueName: \"kubernetes.io/projected/c20e73b6-3188-49c6-836d-e518d93d9457-kube-api-access-mbqkv\") pod \"neutron-19ef-account-create-n7j6v\" (UID: \"c20e73b6-3188-49c6-836d-e518d93d9457\") " pod="openstack/neutron-19ef-account-create-n7j6v" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.242438 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-19ef-account-create-n7j6v"] Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.283358 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b963-account-create-zbgmn" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.364503 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbqkv\" (UniqueName: \"kubernetes.io/projected/c20e73b6-3188-49c6-836d-e518d93d9457-kube-api-access-mbqkv\") pod \"neutron-19ef-account-create-n7j6v\" (UID: \"c20e73b6-3188-49c6-836d-e518d93d9457\") " pod="openstack/neutron-19ef-account-create-n7j6v" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.388874 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hsh5j"] Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.407584 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.458156 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbqkv\" (UniqueName: \"kubernetes.io/projected/c20e73b6-3188-49c6-836d-e518d93d9457-kube-api-access-mbqkv\") pod \"neutron-19ef-account-create-n7j6v\" (UID: \"c20e73b6-3188-49c6-836d-e518d93d9457\") " pod="openstack/neutron-19ef-account-create-n7j6v" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.475910 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-x8pvr"] Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.477427 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.481970 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.482342 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.482827 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-w7wmc" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.487525 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-x8pvr"] Oct 02 02:00:03 crc kubenswrapper[4775]: W1002 02:00:03.523185 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod683f8a41_9042_4a9d_9fda_03784d72e926.slice/crio-1a90b262f1b37086adb48b75163d35f4e27a9b7d7998bb52c7ab1d40acf3c1f1 WatchSource:0}: Error finding container 1a90b262f1b37086adb48b75163d35f4e27a9b7d7998bb52c7ab1d40acf3c1f1: Status 404 returned error can't find the container with id 1a90b262f1b37086adb48b75163d35f4e27a9b7d7998bb52c7ab1d40acf3c1f1 Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.532933 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-19ef-account-create-n7j6v" Oct 02 02:00:03 crc kubenswrapper[4775]: W1002 02:00:03.534830 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12864b75_dbdc_4075_aea7_9d3cc2a6ca59.slice/crio-d8f1d1d00b3490c5599f7fe0eaeb40f3a76aca72ccbf022e9d39d0478cbf05f2 WatchSource:0}: Error finding container d8f1d1d00b3490c5599f7fe0eaeb40f3a76aca72ccbf022e9d39d0478cbf05f2: Status 404 returned error can't find the container with id d8f1d1d00b3490c5599f7fe0eaeb40f3a76aca72ccbf022e9d39d0478cbf05f2 Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.577184 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a3f9c1f-1298-4835-9935-56b808aae84d-etc-machine-id\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.577250 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-scripts\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.577299 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-db-sync-config-data\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.577380 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-combined-ca-bundle\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.577421 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzddh\" (UniqueName: \"kubernetes.io/projected/9a3f9c1f-1298-4835-9935-56b808aae84d-kube-api-access-xzddh\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.577439 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-config-data\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.679183 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-combined-ca-bundle\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.679249 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzddh\" (UniqueName: \"kubernetes.io/projected/9a3f9c1f-1298-4835-9935-56b808aae84d-kube-api-access-xzddh\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.679268 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-config-data\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.679297 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a3f9c1f-1298-4835-9935-56b808aae84d-etc-machine-id\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.679340 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-scripts\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.679378 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-db-sync-config-data\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.680341 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a3f9c1f-1298-4835-9935-56b808aae84d-etc-machine-id\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.686591 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-config-data\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.698736 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-combined-ca-bundle\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.698789 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-scripts\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.698848 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-db-sync-config-data\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.702277 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzddh\" (UniqueName: \"kubernetes.io/projected/9a3f9c1f-1298-4835-9935-56b808aae84d-kube-api-access-xzddh\") pod \"cinder-db-sync-x8pvr\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.775167 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-w7wmc" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.777001 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.786140 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.884665 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-skdqz"] Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.891613 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adbfa780-e6fe-402d-a040-175271a4b1e1-config-volume\") pod \"adbfa780-e6fe-402d-a040-175271a4b1e1\" (UID: \"adbfa780-e6fe-402d-a040-175271a4b1e1\") " Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.891664 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdq9s\" (UniqueName: \"kubernetes.io/projected/adbfa780-e6fe-402d-a040-175271a4b1e1-kube-api-access-vdq9s\") pod \"adbfa780-e6fe-402d-a040-175271a4b1e1\" (UID: \"adbfa780-e6fe-402d-a040-175271a4b1e1\") " Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.891759 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adbfa780-e6fe-402d-a040-175271a4b1e1-secret-volume\") pod \"adbfa780-e6fe-402d-a040-175271a4b1e1\" (UID: \"adbfa780-e6fe-402d-a040-175271a4b1e1\") " Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.893319 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adbfa780-e6fe-402d-a040-175271a4b1e1-config-volume" (OuterVolumeSpecName: "config-volume") pod "adbfa780-e6fe-402d-a040-175271a4b1e1" (UID: "adbfa780-e6fe-402d-a040-175271a4b1e1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.900418 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adbfa780-e6fe-402d-a040-175271a4b1e1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "adbfa780-e6fe-402d-a040-175271a4b1e1" (UID: "adbfa780-e6fe-402d-a040-175271a4b1e1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.902272 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adbfa780-e6fe-402d-a040-175271a4b1e1-kube-api-access-vdq9s" (OuterVolumeSpecName: "kube-api-access-vdq9s") pod "adbfa780-e6fe-402d-a040-175271a4b1e1" (UID: "adbfa780-e6fe-402d-a040-175271a4b1e1"). InnerVolumeSpecName "kube-api-access-vdq9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.973722 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hdtb5"] Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.994261 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/adbfa780-e6fe-402d-a040-175271a4b1e1-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.994290 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdq9s\" (UniqueName: \"kubernetes.io/projected/adbfa780-e6fe-402d-a040-175271a4b1e1-kube-api-access-vdq9s\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:03 crc kubenswrapper[4775]: I1002 02:00:03.994301 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/adbfa780-e6fe-402d-a040-175271a4b1e1-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.047589 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-b963-account-create-zbgmn"] Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.071839 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12864b75-dbdc-4075-aea7-9d3cc2a6ca59","Type":"ContainerStarted","Data":"d8f1d1d00b3490c5599f7fe0eaeb40f3a76aca72ccbf022e9d39d0478cbf05f2"} Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.072039 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.073342 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-19ef-account-create-n7j6v"] Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.073512 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" event={"ID":"adbfa780-e6fe-402d-a040-175271a4b1e1","Type":"ContainerDied","Data":"263fff84a532a44e19563392ed03cb8b541fa92080ac12c9325fcbb282581618"} Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.073536 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="263fff84a532a44e19563392ed03cb8b541fa92080ac12c9325fcbb282581618" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.073572 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.075397 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hsh5j" event={"ID":"683f8a41-9042-4a9d-9fda-03784d72e926","Type":"ContainerStarted","Data":"6896915cc67aadd994ad499da8c5082fb11be528fb9225adc20db0b4e4bee028"} Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.075427 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hsh5j" event={"ID":"683f8a41-9042-4a9d-9fda-03784d72e926","Type":"ContainerStarted","Data":"1a90b262f1b37086adb48b75163d35f4e27a9b7d7998bb52c7ab1d40acf3c1f1"} Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.076965 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" event={"ID":"babe38d7-4bf5-47f7-acb3-22161a9eac50","Type":"ContainerStarted","Data":"de7cc4d85f28ebbb4dbd510296b1659cb218985ba9340aed81a0a9f12ea3da33"} Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.079206 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hdtb5" event={"ID":"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9","Type":"ContainerStarted","Data":"234fdba0bdee12badc1d90753ddb54c667013fa32d8a960666607fd72db604d8"} Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.086600 4775 generic.go:334] "Generic (PLEG): container finished" podID="8f436799-ffa1-45dc-9564-3fbb6b64fffe" containerID="249c72ddd5b95eb2ba324244c00f78be9df33b130908bf4d15f4900d785e4262" exitCode=0 Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.086632 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" event={"ID":"8f436799-ffa1-45dc-9564-3fbb6b64fffe","Type":"ContainerDied","Data":"249c72ddd5b95eb2ba324244c00f78be9df33b130908bf4d15f4900d785e4262"} Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.086657 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" event={"ID":"8f436799-ffa1-45dc-9564-3fbb6b64fffe","Type":"ContainerStarted","Data":"5c18d929c836e2b32543241d40250004e7ef9113d8013ae0d7fc22621b1e419d"} Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.090588 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.102494 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-hsh5j" podStartSLOduration=2.102475326 podStartE2EDuration="2.102475326s" podCreationTimestamp="2025-10-02 02:00:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:04.097864118 +0000 UTC m=+1141.264608158" watchObservedRunningTime="2025-10-02 02:00:04.102475326 +0000 UTC m=+1141.269219366" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.354402 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-x8pvr"] Oct 02 02:00:04 crc kubenswrapper[4775]: W1002 02:00:04.362242 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a3f9c1f_1298_4835_9935_56b808aae84d.slice/crio-003ee24629e2b9cd3dd9930345a1f20f8ff5431e535892d9895a6bb4aa50c018 WatchSource:0}: Error finding container 003ee24629e2b9cd3dd9930345a1f20f8ff5431e535892d9895a6bb4aa50c018: Status 404 returned error can't find the container with id 003ee24629e2b9cd3dd9930345a1f20f8ff5431e535892d9895a6bb4aa50c018 Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.401834 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.508255 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-dns-svc\") pod \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.508596 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-ovsdbserver-nb\") pod \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.508739 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-ovsdbserver-sb\") pod \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.508824 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwv27\" (UniqueName: \"kubernetes.io/projected/8f436799-ffa1-45dc-9564-3fbb6b64fffe-kube-api-access-fwv27\") pod \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.508899 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-config\") pod \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.508979 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-dns-swift-storage-0\") pod \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\" (UID: \"8f436799-ffa1-45dc-9564-3fbb6b64fffe\") " Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.522071 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f436799-ffa1-45dc-9564-3fbb6b64fffe-kube-api-access-fwv27" (OuterVolumeSpecName: "kube-api-access-fwv27") pod "8f436799-ffa1-45dc-9564-3fbb6b64fffe" (UID: "8f436799-ffa1-45dc-9564-3fbb6b64fffe"). InnerVolumeSpecName "kube-api-access-fwv27". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.538485 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8f436799-ffa1-45dc-9564-3fbb6b64fffe" (UID: "8f436799-ffa1-45dc-9564-3fbb6b64fffe"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.539234 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8f436799-ffa1-45dc-9564-3fbb6b64fffe" (UID: "8f436799-ffa1-45dc-9564-3fbb6b64fffe"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.558558 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8f436799-ffa1-45dc-9564-3fbb6b64fffe" (UID: "8f436799-ffa1-45dc-9564-3fbb6b64fffe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.562592 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8f436799-ffa1-45dc-9564-3fbb6b64fffe" (UID: "8f436799-ffa1-45dc-9564-3fbb6b64fffe"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.565098 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-config" (OuterVolumeSpecName: "config") pod "8f436799-ffa1-45dc-9564-3fbb6b64fffe" (UID: "8f436799-ffa1-45dc-9564-3fbb6b64fffe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.610736 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.610764 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwv27\" (UniqueName: \"kubernetes.io/projected/8f436799-ffa1-45dc-9564-3fbb6b64fffe-kube-api-access-fwv27\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.610776 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.610784 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.610793 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:04 crc kubenswrapper[4775]: I1002 02:00:04.610802 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8f436799-ffa1-45dc-9564-3fbb6b64fffe-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.123155 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-x8pvr" event={"ID":"9a3f9c1f-1298-4835-9935-56b808aae84d","Type":"ContainerStarted","Data":"003ee24629e2b9cd3dd9930345a1f20f8ff5431e535892d9895a6bb4aa50c018"} Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.132620 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.133143 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55fff446b9-xrvrd" event={"ID":"8f436799-ffa1-45dc-9564-3fbb6b64fffe","Type":"ContainerDied","Data":"5c18d929c836e2b32543241d40250004e7ef9113d8013ae0d7fc22621b1e419d"} Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.133260 4775 scope.go:117] "RemoveContainer" containerID="249c72ddd5b95eb2ba324244c00f78be9df33b130908bf4d15f4900d785e4262" Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.143978 4775 generic.go:334] "Generic (PLEG): container finished" podID="0b46397b-0c28-46ab-8f20-1426466202b3" containerID="2c422b8f2c034369225bbf9b72a631e551bac35a17c65cc2a633d7e5f7f6a986" exitCode=0 Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.144041 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b963-account-create-zbgmn" event={"ID":"0b46397b-0c28-46ab-8f20-1426466202b3","Type":"ContainerDied","Data":"2c422b8f2c034369225bbf9b72a631e551bac35a17c65cc2a633d7e5f7f6a986"} Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.144072 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b963-account-create-zbgmn" event={"ID":"0b46397b-0c28-46ab-8f20-1426466202b3","Type":"ContainerStarted","Data":"e2136fb29513cd7423fcc0c18a0886a38fdf4527c9fe2867441afda16b7516a4"} Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.151141 4775 generic.go:334] "Generic (PLEG): container finished" podID="c20e73b6-3188-49c6-836d-e518d93d9457" containerID="97fb4e7966718769851726806782e899704e90130857cbf32f13bf170d9d8525" exitCode=0 Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.151192 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-19ef-account-create-n7j6v" event={"ID":"c20e73b6-3188-49c6-836d-e518d93d9457","Type":"ContainerDied","Data":"97fb4e7966718769851726806782e899704e90130857cbf32f13bf170d9d8525"} Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.151215 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-19ef-account-create-n7j6v" event={"ID":"c20e73b6-3188-49c6-836d-e518d93d9457","Type":"ContainerStarted","Data":"9933d8b89bace5bc10ae63307f2d609d332d4120730b2f374434b02bd7840de3"} Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.152529 4775 generic.go:334] "Generic (PLEG): container finished" podID="babe38d7-4bf5-47f7-acb3-22161a9eac50" containerID="032b608b1a7f0df7db55c0d674a323620ff4dceb9728680abc2e92f995715480" exitCode=0 Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.153624 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" event={"ID":"babe38d7-4bf5-47f7-acb3-22161a9eac50","Type":"ContainerDied","Data":"032b608b1a7f0df7db55c0d674a323620ff4dceb9728680abc2e92f995715480"} Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.371067 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-xrvrd"] Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.381449 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55fff446b9-xrvrd"] Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.779983 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f436799-ffa1-45dc-9564-3fbb6b64fffe" path="/var/lib/kubelet/pods/8f436799-ffa1-45dc-9564-3fbb6b64fffe/volumes" Oct 02 02:00:05 crc kubenswrapper[4775]: I1002 02:00:05.882815 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:00:06 crc kubenswrapper[4775]: I1002 02:00:06.165161 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" event={"ID":"babe38d7-4bf5-47f7-acb3-22161a9eac50","Type":"ContainerStarted","Data":"526a7a5bc8b20b157e4b2789ec2fda0e947f785915e94610f2ddb5355f916123"} Oct 02 02:00:06 crc kubenswrapper[4775]: I1002 02:00:06.165640 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:06 crc kubenswrapper[4775]: I1002 02:00:06.187682 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" podStartSLOduration=4.187666549 podStartE2EDuration="4.187666549s" podCreationTimestamp="2025-10-02 02:00:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:06.184026715 +0000 UTC m=+1143.350770775" watchObservedRunningTime="2025-10-02 02:00:06.187666549 +0000 UTC m=+1143.354410579" Oct 02 02:00:06 crc kubenswrapper[4775]: I1002 02:00:06.600084 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-19ef-account-create-n7j6v" Oct 02 02:00:06 crc kubenswrapper[4775]: I1002 02:00:06.611925 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b963-account-create-zbgmn" Oct 02 02:00:06 crc kubenswrapper[4775]: I1002 02:00:06.759294 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbqkv\" (UniqueName: \"kubernetes.io/projected/c20e73b6-3188-49c6-836d-e518d93d9457-kube-api-access-mbqkv\") pod \"c20e73b6-3188-49c6-836d-e518d93d9457\" (UID: \"c20e73b6-3188-49c6-836d-e518d93d9457\") " Oct 02 02:00:06 crc kubenswrapper[4775]: I1002 02:00:06.759442 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbhgk\" (UniqueName: \"kubernetes.io/projected/0b46397b-0c28-46ab-8f20-1426466202b3-kube-api-access-fbhgk\") pod \"0b46397b-0c28-46ab-8f20-1426466202b3\" (UID: \"0b46397b-0c28-46ab-8f20-1426466202b3\") " Oct 02 02:00:06 crc kubenswrapper[4775]: I1002 02:00:06.765168 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b46397b-0c28-46ab-8f20-1426466202b3-kube-api-access-fbhgk" (OuterVolumeSpecName: "kube-api-access-fbhgk") pod "0b46397b-0c28-46ab-8f20-1426466202b3" (UID: "0b46397b-0c28-46ab-8f20-1426466202b3"). InnerVolumeSpecName "kube-api-access-fbhgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:06 crc kubenswrapper[4775]: I1002 02:00:06.766552 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c20e73b6-3188-49c6-836d-e518d93d9457-kube-api-access-mbqkv" (OuterVolumeSpecName: "kube-api-access-mbqkv") pod "c20e73b6-3188-49c6-836d-e518d93d9457" (UID: "c20e73b6-3188-49c6-836d-e518d93d9457"). InnerVolumeSpecName "kube-api-access-mbqkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:06 crc kubenswrapper[4775]: I1002 02:00:06.861826 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbqkv\" (UniqueName: \"kubernetes.io/projected/c20e73b6-3188-49c6-836d-e518d93d9457-kube-api-access-mbqkv\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:06 crc kubenswrapper[4775]: I1002 02:00:06.861859 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbhgk\" (UniqueName: \"kubernetes.io/projected/0b46397b-0c28-46ab-8f20-1426466202b3-kube-api-access-fbhgk\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:07 crc kubenswrapper[4775]: I1002 02:00:07.183831 4775 generic.go:334] "Generic (PLEG): container finished" podID="8df5f08f-cb36-4414-adb4-9bd68db8a41c" containerID="8b262fc92c0f625440f236f2622f5d81981f9d0a4057e92dd3a337e627891bd7" exitCode=0 Oct 02 02:00:07 crc kubenswrapper[4775]: I1002 02:00:07.183892 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-q2bd5" event={"ID":"8df5f08f-cb36-4414-adb4-9bd68db8a41c","Type":"ContainerDied","Data":"8b262fc92c0f625440f236f2622f5d81981f9d0a4057e92dd3a337e627891bd7"} Oct 02 02:00:07 crc kubenswrapper[4775]: I1002 02:00:07.185825 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-19ef-account-create-n7j6v" event={"ID":"c20e73b6-3188-49c6-836d-e518d93d9457","Type":"ContainerDied","Data":"9933d8b89bace5bc10ae63307f2d609d332d4120730b2f374434b02bd7840de3"} Oct 02 02:00:07 crc kubenswrapper[4775]: I1002 02:00:07.185845 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9933d8b89bace5bc10ae63307f2d609d332d4120730b2f374434b02bd7840de3" Oct 02 02:00:07 crc kubenswrapper[4775]: I1002 02:00:07.185887 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-19ef-account-create-n7j6v" Oct 02 02:00:07 crc kubenswrapper[4775]: I1002 02:00:07.190252 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-b963-account-create-zbgmn" Oct 02 02:00:07 crc kubenswrapper[4775]: I1002 02:00:07.191973 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-b963-account-create-zbgmn" event={"ID":"0b46397b-0c28-46ab-8f20-1426466202b3","Type":"ContainerDied","Data":"e2136fb29513cd7423fcc0c18a0886a38fdf4527c9fe2867441afda16b7516a4"} Oct 02 02:00:07 crc kubenswrapper[4775]: I1002 02:00:07.192012 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2136fb29513cd7423fcc0c18a0886a38fdf4527c9fe2867441afda16b7516a4" Oct 02 02:00:07 crc kubenswrapper[4775]: I1002 02:00:07.233301 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:00:07 crc kubenswrapper[4775]: I1002 02:00:07.233350 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:00:07 crc kubenswrapper[4775]: I1002 02:00:07.233390 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 02:00:07 crc kubenswrapper[4775]: I1002 02:00:07.234078 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"61c8c995ae4362ef1e859023c2c2ffc70bf9e259d47718d71624878bd8dc02ec"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:00:07 crc kubenswrapper[4775]: I1002 02:00:07.234343 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://61c8c995ae4362ef1e859023c2c2ffc70bf9e259d47718d71624878bd8dc02ec" gracePeriod=600 Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.201035 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="61c8c995ae4362ef1e859023c2c2ffc70bf9e259d47718d71624878bd8dc02ec" exitCode=0 Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.201206 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"61c8c995ae4362ef1e859023c2c2ffc70bf9e259d47718d71624878bd8dc02ec"} Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.201315 4775 scope.go:117] "RemoveContainer" containerID="8d8df09487b33326d5ae12f2261c6dc85c52809ba9bfd51eac55d13c8fc891e6" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.203209 4775 generic.go:334] "Generic (PLEG): container finished" podID="683f8a41-9042-4a9d-9fda-03784d72e926" containerID="6896915cc67aadd994ad499da8c5082fb11be528fb9225adc20db0b4e4bee028" exitCode=0 Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.204567 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hsh5j" event={"ID":"683f8a41-9042-4a9d-9fda-03784d72e926","Type":"ContainerDied","Data":"6896915cc67aadd994ad499da8c5082fb11be528fb9225adc20db0b4e4bee028"} Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.239756 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-dw9gs"] Oct 02 02:00:08 crc kubenswrapper[4775]: E1002 02:00:08.240240 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c20e73b6-3188-49c6-836d-e518d93d9457" containerName="mariadb-account-create" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.240254 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c20e73b6-3188-49c6-836d-e518d93d9457" containerName="mariadb-account-create" Oct 02 02:00:08 crc kubenswrapper[4775]: E1002 02:00:08.240285 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b46397b-0c28-46ab-8f20-1426466202b3" containerName="mariadb-account-create" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.240291 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b46397b-0c28-46ab-8f20-1426466202b3" containerName="mariadb-account-create" Oct 02 02:00:08 crc kubenswrapper[4775]: E1002 02:00:08.240307 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f436799-ffa1-45dc-9564-3fbb6b64fffe" containerName="init" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.240314 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f436799-ffa1-45dc-9564-3fbb6b64fffe" containerName="init" Oct 02 02:00:08 crc kubenswrapper[4775]: E1002 02:00:08.240321 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adbfa780-e6fe-402d-a040-175271a4b1e1" containerName="collect-profiles" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.240327 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="adbfa780-e6fe-402d-a040-175271a4b1e1" containerName="collect-profiles" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.240496 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f436799-ffa1-45dc-9564-3fbb6b64fffe" containerName="init" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.240507 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c20e73b6-3188-49c6-836d-e518d93d9457" containerName="mariadb-account-create" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.240522 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="adbfa780-e6fe-402d-a040-175271a4b1e1" containerName="collect-profiles" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.240533 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b46397b-0c28-46ab-8f20-1426466202b3" containerName="mariadb-account-create" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.241120 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dw9gs" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.244589 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-cwwrx" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.244792 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.247860 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dw9gs"] Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.390861 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d05cdd83-5605-46ed-88e8-168edbbcf7c1-db-sync-config-data\") pod \"barbican-db-sync-dw9gs\" (UID: \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\") " pod="openstack/barbican-db-sync-dw9gs" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.390924 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05cdd83-5605-46ed-88e8-168edbbcf7c1-combined-ca-bundle\") pod \"barbican-db-sync-dw9gs\" (UID: \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\") " pod="openstack/barbican-db-sync-dw9gs" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.390981 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh2gs\" (UniqueName: \"kubernetes.io/projected/d05cdd83-5605-46ed-88e8-168edbbcf7c1-kube-api-access-xh2gs\") pod \"barbican-db-sync-dw9gs\" (UID: \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\") " pod="openstack/barbican-db-sync-dw9gs" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.492349 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh2gs\" (UniqueName: \"kubernetes.io/projected/d05cdd83-5605-46ed-88e8-168edbbcf7c1-kube-api-access-xh2gs\") pod \"barbican-db-sync-dw9gs\" (UID: \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\") " pod="openstack/barbican-db-sync-dw9gs" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.492487 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d05cdd83-5605-46ed-88e8-168edbbcf7c1-db-sync-config-data\") pod \"barbican-db-sync-dw9gs\" (UID: \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\") " pod="openstack/barbican-db-sync-dw9gs" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.492544 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05cdd83-5605-46ed-88e8-168edbbcf7c1-combined-ca-bundle\") pod \"barbican-db-sync-dw9gs\" (UID: \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\") " pod="openstack/barbican-db-sync-dw9gs" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.513557 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d05cdd83-5605-46ed-88e8-168edbbcf7c1-db-sync-config-data\") pod \"barbican-db-sync-dw9gs\" (UID: \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\") " pod="openstack/barbican-db-sync-dw9gs" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.518706 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05cdd83-5605-46ed-88e8-168edbbcf7c1-combined-ca-bundle\") pod \"barbican-db-sync-dw9gs\" (UID: \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\") " pod="openstack/barbican-db-sync-dw9gs" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.542810 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh2gs\" (UniqueName: \"kubernetes.io/projected/d05cdd83-5605-46ed-88e8-168edbbcf7c1-kube-api-access-xh2gs\") pod \"barbican-db-sync-dw9gs\" (UID: \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\") " pod="openstack/barbican-db-sync-dw9gs" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.555483 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-hjvr2"] Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.557762 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hjvr2" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.561607 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dw9gs" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.563392 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qfnwq" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.563627 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.563761 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.595392 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-hjvr2"] Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.695855 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-combined-ca-bundle\") pod \"neutron-db-sync-hjvr2\" (UID: \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\") " pod="openstack/neutron-db-sync-hjvr2" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.696234 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-config\") pod \"neutron-db-sync-hjvr2\" (UID: \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\") " pod="openstack/neutron-db-sync-hjvr2" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.699152 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-696hf\" (UniqueName: \"kubernetes.io/projected/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-kube-api-access-696hf\") pod \"neutron-db-sync-hjvr2\" (UID: \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\") " pod="openstack/neutron-db-sync-hjvr2" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.801219 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-combined-ca-bundle\") pod \"neutron-db-sync-hjvr2\" (UID: \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\") " pod="openstack/neutron-db-sync-hjvr2" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.801289 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-config\") pod \"neutron-db-sync-hjvr2\" (UID: \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\") " pod="openstack/neutron-db-sync-hjvr2" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.801340 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-696hf\" (UniqueName: \"kubernetes.io/projected/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-kube-api-access-696hf\") pod \"neutron-db-sync-hjvr2\" (UID: \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\") " pod="openstack/neutron-db-sync-hjvr2" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.807731 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-combined-ca-bundle\") pod \"neutron-db-sync-hjvr2\" (UID: \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\") " pod="openstack/neutron-db-sync-hjvr2" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.809968 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-config\") pod \"neutron-db-sync-hjvr2\" (UID: \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\") " pod="openstack/neutron-db-sync-hjvr2" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.817285 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-696hf\" (UniqueName: \"kubernetes.io/projected/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-kube-api-access-696hf\") pod \"neutron-db-sync-hjvr2\" (UID: \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\") " pod="openstack/neutron-db-sync-hjvr2" Oct 02 02:00:08 crc kubenswrapper[4775]: I1002 02:00:08.926704 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hjvr2" Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.802658 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.822746 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-q2bd5" Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.928555 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-fernet-keys\") pod \"683f8a41-9042-4a9d-9fda-03784d72e926\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.928670 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-db-sync-config-data\") pod \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.928757 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-scripts\") pod \"683f8a41-9042-4a9d-9fda-03784d72e926\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.928821 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn9sd\" (UniqueName: \"kubernetes.io/projected/8df5f08f-cb36-4414-adb4-9bd68db8a41c-kube-api-access-sn9sd\") pod \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.928854 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-config-data\") pod \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.928897 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-config-data\") pod \"683f8a41-9042-4a9d-9fda-03784d72e926\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.928945 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-combined-ca-bundle\") pod \"683f8a41-9042-4a9d-9fda-03784d72e926\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.929034 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-credential-keys\") pod \"683f8a41-9042-4a9d-9fda-03784d72e926\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.929072 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-combined-ca-bundle\") pod \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\" (UID: \"8df5f08f-cb36-4414-adb4-9bd68db8a41c\") " Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.929098 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ntbg\" (UniqueName: \"kubernetes.io/projected/683f8a41-9042-4a9d-9fda-03784d72e926-kube-api-access-2ntbg\") pod \"683f8a41-9042-4a9d-9fda-03784d72e926\" (UID: \"683f8a41-9042-4a9d-9fda-03784d72e926\") " Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.934058 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/683f8a41-9042-4a9d-9fda-03784d72e926-kube-api-access-2ntbg" (OuterVolumeSpecName: "kube-api-access-2ntbg") pod "683f8a41-9042-4a9d-9fda-03784d72e926" (UID: "683f8a41-9042-4a9d-9fda-03784d72e926"). InnerVolumeSpecName "kube-api-access-2ntbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.941713 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "683f8a41-9042-4a9d-9fda-03784d72e926" (UID: "683f8a41-9042-4a9d-9fda-03784d72e926"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.941805 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "683f8a41-9042-4a9d-9fda-03784d72e926" (UID: "683f8a41-9042-4a9d-9fda-03784d72e926"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.942453 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8df5f08f-cb36-4414-adb4-9bd68db8a41c" (UID: "8df5f08f-cb36-4414-adb4-9bd68db8a41c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.955210 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8df5f08f-cb36-4414-adb4-9bd68db8a41c-kube-api-access-sn9sd" (OuterVolumeSpecName: "kube-api-access-sn9sd") pod "8df5f08f-cb36-4414-adb4-9bd68db8a41c" (UID: "8df5f08f-cb36-4414-adb4-9bd68db8a41c"). InnerVolumeSpecName "kube-api-access-sn9sd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.956186 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-scripts" (OuterVolumeSpecName: "scripts") pod "683f8a41-9042-4a9d-9fda-03784d72e926" (UID: "683f8a41-9042-4a9d-9fda-03784d72e926"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.962770 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-config-data" (OuterVolumeSpecName: "config-data") pod "683f8a41-9042-4a9d-9fda-03784d72e926" (UID: "683f8a41-9042-4a9d-9fda-03784d72e926"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.964906 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8df5f08f-cb36-4414-adb4-9bd68db8a41c" (UID: "8df5f08f-cb36-4414-adb4-9bd68db8a41c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.965396 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "683f8a41-9042-4a9d-9fda-03784d72e926" (UID: "683f8a41-9042-4a9d-9fda-03784d72e926"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:09 crc kubenswrapper[4775]: I1002 02:00:09.994072 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-config-data" (OuterVolumeSpecName: "config-data") pod "8df5f08f-cb36-4414-adb4-9bd68db8a41c" (UID: "8df5f08f-cb36-4414-adb4-9bd68db8a41c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.031213 4775 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.031250 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.031266 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ntbg\" (UniqueName: \"kubernetes.io/projected/683f8a41-9042-4a9d-9fda-03784d72e926-kube-api-access-2ntbg\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.031281 4775 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.031293 4775 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.031305 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.031316 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn9sd\" (UniqueName: \"kubernetes.io/projected/8df5f08f-cb36-4414-adb4-9bd68db8a41c-kube-api-access-sn9sd\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.031328 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8df5f08f-cb36-4414-adb4-9bd68db8a41c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.031337 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.031348 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/683f8a41-9042-4a9d-9fda-03784d72e926-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.228102 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-q2bd5" event={"ID":"8df5f08f-cb36-4414-adb4-9bd68db8a41c","Type":"ContainerDied","Data":"7bfbc9e9f62e78d6dbed72ae8e234ccdf7983c37e2518f17f4a28e4542086dcc"} Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.228148 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bfbc9e9f62e78d6dbed72ae8e234ccdf7983c37e2518f17f4a28e4542086dcc" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.228118 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-q2bd5" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.232302 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hsh5j" event={"ID":"683f8a41-9042-4a9d-9fda-03784d72e926","Type":"ContainerDied","Data":"1a90b262f1b37086adb48b75163d35f4e27a9b7d7998bb52c7ab1d40acf3c1f1"} Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.232351 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a90b262f1b37086adb48b75163d35f4e27a9b7d7998bb52c7ab1d40acf3c1f1" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.232425 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hsh5j" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.306165 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-hsh5j"] Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.309647 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-hsh5j"] Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.439981 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2smp7"] Oct 02 02:00:10 crc kubenswrapper[4775]: E1002 02:00:10.442894 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="683f8a41-9042-4a9d-9fda-03784d72e926" containerName="keystone-bootstrap" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.442931 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="683f8a41-9042-4a9d-9fda-03784d72e926" containerName="keystone-bootstrap" Oct 02 02:00:10 crc kubenswrapper[4775]: E1002 02:00:10.443035 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8df5f08f-cb36-4414-adb4-9bd68db8a41c" containerName="glance-db-sync" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.443042 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8df5f08f-cb36-4414-adb4-9bd68db8a41c" containerName="glance-db-sync" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.443863 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8df5f08f-cb36-4414-adb4-9bd68db8a41c" containerName="glance-db-sync" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.443894 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="683f8a41-9042-4a9d-9fda-03784d72e926" containerName="keystone-bootstrap" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.445635 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.448572 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-shncl" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.448741 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.448935 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.450513 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.457376 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2smp7"] Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.543935 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-config-data\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.544020 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-credential-keys\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.544053 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-combined-ca-bundle\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.544199 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-fernet-keys\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.544234 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-scripts\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.544261 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtq4m\" (UniqueName: \"kubernetes.io/projected/272c8e33-553c-4932-951d-e1944cffc9e7-kube-api-access-vtq4m\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.646782 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-fernet-keys\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.646865 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-scripts\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.646913 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtq4m\" (UniqueName: \"kubernetes.io/projected/272c8e33-553c-4932-951d-e1944cffc9e7-kube-api-access-vtq4m\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.646991 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-config-data\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.647010 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-credential-keys\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.647039 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-combined-ca-bundle\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.652308 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-fernet-keys\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.652767 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-config-data\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.653898 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-credential-keys\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.657467 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-combined-ca-bundle\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.664361 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-scripts\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.671001 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtq4m\" (UniqueName: \"kubernetes.io/projected/272c8e33-553c-4932-951d-e1944cffc9e7-kube-api-access-vtq4m\") pod \"keystone-bootstrap-2smp7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:10 crc kubenswrapper[4775]: I1002 02:00:10.767279 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.106992 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-skdqz"] Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.107239 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" podUID="babe38d7-4bf5-47f7-acb3-22161a9eac50" containerName="dnsmasq-dns" containerID="cri-o://526a7a5bc8b20b157e4b2789ec2fda0e947f785915e94610f2ddb5355f916123" gracePeriod=10 Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.108136 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.139815 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-bczkw"] Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.149971 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.157048 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-bczkw"] Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.261033 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.261070 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mchn\" (UniqueName: \"kubernetes.io/projected/9e78aa9b-9107-461c-b94f-c0f166b489b4-kube-api-access-4mchn\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.261091 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.261145 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.261164 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-config\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.261216 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.362829 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.362866 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mchn\" (UniqueName: \"kubernetes.io/projected/9e78aa9b-9107-461c-b94f-c0f166b489b4-kube-api-access-4mchn\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.362888 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.362941 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-config\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.362970 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.363010 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.364058 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.364328 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.364788 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-config\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.364973 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.368325 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.417873 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mchn\" (UniqueName: \"kubernetes.io/projected/9e78aa9b-9107-461c-b94f-c0f166b489b4-kube-api-access-4mchn\") pod \"dnsmasq-dns-8b5c85b87-bczkw\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: E1002 02:00:11.445507 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbabe38d7_4bf5_47f7_acb3_22161a9eac50.slice/crio-conmon-526a7a5bc8b20b157e4b2789ec2fda0e947f785915e94610f2ddb5355f916123.scope\": RecentStats: unable to find data in memory cache]" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.476852 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:11 crc kubenswrapper[4775]: I1002 02:00:11.774707 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="683f8a41-9042-4a9d-9fda-03784d72e926" path="/var/lib/kubelet/pods/683f8a41-9042-4a9d-9fda-03784d72e926/volumes" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.130145 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.132002 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.134205 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-5shrf" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.134460 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.134578 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.139548 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.246498 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.248422 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.251354 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.258876 4775 generic.go:334] "Generic (PLEG): container finished" podID="babe38d7-4bf5-47f7-acb3-22161a9eac50" containerID="526a7a5bc8b20b157e4b2789ec2fda0e947f785915e94610f2ddb5355f916123" exitCode=0 Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.258974 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" event={"ID":"babe38d7-4bf5-47f7-acb3-22161a9eac50","Type":"ContainerDied","Data":"526a7a5bc8b20b157e4b2789ec2fda0e947f785915e94610f2ddb5355f916123"} Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.264944 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.280708 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-scripts\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.280936 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2qms\" (UniqueName: \"kubernetes.io/projected/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-kube-api-access-z2qms\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.281071 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-logs\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.281185 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.281320 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.281470 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.281574 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-config-data\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.383738 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.383831 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.383862 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.383993 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/94dc1456-86ef-4000-85d1-5dee3eaca6aa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.384041 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.384068 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.384094 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.384120 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.384139 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94dc1456-86ef-4000-85d1-5dee3eaca6aa-logs\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.384169 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-config-data\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.384204 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xghn7\" (UniqueName: \"kubernetes.io/projected/94dc1456-86ef-4000-85d1-5dee3eaca6aa-kube-api-access-xghn7\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.384238 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-scripts\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.384252 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2qms\" (UniqueName: \"kubernetes.io/projected/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-kube-api-access-z2qms\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.384319 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-logs\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.384797 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-logs\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.385156 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.385209 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.391010 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.396008 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-scripts\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.403005 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-config-data\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.410601 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2qms\" (UniqueName: \"kubernetes.io/projected/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-kube-api-access-z2qms\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.440111 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.451630 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.488390 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.488447 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/94dc1456-86ef-4000-85d1-5dee3eaca6aa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.488489 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.488509 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.488527 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.488543 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94dc1456-86ef-4000-85d1-5dee3eaca6aa-logs\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.488579 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xghn7\" (UniqueName: \"kubernetes.io/projected/94dc1456-86ef-4000-85d1-5dee3eaca6aa-kube-api-access-xghn7\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.489507 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/94dc1456-86ef-4000-85d1-5dee3eaca6aa-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.490188 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.492076 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94dc1456-86ef-4000-85d1-5dee3eaca6aa-logs\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.495267 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-config-data\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.495525 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-scripts\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.498989 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.505581 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xghn7\" (UniqueName: \"kubernetes.io/projected/94dc1456-86ef-4000-85d1-5dee3eaca6aa-kube-api-access-xghn7\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.516941 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:12 crc kubenswrapper[4775]: I1002 02:00:12.580079 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:13 crc kubenswrapper[4775]: I1002 02:00:13.061056 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" podUID="babe38d7-4bf5-47f7-acb3-22161a9eac50" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.139:5353: connect: connection refused" Oct 02 02:00:14 crc kubenswrapper[4775]: I1002 02:00:14.113144 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:00:14 crc kubenswrapper[4775]: I1002 02:00:14.212550 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-hjvr2"] Oct 02 02:00:14 crc kubenswrapper[4775]: I1002 02:00:14.238677 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:00:21 crc kubenswrapper[4775]: W1002 02:00:21.750767 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc34b79a8_c071_4fff_8dbc_cab3601cc3f8.slice/crio-770927a770e1eae3d2c6eb6cc614d179524138eb7d78c21181fb7a76f2a356c6 WatchSource:0}: Error finding container 770927a770e1eae3d2c6eb6cc614d179524138eb7d78c21181fb7a76f2a356c6: Status 404 returned error can't find the container with id 770927a770e1eae3d2c6eb6cc614d179524138eb7d78c21181fb7a76f2a356c6 Oct 02 02:00:21 crc kubenswrapper[4775]: I1002 02:00:21.841380 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:21 crc kubenswrapper[4775]: I1002 02:00:21.963394 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfqg5\" (UniqueName: \"kubernetes.io/projected/babe38d7-4bf5-47f7-acb3-22161a9eac50-kube-api-access-hfqg5\") pod \"babe38d7-4bf5-47f7-acb3-22161a9eac50\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " Oct 02 02:00:21 crc kubenswrapper[4775]: I1002 02:00:21.963484 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-ovsdbserver-sb\") pod \"babe38d7-4bf5-47f7-acb3-22161a9eac50\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " Oct 02 02:00:21 crc kubenswrapper[4775]: I1002 02:00:21.963542 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-dns-svc\") pod \"babe38d7-4bf5-47f7-acb3-22161a9eac50\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " Oct 02 02:00:21 crc kubenswrapper[4775]: I1002 02:00:21.963568 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-ovsdbserver-nb\") pod \"babe38d7-4bf5-47f7-acb3-22161a9eac50\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " Oct 02 02:00:21 crc kubenswrapper[4775]: I1002 02:00:21.963609 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-dns-swift-storage-0\") pod \"babe38d7-4bf5-47f7-acb3-22161a9eac50\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " Oct 02 02:00:21 crc kubenswrapper[4775]: I1002 02:00:21.963715 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-config\") pod \"babe38d7-4bf5-47f7-acb3-22161a9eac50\" (UID: \"babe38d7-4bf5-47f7-acb3-22161a9eac50\") " Oct 02 02:00:21 crc kubenswrapper[4775]: I1002 02:00:21.971310 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/babe38d7-4bf5-47f7-acb3-22161a9eac50-kube-api-access-hfqg5" (OuterVolumeSpecName: "kube-api-access-hfqg5") pod "babe38d7-4bf5-47f7-acb3-22161a9eac50" (UID: "babe38d7-4bf5-47f7-acb3-22161a9eac50"). InnerVolumeSpecName "kube-api-access-hfqg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.016127 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "babe38d7-4bf5-47f7-acb3-22161a9eac50" (UID: "babe38d7-4bf5-47f7-acb3-22161a9eac50"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.024876 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "babe38d7-4bf5-47f7-acb3-22161a9eac50" (UID: "babe38d7-4bf5-47f7-acb3-22161a9eac50"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.027906 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "babe38d7-4bf5-47f7-acb3-22161a9eac50" (UID: "babe38d7-4bf5-47f7-acb3-22161a9eac50"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.046552 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-config" (OuterVolumeSpecName: "config") pod "babe38d7-4bf5-47f7-acb3-22161a9eac50" (UID: "babe38d7-4bf5-47f7-acb3-22161a9eac50"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.051019 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "babe38d7-4bf5-47f7-acb3-22161a9eac50" (UID: "babe38d7-4bf5-47f7-acb3-22161a9eac50"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.065397 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.065435 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfqg5\" (UniqueName: \"kubernetes.io/projected/babe38d7-4bf5-47f7-acb3-22161a9eac50-kube-api-access-hfqg5\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.065446 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.065454 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.065464 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.065472 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/babe38d7-4bf5-47f7-acb3-22161a9eac50-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.409671 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" event={"ID":"babe38d7-4bf5-47f7-acb3-22161a9eac50","Type":"ContainerDied","Data":"de7cc4d85f28ebbb4dbd510296b1659cb218985ba9340aed81a0a9f12ea3da33"} Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.409736 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.409809 4775 scope.go:117] "RemoveContainer" containerID="526a7a5bc8b20b157e4b2789ec2fda0e947f785915e94610f2ddb5355f916123" Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.426237 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hjvr2" event={"ID":"c34b79a8-c071-4fff-8dbc-cab3601cc3f8","Type":"ContainerStarted","Data":"770927a770e1eae3d2c6eb6cc614d179524138eb7d78c21181fb7a76f2a356c6"} Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.439236 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-skdqz"] Oct 02 02:00:22 crc kubenswrapper[4775]: I1002 02:00:22.448812 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-76fcf4b695-skdqz"] Oct 02 02:00:23 crc kubenswrapper[4775]: I1002 02:00:23.061092 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-76fcf4b695-skdqz" podUID="babe38d7-4bf5-47f7-acb3-22161a9eac50" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.139:5353: i/o timeout" Oct 02 02:00:23 crc kubenswrapper[4775]: E1002 02:00:23.118733 4775 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 02 02:00:23 crc kubenswrapper[4775]: E1002 02:00:23.119550 4775 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xzddh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-x8pvr_openstack(9a3f9c1f-1298-4835-9935-56b808aae84d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 02 02:00:23 crc kubenswrapper[4775]: E1002 02:00:23.121284 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-x8pvr" podUID="9a3f9c1f-1298-4835-9935-56b808aae84d" Oct 02 02:00:23 crc kubenswrapper[4775]: I1002 02:00:23.154634 4775 scope.go:117] "RemoveContainer" containerID="032b608b1a7f0df7db55c0d674a323620ff4dceb9728680abc2e92f995715480" Oct 02 02:00:23 crc kubenswrapper[4775]: E1002 02:00:23.435568 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-x8pvr" podUID="9a3f9c1f-1298-4835-9935-56b808aae84d" Oct 02 02:00:23 crc kubenswrapper[4775]: I1002 02:00:23.664853 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2smp7"] Oct 02 02:00:23 crc kubenswrapper[4775]: I1002 02:00:23.671217 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-dw9gs"] Oct 02 02:00:23 crc kubenswrapper[4775]: W1002 02:00:23.677396 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd05cdd83_5605_46ed_88e8_168edbbcf7c1.slice/crio-2a660c8f302edfa8723a629dde6891ac96c10e9f88b1ba6d5e926694b4cbe3c5 WatchSource:0}: Error finding container 2a660c8f302edfa8723a629dde6891ac96c10e9f88b1ba6d5e926694b4cbe3c5: Status 404 returned error can't find the container with id 2a660c8f302edfa8723a629dde6891ac96c10e9f88b1ba6d5e926694b4cbe3c5 Oct 02 02:00:23 crc kubenswrapper[4775]: I1002 02:00:23.759660 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:00:23 crc kubenswrapper[4775]: I1002 02:00:23.795812 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="babe38d7-4bf5-47f7-acb3-22161a9eac50" path="/var/lib/kubelet/pods/babe38d7-4bf5-47f7-acb3-22161a9eac50/volumes" Oct 02 02:00:23 crc kubenswrapper[4775]: I1002 02:00:23.846461 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-bczkw"] Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.447218 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2smp7" event={"ID":"272c8e33-553c-4932-951d-e1944cffc9e7","Type":"ContainerStarted","Data":"f03365d9159046e89cfea91c9f68ea0948f70fcf7620f4ce9ed01400f90dc48d"} Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.447674 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2smp7" event={"ID":"272c8e33-553c-4932-951d-e1944cffc9e7","Type":"ContainerStarted","Data":"974ef4bc229645b36a33468709ef973675ff7d7f82bec6a0c291c4d9f7f09ad5"} Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.449793 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7","Type":"ContainerStarted","Data":"10331c02da26bcaf3cb689b3b182c8340f83075a4cef4bbe4c7347441ef53283"} Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.458321 4775 generic.go:334] "Generic (PLEG): container finished" podID="9e78aa9b-9107-461c-b94f-c0f166b489b4" containerID="fb3313fc8fe1d8fb265ea0d218297b58a6c957b9693cc2db9285b5a3bdd5aa98" exitCode=0 Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.458423 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" event={"ID":"9e78aa9b-9107-461c-b94f-c0f166b489b4","Type":"ContainerDied","Data":"fb3313fc8fe1d8fb265ea0d218297b58a6c957b9693cc2db9285b5a3bdd5aa98"} Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.458454 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" event={"ID":"9e78aa9b-9107-461c-b94f-c0f166b489b4","Type":"ContainerStarted","Data":"1cde4bf949e3ada68147efe3390db54f06c28389d308e1852c50298736a76357"} Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.470570 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2smp7" podStartSLOduration=14.470544561 podStartE2EDuration="14.470544561s" podCreationTimestamp="2025-10-02 02:00:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:24.466834465 +0000 UTC m=+1161.633578505" watchObservedRunningTime="2025-10-02 02:00:24.470544561 +0000 UTC m=+1161.637288611" Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.471668 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"08d56b32ae1f08ce86ec95ee88e05d04f2a7f9064777b463fc71e568251c9583"} Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.477741 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dw9gs" event={"ID":"d05cdd83-5605-46ed-88e8-168edbbcf7c1","Type":"ContainerStarted","Data":"2a660c8f302edfa8723a629dde6891ac96c10e9f88b1ba6d5e926694b4cbe3c5"} Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.481598 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hdtb5" event={"ID":"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9","Type":"ContainerStarted","Data":"1cde1b09d6db770bfcaa3f911697783619cbc8a288399077a17545ecd5108f0f"} Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.503404 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12864b75-dbdc-4075-aea7-9d3cc2a6ca59","Type":"ContainerStarted","Data":"2d8a19e074a59869fa182becc84c8c6153a9d08b9d345d4f7d9e637c21100594"} Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.509009 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hjvr2" event={"ID":"c34b79a8-c071-4fff-8dbc-cab3601cc3f8","Type":"ContainerStarted","Data":"3fc5ef432fab421b40f1e30656d3ab99d83e495569dde9d28576177506951a8b"} Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.524227 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-hdtb5" podStartSLOduration=3.446631465 podStartE2EDuration="22.524211851s" podCreationTimestamp="2025-10-02 02:00:02 +0000 UTC" firstStartedPulling="2025-10-02 02:00:04.018852745 +0000 UTC m=+1141.185596785" lastFinishedPulling="2025-10-02 02:00:23.096433091 +0000 UTC m=+1160.263177171" observedRunningTime="2025-10-02 02:00:24.523524024 +0000 UTC m=+1161.690268064" watchObservedRunningTime="2025-10-02 02:00:24.524211851 +0000 UTC m=+1161.690955891" Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.559894 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-hjvr2" podStartSLOduration=16.559873019 podStartE2EDuration="16.559873019s" podCreationTimestamp="2025-10-02 02:00:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:24.540275515 +0000 UTC m=+1161.707019555" watchObservedRunningTime="2025-10-02 02:00:24.559873019 +0000 UTC m=+1161.726617059" Oct 02 02:00:24 crc kubenswrapper[4775]: I1002 02:00:24.802177 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:00:25 crc kubenswrapper[4775]: W1002 02:00:25.181159 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94dc1456_86ef_4000_85d1_5dee3eaca6aa.slice/crio-63fe9cfbea498a0f154a4f232049ec42aaeb08376fa0f0f51a165d907b182237 WatchSource:0}: Error finding container 63fe9cfbea498a0f154a4f232049ec42aaeb08376fa0f0f51a165d907b182237: Status 404 returned error can't find the container with id 63fe9cfbea498a0f154a4f232049ec42aaeb08376fa0f0f51a165d907b182237 Oct 02 02:00:25 crc kubenswrapper[4775]: I1002 02:00:25.528469 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7","Type":"ContainerStarted","Data":"cb4f7abc08a424f9e6df5fba5262981be11dfb06ec281400221a16be4dcc797a"} Oct 02 02:00:25 crc kubenswrapper[4775]: I1002 02:00:25.533913 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"94dc1456-86ef-4000-85d1-5dee3eaca6aa","Type":"ContainerStarted","Data":"63fe9cfbea498a0f154a4f232049ec42aaeb08376fa0f0f51a165d907b182237"} Oct 02 02:00:26 crc kubenswrapper[4775]: I1002 02:00:26.548293 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" containerName="glance-log" containerID="cri-o://cb4f7abc08a424f9e6df5fba5262981be11dfb06ec281400221a16be4dcc797a" gracePeriod=30 Oct 02 02:00:26 crc kubenswrapper[4775]: I1002 02:00:26.548284 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7","Type":"ContainerStarted","Data":"55f965f44337a0a5e312c666bc56a01c4c562e7553be4d40c1d84f16fddbcaba"} Oct 02 02:00:26 crc kubenswrapper[4775]: I1002 02:00:26.548876 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" containerName="glance-httpd" containerID="cri-o://55f965f44337a0a5e312c666bc56a01c4c562e7553be4d40c1d84f16fddbcaba" gracePeriod=30 Oct 02 02:00:26 crc kubenswrapper[4775]: I1002 02:00:26.550904 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"94dc1456-86ef-4000-85d1-5dee3eaca6aa","Type":"ContainerStarted","Data":"399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765"} Oct 02 02:00:26 crc kubenswrapper[4775]: I1002 02:00:26.554784 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" event={"ID":"9e78aa9b-9107-461c-b94f-c0f166b489b4","Type":"ContainerStarted","Data":"10a348acd950eaad46c525129402d791195a91d90d80d01aa576064a8aeec468"} Oct 02 02:00:26 crc kubenswrapper[4775]: I1002 02:00:26.554916 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:26 crc kubenswrapper[4775]: I1002 02:00:26.569267 4775 generic.go:334] "Generic (PLEG): container finished" podID="fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9" containerID="1cde1b09d6db770bfcaa3f911697783619cbc8a288399077a17545ecd5108f0f" exitCode=0 Oct 02 02:00:26 crc kubenswrapper[4775]: I1002 02:00:26.569422 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hdtb5" event={"ID":"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9","Type":"ContainerDied","Data":"1cde1b09d6db770bfcaa3f911697783619cbc8a288399077a17545ecd5108f0f"} Oct 02 02:00:26 crc kubenswrapper[4775]: I1002 02:00:26.574453 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=15.574434474 podStartE2EDuration="15.574434474s" podCreationTimestamp="2025-10-02 02:00:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:26.572824532 +0000 UTC m=+1163.739568572" watchObservedRunningTime="2025-10-02 02:00:26.574434474 +0000 UTC m=+1163.741178514" Oct 02 02:00:26 crc kubenswrapper[4775]: I1002 02:00:26.581088 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12864b75-dbdc-4075-aea7-9d3cc2a6ca59","Type":"ContainerStarted","Data":"89a042d772657ba277e82d383099114e18ac75f765966f3219b0c0ca7ba4fc91"} Oct 02 02:00:26 crc kubenswrapper[4775]: I1002 02:00:26.621598 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" podStartSLOduration=15.621582097 podStartE2EDuration="15.621582097s" podCreationTimestamp="2025-10-02 02:00:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:26.611815545 +0000 UTC m=+1163.778559585" watchObservedRunningTime="2025-10-02 02:00:26.621582097 +0000 UTC m=+1163.788326127" Oct 02 02:00:27 crc kubenswrapper[4775]: I1002 02:00:27.592831 4775 generic.go:334] "Generic (PLEG): container finished" podID="574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" containerID="55f965f44337a0a5e312c666bc56a01c4c562e7553be4d40c1d84f16fddbcaba" exitCode=0 Oct 02 02:00:27 crc kubenswrapper[4775]: I1002 02:00:27.592860 4775 generic.go:334] "Generic (PLEG): container finished" podID="574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" containerID="cb4f7abc08a424f9e6df5fba5262981be11dfb06ec281400221a16be4dcc797a" exitCode=143 Oct 02 02:00:27 crc kubenswrapper[4775]: I1002 02:00:27.592904 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7","Type":"ContainerDied","Data":"55f965f44337a0a5e312c666bc56a01c4c562e7553be4d40c1d84f16fddbcaba"} Oct 02 02:00:27 crc kubenswrapper[4775]: I1002 02:00:27.592945 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7","Type":"ContainerDied","Data":"cb4f7abc08a424f9e6df5fba5262981be11dfb06ec281400221a16be4dcc797a"} Oct 02 02:00:27 crc kubenswrapper[4775]: I1002 02:00:27.597396 4775 generic.go:334] "Generic (PLEG): container finished" podID="272c8e33-553c-4932-951d-e1944cffc9e7" containerID="f03365d9159046e89cfea91c9f68ea0948f70fcf7620f4ce9ed01400f90dc48d" exitCode=0 Oct 02 02:00:27 crc kubenswrapper[4775]: I1002 02:00:27.597437 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2smp7" event={"ID":"272c8e33-553c-4932-951d-e1944cffc9e7","Type":"ContainerDied","Data":"f03365d9159046e89cfea91c9f68ea0948f70fcf7620f4ce9ed01400f90dc48d"} Oct 02 02:00:28 crc kubenswrapper[4775]: I1002 02:00:28.914899 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:28 crc kubenswrapper[4775]: I1002 02:00:28.965436 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-logs\") pod \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " Oct 02 02:00:28 crc kubenswrapper[4775]: I1002 02:00:28.965495 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-combined-ca-bundle\") pod \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " Oct 02 02:00:28 crc kubenswrapper[4775]: I1002 02:00:28.965558 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-scripts\") pod \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " Oct 02 02:00:28 crc kubenswrapper[4775]: I1002 02:00:28.965579 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-config-data\") pod \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " Oct 02 02:00:28 crc kubenswrapper[4775]: I1002 02:00:28.965614 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njjkh\" (UniqueName: \"kubernetes.io/projected/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-kube-api-access-njjkh\") pod \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\" (UID: \"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9\") " Oct 02 02:00:28 crc kubenswrapper[4775]: I1002 02:00:28.967035 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-logs" (OuterVolumeSpecName: "logs") pod "fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9" (UID: "fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:28 crc kubenswrapper[4775]: I1002 02:00:28.972358 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-scripts" (OuterVolumeSpecName: "scripts") pod "fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9" (UID: "fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:28 crc kubenswrapper[4775]: I1002 02:00:28.972533 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-kube-api-access-njjkh" (OuterVolumeSpecName: "kube-api-access-njjkh") pod "fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9" (UID: "fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9"). InnerVolumeSpecName "kube-api-access-njjkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:28 crc kubenswrapper[4775]: I1002 02:00:28.998882 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-config-data" (OuterVolumeSpecName: "config-data") pod "fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9" (UID: "fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.006633 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9" (UID: "fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.017081 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.070420 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.070654 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.070665 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njjkh\" (UniqueName: \"kubernetes.io/projected/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-kube-api-access-njjkh\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.070676 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.070684 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.172037 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-config-data\") pod \"272c8e33-553c-4932-951d-e1944cffc9e7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.172186 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-credential-keys\") pod \"272c8e33-553c-4932-951d-e1944cffc9e7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.172347 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-scripts\") pod \"272c8e33-553c-4932-951d-e1944cffc9e7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.172417 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtq4m\" (UniqueName: \"kubernetes.io/projected/272c8e33-553c-4932-951d-e1944cffc9e7-kube-api-access-vtq4m\") pod \"272c8e33-553c-4932-951d-e1944cffc9e7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.172467 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-fernet-keys\") pod \"272c8e33-553c-4932-951d-e1944cffc9e7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.172485 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-combined-ca-bundle\") pod \"272c8e33-553c-4932-951d-e1944cffc9e7\" (UID: \"272c8e33-553c-4932-951d-e1944cffc9e7\") " Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.180605 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/272c8e33-553c-4932-951d-e1944cffc9e7-kube-api-access-vtq4m" (OuterVolumeSpecName: "kube-api-access-vtq4m") pod "272c8e33-553c-4932-951d-e1944cffc9e7" (UID: "272c8e33-553c-4932-951d-e1944cffc9e7"). InnerVolumeSpecName "kube-api-access-vtq4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.180883 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-scripts" (OuterVolumeSpecName: "scripts") pod "272c8e33-553c-4932-951d-e1944cffc9e7" (UID: "272c8e33-553c-4932-951d-e1944cffc9e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.184070 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "272c8e33-553c-4932-951d-e1944cffc9e7" (UID: "272c8e33-553c-4932-951d-e1944cffc9e7"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.189925 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "272c8e33-553c-4932-951d-e1944cffc9e7" (UID: "272c8e33-553c-4932-951d-e1944cffc9e7"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.228392 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "272c8e33-553c-4932-951d-e1944cffc9e7" (UID: "272c8e33-553c-4932-951d-e1944cffc9e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.241182 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-config-data" (OuterVolumeSpecName: "config-data") pod "272c8e33-553c-4932-951d-e1944cffc9e7" (UID: "272c8e33-553c-4932-951d-e1944cffc9e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.274443 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtq4m\" (UniqueName: \"kubernetes.io/projected/272c8e33-553c-4932-951d-e1944cffc9e7-kube-api-access-vtq4m\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.274474 4775 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.274484 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.274491 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.274502 4775 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.274510 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/272c8e33-553c-4932-951d-e1944cffc9e7-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.343302 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.375064 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.375106 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-scripts\") pod \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.375140 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-httpd-run\") pod \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.375205 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2qms\" (UniqueName: \"kubernetes.io/projected/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-kube-api-access-z2qms\") pod \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.375221 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-combined-ca-bundle\") pod \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.375241 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-config-data\") pod \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.375256 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-logs\") pod \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\" (UID: \"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7\") " Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.375614 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" (UID: "574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.375696 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-logs" (OuterVolumeSpecName: "logs") pod "574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" (UID: "574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.379573 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-scripts" (OuterVolumeSpecName: "scripts") pod "574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" (UID: "574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.379650 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-kube-api-access-z2qms" (OuterVolumeSpecName: "kube-api-access-z2qms") pod "574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" (UID: "574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7"). InnerVolumeSpecName "kube-api-access-z2qms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.379709 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" (UID: "574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.398728 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" (UID: "574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.427105 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-config-data" (OuterVolumeSpecName: "config-data") pod "574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" (UID: "574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.485079 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.485301 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.485311 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.485343 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2qms\" (UniqueName: \"kubernetes.io/projected/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-kube-api-access-z2qms\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.485355 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.485383 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.485392 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.496500 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.586523 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.618688 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hdtb5" event={"ID":"fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9","Type":"ContainerDied","Data":"234fdba0bdee12badc1d90753ddb54c667013fa32d8a960666607fd72db604d8"} Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.618725 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="234fdba0bdee12badc1d90753ddb54c667013fa32d8a960666607fd72db604d8" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.618771 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hdtb5" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.621820 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2smp7" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.621841 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2smp7" event={"ID":"272c8e33-553c-4932-951d-e1944cffc9e7","Type":"ContainerDied","Data":"974ef4bc229645b36a33468709ef973675ff7d7f82bec6a0c291c4d9f7f09ad5"} Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.621872 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="974ef4bc229645b36a33468709ef973675ff7d7f82bec6a0c291c4d9f7f09ad5" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.624284 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.624329 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7","Type":"ContainerDied","Data":"10331c02da26bcaf3cb689b3b182c8340f83075a4cef4bbe4c7347441ef53283"} Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.624385 4775 scope.go:117] "RemoveContainer" containerID="55f965f44337a0a5e312c666bc56a01c4c562e7553be4d40c1d84f16fddbcaba" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.634901 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"94dc1456-86ef-4000-85d1-5dee3eaca6aa","Type":"ContainerStarted","Data":"b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3"} Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.635103 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="94dc1456-86ef-4000-85d1-5dee3eaca6aa" containerName="glance-log" containerID="cri-o://399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765" gracePeriod=30 Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.635441 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="94dc1456-86ef-4000-85d1-5dee3eaca6aa" containerName="glance-httpd" containerID="cri-o://b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3" gracePeriod=30 Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.641945 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dw9gs" event={"ID":"d05cdd83-5605-46ed-88e8-168edbbcf7c1","Type":"ContainerStarted","Data":"63635a39be96c75e16a0cd5c161326889020e66580536e869bd6d92c322a7bb6"} Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.673527 4775 scope.go:117] "RemoveContainer" containerID="cb4f7abc08a424f9e6df5fba5262981be11dfb06ec281400221a16be4dcc797a" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.703043 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=18.703028118 podStartE2EDuration="18.703028118s" podCreationTimestamp="2025-10-02 02:00:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:29.667323819 +0000 UTC m=+1166.834067889" watchObservedRunningTime="2025-10-02 02:00:29.703028118 +0000 UTC m=+1166.869772158" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.717184 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-dw9gs" podStartSLOduration=16.427569764 podStartE2EDuration="21.717164421s" podCreationTimestamp="2025-10-02 02:00:08 +0000 UTC" firstStartedPulling="2025-10-02 02:00:23.68182351 +0000 UTC m=+1160.848567550" lastFinishedPulling="2025-10-02 02:00:28.971418167 +0000 UTC m=+1166.138162207" observedRunningTime="2025-10-02 02:00:29.689658214 +0000 UTC m=+1166.856402254" watchObservedRunningTime="2025-10-02 02:00:29.717164421 +0000 UTC m=+1166.883908461" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.731452 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.737739 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-564487c87-zppvw"] Oct 02 02:00:29 crc kubenswrapper[4775]: E1002 02:00:29.738155 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9" containerName="placement-db-sync" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.738169 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9" containerName="placement-db-sync" Oct 02 02:00:29 crc kubenswrapper[4775]: E1002 02:00:29.738195 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="babe38d7-4bf5-47f7-acb3-22161a9eac50" containerName="dnsmasq-dns" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.738201 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="babe38d7-4bf5-47f7-acb3-22161a9eac50" containerName="dnsmasq-dns" Oct 02 02:00:29 crc kubenswrapper[4775]: E1002 02:00:29.738214 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" containerName="glance-log" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.738223 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" containerName="glance-log" Oct 02 02:00:29 crc kubenswrapper[4775]: E1002 02:00:29.738237 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="babe38d7-4bf5-47f7-acb3-22161a9eac50" containerName="init" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.738242 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="babe38d7-4bf5-47f7-acb3-22161a9eac50" containerName="init" Oct 02 02:00:29 crc kubenswrapper[4775]: E1002 02:00:29.738251 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="272c8e33-553c-4932-951d-e1944cffc9e7" containerName="keystone-bootstrap" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.738258 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="272c8e33-553c-4932-951d-e1944cffc9e7" containerName="keystone-bootstrap" Oct 02 02:00:29 crc kubenswrapper[4775]: E1002 02:00:29.738268 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" containerName="glance-httpd" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.738274 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" containerName="glance-httpd" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.738432 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" containerName="glance-log" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.738443 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9" containerName="placement-db-sync" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.738458 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="babe38d7-4bf5-47f7-acb3-22161a9eac50" containerName="dnsmasq-dns" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.738469 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="272c8e33-553c-4932-951d-e1944cffc9e7" containerName="keystone-bootstrap" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.738480 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" containerName="glance-httpd" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.739041 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.747937 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.785786 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.786434 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.786826 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-shncl" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.787324 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.789231 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.789390 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.801946 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-scripts\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.802015 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-public-tls-certs\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.802172 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-internal-tls-certs\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.802203 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-credential-keys\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.802222 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-config-data\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.802243 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lpv5\" (UniqueName: \"kubernetes.io/projected/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-kube-api-access-7lpv5\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.802303 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-fernet-keys\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.802333 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-combined-ca-bundle\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.822695 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7" path="/var/lib/kubelet/pods/574e8e63-8c5e-454c-b6f9-d2a3f55d7cc7/volumes" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.823529 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-564487c87-zppvw"] Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.823553 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.824924 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.825575 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.827598 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.829090 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.903524 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-scripts\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.903578 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-public-tls-certs\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.903610 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.903682 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-scripts\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.903705 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-config-data\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.903725 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.903825 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.903850 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.903867 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-internal-tls-certs\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.903883 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-credential-keys\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.903900 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-config-data\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.903996 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lpv5\" (UniqueName: \"kubernetes.io/projected/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-kube-api-access-7lpv5\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.904046 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkq55\" (UniqueName: \"kubernetes.io/projected/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-kube-api-access-lkq55\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.904107 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-fernet-keys\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.904138 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-combined-ca-bundle\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.904193 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-logs\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.909923 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-public-tls-certs\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.911395 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-fernet-keys\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.911518 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-combined-ca-bundle\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.911838 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-internal-tls-certs\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.912766 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-scripts\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.914272 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-credential-keys\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.914786 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-config-data\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:29 crc kubenswrapper[4775]: I1002 02:00:29.921233 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lpv5\" (UniqueName: \"kubernetes.io/projected/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-kube-api-access-7lpv5\") pod \"keystone-564487c87-zppvw\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.005822 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkq55\" (UniqueName: \"kubernetes.io/projected/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-kube-api-access-lkq55\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.006125 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-logs\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.006190 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.006215 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-scripts\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.006232 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-config-data\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.006252 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.006287 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.006311 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.007716 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.008014 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-logs\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.008739 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.012264 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.012292 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-config-data\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.012620 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.023998 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-scripts\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.026465 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkq55\" (UniqueName: \"kubernetes.io/projected/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-kube-api-access-lkq55\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.045180 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.133283 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-777987c7d4-d5vzz"] Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.135406 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.144180 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.144440 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.144304 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.144639 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-5smvp" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.145637 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.159546 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.172886 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.191067 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-777987c7d4-d5vzz"] Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.209568 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-public-tls-certs\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.209704 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-scripts\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.209781 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56813183-3dfc-430c-aaac-bda27d91340b-logs\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.209853 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-internal-tls-certs\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.209922 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-combined-ca-bundle\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.210011 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb6zc\" (UniqueName: \"kubernetes.io/projected/56813183-3dfc-430c-aaac-bda27d91340b-kube-api-access-cb6zc\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.210125 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-config-data\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.311350 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-combined-ca-bundle\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.311413 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb6zc\" (UniqueName: \"kubernetes.io/projected/56813183-3dfc-430c-aaac-bda27d91340b-kube-api-access-cb6zc\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.311505 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-config-data\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.311566 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-public-tls-certs\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.311613 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-scripts\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.311644 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56813183-3dfc-430c-aaac-bda27d91340b-logs\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.311673 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-internal-tls-certs\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.323460 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56813183-3dfc-430c-aaac-bda27d91340b-logs\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.337595 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-config-data\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.340217 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-internal-tls-certs\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.373365 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-scripts\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.380629 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb6zc\" (UniqueName: \"kubernetes.io/projected/56813183-3dfc-430c-aaac-bda27d91340b-kube-api-access-cb6zc\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.383677 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-public-tls-certs\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.387655 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-combined-ca-bundle\") pod \"placement-777987c7d4-d5vzz\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.560031 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.574264 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.617195 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-combined-ca-bundle\") pod \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.617276 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94dc1456-86ef-4000-85d1-5dee3eaca6aa-logs\") pod \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.617304 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-config-data\") pod \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.617326 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xghn7\" (UniqueName: \"kubernetes.io/projected/94dc1456-86ef-4000-85d1-5dee3eaca6aa-kube-api-access-xghn7\") pod \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.617348 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/94dc1456-86ef-4000-85d1-5dee3eaca6aa-httpd-run\") pod \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.617368 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-scripts\") pod \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.617383 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\" (UID: \"94dc1456-86ef-4000-85d1-5dee3eaca6aa\") " Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.618919 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94dc1456-86ef-4000-85d1-5dee3eaca6aa-logs" (OuterVolumeSpecName: "logs") pod "94dc1456-86ef-4000-85d1-5dee3eaca6aa" (UID: "94dc1456-86ef-4000-85d1-5dee3eaca6aa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.619157 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94dc1456-86ef-4000-85d1-5dee3eaca6aa-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "94dc1456-86ef-4000-85d1-5dee3eaca6aa" (UID: "94dc1456-86ef-4000-85d1-5dee3eaca6aa"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.624789 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "94dc1456-86ef-4000-85d1-5dee3eaca6aa" (UID: "94dc1456-86ef-4000-85d1-5dee3eaca6aa"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.632465 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94dc1456-86ef-4000-85d1-5dee3eaca6aa-kube-api-access-xghn7" (OuterVolumeSpecName: "kube-api-access-xghn7") pod "94dc1456-86ef-4000-85d1-5dee3eaca6aa" (UID: "94dc1456-86ef-4000-85d1-5dee3eaca6aa"). InnerVolumeSpecName "kube-api-access-xghn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.639182 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-scripts" (OuterVolumeSpecName: "scripts") pod "94dc1456-86ef-4000-85d1-5dee3eaca6aa" (UID: "94dc1456-86ef-4000-85d1-5dee3eaca6aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.675239 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94dc1456-86ef-4000-85d1-5dee3eaca6aa" (UID: "94dc1456-86ef-4000-85d1-5dee3eaca6aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.698979 4775 generic.go:334] "Generic (PLEG): container finished" podID="94dc1456-86ef-4000-85d1-5dee3eaca6aa" containerID="b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3" exitCode=143 Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.699022 4775 generic.go:334] "Generic (PLEG): container finished" podID="94dc1456-86ef-4000-85d1-5dee3eaca6aa" containerID="399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765" exitCode=143 Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.699775 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.700200 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"94dc1456-86ef-4000-85d1-5dee3eaca6aa","Type":"ContainerDied","Data":"b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3"} Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.700233 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"94dc1456-86ef-4000-85d1-5dee3eaca6aa","Type":"ContainerDied","Data":"399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765"} Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.700244 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"94dc1456-86ef-4000-85d1-5dee3eaca6aa","Type":"ContainerDied","Data":"63fe9cfbea498a0f154a4f232049ec42aaeb08376fa0f0f51a165d907b182237"} Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.700258 4775 scope.go:117] "RemoveContainer" containerID="b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.723625 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94dc1456-86ef-4000-85d1-5dee3eaca6aa-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.723760 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xghn7\" (UniqueName: \"kubernetes.io/projected/94dc1456-86ef-4000-85d1-5dee3eaca6aa-kube-api-access-xghn7\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.723818 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/94dc1456-86ef-4000-85d1-5dee3eaca6aa-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.723940 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.724033 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.724224 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.743343 4775 scope.go:117] "RemoveContainer" containerID="399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.745045 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-config-data" (OuterVolumeSpecName: "config-data") pod "94dc1456-86ef-4000-85d1-5dee3eaca6aa" (UID: "94dc1456-86ef-4000-85d1-5dee3eaca6aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.745564 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.827807 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94dc1456-86ef-4000-85d1-5dee3eaca6aa-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.827865 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:30 crc kubenswrapper[4775]: I1002 02:00:30.967745 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-564487c87-zppvw"] Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.030033 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.039815 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.044672 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.069990 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:00:31 crc kubenswrapper[4775]: E1002 02:00:31.070394 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94dc1456-86ef-4000-85d1-5dee3eaca6aa" containerName="glance-httpd" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.070419 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="94dc1456-86ef-4000-85d1-5dee3eaca6aa" containerName="glance-httpd" Oct 02 02:00:31 crc kubenswrapper[4775]: E1002 02:00:31.070450 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94dc1456-86ef-4000-85d1-5dee3eaca6aa" containerName="glance-log" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.070457 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="94dc1456-86ef-4000-85d1-5dee3eaca6aa" containerName="glance-log" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.070619 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="94dc1456-86ef-4000-85d1-5dee3eaca6aa" containerName="glance-httpd" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.070637 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="94dc1456-86ef-4000-85d1-5dee3eaca6aa" containerName="glance-log" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.071427 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.071509 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.076326 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.076479 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.152678 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-777987c7d4-d5vzz"] Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.233162 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hkdv\" (UniqueName: \"kubernetes.io/projected/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-kube-api-access-4hkdv\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.233240 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-logs\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.233280 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.233340 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.233404 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.233439 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.233470 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.233694 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.335287 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.335365 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.335413 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.335508 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hkdv\" (UniqueName: \"kubernetes.io/projected/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-kube-api-access-4hkdv\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.335532 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-logs\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.335574 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.335601 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.335644 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.336175 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.336182 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.336450 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-logs\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.340844 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.341548 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.342023 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.353102 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.357745 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hkdv\" (UniqueName: \"kubernetes.io/projected/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-kube-api-access-4hkdv\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.375470 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.461218 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.478165 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.530488 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-nx6tz"] Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.530786 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" podUID="11558ec3-d684-403c-9afe-f08cb76fb18a" containerName="dnsmasq-dns" containerID="cri-o://372a057f47923d18519533578d61e7d8f697805ca65253b9dde7c9953336d300" gracePeriod=10 Oct 02 02:00:31 crc kubenswrapper[4775]: I1002 02:00:31.773871 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94dc1456-86ef-4000-85d1-5dee3eaca6aa" path="/var/lib/kubelet/pods/94dc1456-86ef-4000-85d1-5dee3eaca6aa/volumes" Oct 02 02:00:32 crc kubenswrapper[4775]: I1002 02:00:32.717446 4775 generic.go:334] "Generic (PLEG): container finished" podID="11558ec3-d684-403c-9afe-f08cb76fb18a" containerID="372a057f47923d18519533578d61e7d8f697805ca65253b9dde7c9953336d300" exitCode=0 Oct 02 02:00:32 crc kubenswrapper[4775]: I1002 02:00:32.717520 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" event={"ID":"11558ec3-d684-403c-9afe-f08cb76fb18a","Type":"ContainerDied","Data":"372a057f47923d18519533578d61e7d8f697805ca65253b9dde7c9953336d300"} Oct 02 02:00:34 crc kubenswrapper[4775]: I1002 02:00:34.650378 4775 scope.go:117] "RemoveContainer" containerID="b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3" Oct 02 02:00:34 crc kubenswrapper[4775]: E1002 02:00:34.651576 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3\": container with ID starting with b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3 not found: ID does not exist" containerID="b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3" Oct 02 02:00:34 crc kubenswrapper[4775]: I1002 02:00:34.651640 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3"} err="failed to get container status \"b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3\": rpc error: code = NotFound desc = could not find container \"b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3\": container with ID starting with b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3 not found: ID does not exist" Oct 02 02:00:34 crc kubenswrapper[4775]: I1002 02:00:34.651687 4775 scope.go:117] "RemoveContainer" containerID="399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765" Oct 02 02:00:34 crc kubenswrapper[4775]: E1002 02:00:34.652158 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765\": container with ID starting with 399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765 not found: ID does not exist" containerID="399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765" Oct 02 02:00:34 crc kubenswrapper[4775]: I1002 02:00:34.652230 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765"} err="failed to get container status \"399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765\": rpc error: code = NotFound desc = could not find container \"399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765\": container with ID starting with 399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765 not found: ID does not exist" Oct 02 02:00:34 crc kubenswrapper[4775]: I1002 02:00:34.652258 4775 scope.go:117] "RemoveContainer" containerID="b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3" Oct 02 02:00:34 crc kubenswrapper[4775]: I1002 02:00:34.652728 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3"} err="failed to get container status \"b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3\": rpc error: code = NotFound desc = could not find container \"b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3\": container with ID starting with b933272a0eab430d9d42bf9fd1dd690c137b2e29f043e086dc90e9ac3d337da3 not found: ID does not exist" Oct 02 02:00:34 crc kubenswrapper[4775]: I1002 02:00:34.652787 4775 scope.go:117] "RemoveContainer" containerID="399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765" Oct 02 02:00:34 crc kubenswrapper[4775]: I1002 02:00:34.653367 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765"} err="failed to get container status \"399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765\": rpc error: code = NotFound desc = could not find container \"399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765\": container with ID starting with 399380ec42c6ff0d804d35d6e9f95961c4ce41e0442743d26987a1a12dc9f765 not found: ID does not exist" Oct 02 02:00:34 crc kubenswrapper[4775]: W1002 02:00:34.668582 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb59797fc_1aa1_4106_ba3f_7239d6f4c7fb.slice/crio-586fe3ca1019c7dbd5e301cb818102473c665e29499d8629e9cb0d7dcd2f8cea WatchSource:0}: Error finding container 586fe3ca1019c7dbd5e301cb818102473c665e29499d8629e9cb0d7dcd2f8cea: Status 404 returned error can't find the container with id 586fe3ca1019c7dbd5e301cb818102473c665e29499d8629e9cb0d7dcd2f8cea Oct 02 02:00:34 crc kubenswrapper[4775]: I1002 02:00:34.752377 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-564487c87-zppvw" event={"ID":"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e","Type":"ContainerStarted","Data":"804847e082ad32e4c521651e8fb68070dd2ed6848268db247f5369cd1293a592"} Oct 02 02:00:34 crc kubenswrapper[4775]: I1002 02:00:34.760259 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb","Type":"ContainerStarted","Data":"586fe3ca1019c7dbd5e301cb818102473c665e29499d8629e9cb0d7dcd2f8cea"} Oct 02 02:00:34 crc kubenswrapper[4775]: I1002 02:00:34.881272 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-777987c7d4-d5vzz" event={"ID":"56813183-3dfc-430c-aaac-bda27d91340b","Type":"ContainerStarted","Data":"8baa28cbfb413992b0822f7aa986dd152be25ca69358d92c31e610054a7e6493"} Oct 02 02:00:35 crc kubenswrapper[4775]: I1002 02:00:35.252910 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:00:35 crc kubenswrapper[4775]: I1002 02:00:35.902534 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-564487c87-zppvw" event={"ID":"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e","Type":"ContainerStarted","Data":"7ad48d14db57c9f5d0c7e7d3fcaf64adb99c760e739a396090a74eb51c2a9919"} Oct 02 02:00:35 crc kubenswrapper[4775]: I1002 02:00:35.902829 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-564487c87-zppvw" Oct 02 02:00:35 crc kubenswrapper[4775]: I1002 02:00:35.909224 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb","Type":"ContainerStarted","Data":"6c7739e2aa0e5396b03e846410f94dee4c0629af1af171c37f25be6c5d7aee75"} Oct 02 02:00:35 crc kubenswrapper[4775]: I1002 02:00:35.911137 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-777987c7d4-d5vzz" event={"ID":"56813183-3dfc-430c-aaac-bda27d91340b","Type":"ContainerStarted","Data":"709d8a7bd6eb1d47bf0d40f08c279c162aa59c94b3583a614ae61982e1165eea"} Oct 02 02:00:35 crc kubenswrapper[4775]: I1002 02:00:35.922093 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-564487c87-zppvw" podStartSLOduration=6.922073305 podStartE2EDuration="6.922073305s" podCreationTimestamp="2025-10-02 02:00:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:35.916997204 +0000 UTC m=+1173.083741244" watchObservedRunningTime="2025-10-02 02:00:35.922073305 +0000 UTC m=+1173.088817345" Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.762405 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.856237 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tj4cb\" (UniqueName: \"kubernetes.io/projected/11558ec3-d684-403c-9afe-f08cb76fb18a-kube-api-access-tj4cb\") pod \"11558ec3-d684-403c-9afe-f08cb76fb18a\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.856546 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-dns-swift-storage-0\") pod \"11558ec3-d684-403c-9afe-f08cb76fb18a\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.856586 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-config\") pod \"11558ec3-d684-403c-9afe-f08cb76fb18a\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.856689 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-dns-svc\") pod \"11558ec3-d684-403c-9afe-f08cb76fb18a\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.856709 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-ovsdbserver-sb\") pod \"11558ec3-d684-403c-9afe-f08cb76fb18a\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.856732 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-ovsdbserver-nb\") pod \"11558ec3-d684-403c-9afe-f08cb76fb18a\" (UID: \"11558ec3-d684-403c-9afe-f08cb76fb18a\") " Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.872675 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11558ec3-d684-403c-9afe-f08cb76fb18a-kube-api-access-tj4cb" (OuterVolumeSpecName: "kube-api-access-tj4cb") pod "11558ec3-d684-403c-9afe-f08cb76fb18a" (UID: "11558ec3-d684-403c-9afe-f08cb76fb18a"). InnerVolumeSpecName "kube-api-access-tj4cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.912174 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-config" (OuterVolumeSpecName: "config") pod "11558ec3-d684-403c-9afe-f08cb76fb18a" (UID: "11558ec3-d684-403c-9afe-f08cb76fb18a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.938108 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" event={"ID":"11558ec3-d684-403c-9afe-f08cb76fb18a","Type":"ContainerDied","Data":"c479cd7a690e0c2e3e22dfa9f63c6404211aa4de14826c477648d0508f77fab6"} Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.938220 4775 scope.go:117] "RemoveContainer" containerID="372a057f47923d18519533578d61e7d8f697805ca65253b9dde7c9953336d300" Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.938131 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.939552 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08","Type":"ContainerStarted","Data":"bec1ab733d5e9f88875433a033f9f18e650c465e9b2bf5fafa524188bbcf1c8d"} Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.958543 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tj4cb\" (UniqueName: \"kubernetes.io/projected/11558ec3-d684-403c-9afe-f08cb76fb18a-kube-api-access-tj4cb\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.958573 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.963447 4775 scope.go:117] "RemoveContainer" containerID="a62f05a1cf2565d94fb4a6d347e31465959d96929cca83a05d2567b5bcd0704c" Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.979062 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "11558ec3-d684-403c-9afe-f08cb76fb18a" (UID: "11558ec3-d684-403c-9afe-f08cb76fb18a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.979854 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "11558ec3-d684-403c-9afe-f08cb76fb18a" (UID: "11558ec3-d684-403c-9afe-f08cb76fb18a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.982913 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "11558ec3-d684-403c-9afe-f08cb76fb18a" (UID: "11558ec3-d684-403c-9afe-f08cb76fb18a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:37 crc kubenswrapper[4775]: I1002 02:00:37.990841 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "11558ec3-d684-403c-9afe-f08cb76fb18a" (UID: "11558ec3-d684-403c-9afe-f08cb76fb18a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.060799 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.060839 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.060854 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.060867 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/11558ec3-d684-403c-9afe-f08cb76fb18a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.288158 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-nx6tz"] Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.293730 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77585f5f8c-nx6tz"] Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.948913 4775 generic.go:334] "Generic (PLEG): container finished" podID="d05cdd83-5605-46ed-88e8-168edbbcf7c1" containerID="63635a39be96c75e16a0cd5c161326889020e66580536e869bd6d92c322a7bb6" exitCode=0 Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.949541 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dw9gs" event={"ID":"d05cdd83-5605-46ed-88e8-168edbbcf7c1","Type":"ContainerDied","Data":"63635a39be96c75e16a0cd5c161326889020e66580536e869bd6d92c322a7bb6"} Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.953414 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-x8pvr" event={"ID":"9a3f9c1f-1298-4835-9935-56b808aae84d","Type":"ContainerStarted","Data":"b2061818493c7906d4c6165a2929e0f522ff078e37ced6c3df22b374b06afbf0"} Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.956322 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb","Type":"ContainerStarted","Data":"1e4b5c78eaff0fb64d29ae3fd2247940e4dbd27d85e5d5983cb6837acb633966"} Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.958557 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-777987c7d4-d5vzz" event={"ID":"56813183-3dfc-430c-aaac-bda27d91340b","Type":"ContainerStarted","Data":"d46af6901507bf3d2c6f5cd7933a7df11f703991822ecd42a838e772bb645206"} Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.959392 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.959444 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.962053 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12864b75-dbdc-4075-aea7-9d3cc2a6ca59","Type":"ContainerStarted","Data":"5b0b554f9e776b0b4e1d869b769c0e6122573dae6943b653687ab9c3b73cafa6"} Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.967849 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08","Type":"ContainerStarted","Data":"4b96b0239299787f213eec6812a29703fc5113d42a71f339aa0bca3867dc4fa0"} Oct 02 02:00:38 crc kubenswrapper[4775]: I1002 02:00:38.967896 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08","Type":"ContainerStarted","Data":"229dd58f2fa05172f374734e3ca5ed77fc063711c568f4fe9c41b9a071b85798"} Oct 02 02:00:39 crc kubenswrapper[4775]: I1002 02:00:39.019661 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=10.019637221 podStartE2EDuration="10.019637221s" podCreationTimestamp="2025-10-02 02:00:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:39.007739175 +0000 UTC m=+1176.174483225" watchObservedRunningTime="2025-10-02 02:00:39.019637221 +0000 UTC m=+1176.186381271" Oct 02 02:00:39 crc kubenswrapper[4775]: I1002 02:00:39.049880 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-777987c7d4-d5vzz" podStartSLOduration=9.049859049 podStartE2EDuration="9.049859049s" podCreationTimestamp="2025-10-02 02:00:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:39.03981598 +0000 UTC m=+1176.206560090" watchObservedRunningTime="2025-10-02 02:00:39.049859049 +0000 UTC m=+1176.216603119" Oct 02 02:00:39 crc kubenswrapper[4775]: I1002 02:00:39.071614 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.071590508 podStartE2EDuration="8.071590508s" podCreationTimestamp="2025-10-02 02:00:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:39.06583988 +0000 UTC m=+1176.232583930" watchObservedRunningTime="2025-10-02 02:00:39.071590508 +0000 UTC m=+1176.238334568" Oct 02 02:00:39 crc kubenswrapper[4775]: I1002 02:00:39.777006 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11558ec3-d684-403c-9afe-f08cb76fb18a" path="/var/lib/kubelet/pods/11558ec3-d684-403c-9afe-f08cb76fb18a/volumes" Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.175540 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.175574 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.211909 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.233002 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.239821 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-x8pvr" podStartSLOduration=3.836054909 podStartE2EDuration="37.23980205s" podCreationTimestamp="2025-10-02 02:00:03 +0000 UTC" firstStartedPulling="2025-10-02 02:00:04.372495553 +0000 UTC m=+1141.539239593" lastFinishedPulling="2025-10-02 02:00:37.776242704 +0000 UTC m=+1174.942986734" observedRunningTime="2025-10-02 02:00:39.094415695 +0000 UTC m=+1176.261159775" watchObservedRunningTime="2025-10-02 02:00:40.23980205 +0000 UTC m=+1177.406546080" Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.261439 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-77585f5f8c-nx6tz" podUID="11558ec3-d684-403c-9afe-f08cb76fb18a" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.124:5353: i/o timeout" Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.327108 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dw9gs" Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.406158 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05cdd83-5605-46ed-88e8-168edbbcf7c1-combined-ca-bundle\") pod \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\" (UID: \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\") " Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.406206 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh2gs\" (UniqueName: \"kubernetes.io/projected/d05cdd83-5605-46ed-88e8-168edbbcf7c1-kube-api-access-xh2gs\") pod \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\" (UID: \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\") " Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.406313 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d05cdd83-5605-46ed-88e8-168edbbcf7c1-db-sync-config-data\") pod \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\" (UID: \"d05cdd83-5605-46ed-88e8-168edbbcf7c1\") " Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.416208 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d05cdd83-5605-46ed-88e8-168edbbcf7c1-kube-api-access-xh2gs" (OuterVolumeSpecName: "kube-api-access-xh2gs") pod "d05cdd83-5605-46ed-88e8-168edbbcf7c1" (UID: "d05cdd83-5605-46ed-88e8-168edbbcf7c1"). InnerVolumeSpecName "kube-api-access-xh2gs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.416417 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05cdd83-5605-46ed-88e8-168edbbcf7c1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d05cdd83-5605-46ed-88e8-168edbbcf7c1" (UID: "d05cdd83-5605-46ed-88e8-168edbbcf7c1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.443657 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d05cdd83-5605-46ed-88e8-168edbbcf7c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d05cdd83-5605-46ed-88e8-168edbbcf7c1" (UID: "d05cdd83-5605-46ed-88e8-168edbbcf7c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.508155 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d05cdd83-5605-46ed-88e8-168edbbcf7c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.508187 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh2gs\" (UniqueName: \"kubernetes.io/projected/d05cdd83-5605-46ed-88e8-168edbbcf7c1-kube-api-access-xh2gs\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:40 crc kubenswrapper[4775]: I1002 02:00:40.508198 4775 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d05cdd83-5605-46ed-88e8-168edbbcf7c1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.022432 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-dw9gs" event={"ID":"d05cdd83-5605-46ed-88e8-168edbbcf7c1","Type":"ContainerDied","Data":"2a660c8f302edfa8723a629dde6891ac96c10e9f88b1ba6d5e926694b4cbe3c5"} Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.022510 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a660c8f302edfa8723a629dde6891ac96c10e9f88b1ba6d5e926694b4cbe3c5" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.022887 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-dw9gs" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.023510 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.023738 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.230214 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-547cbfbdcc-nk6pz"] Oct 02 02:00:41 crc kubenswrapper[4775]: E1002 02:00:41.230800 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11558ec3-d684-403c-9afe-f08cb76fb18a" containerName="dnsmasq-dns" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.230812 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="11558ec3-d684-403c-9afe-f08cb76fb18a" containerName="dnsmasq-dns" Oct 02 02:00:41 crc kubenswrapper[4775]: E1002 02:00:41.230846 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11558ec3-d684-403c-9afe-f08cb76fb18a" containerName="init" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.230852 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="11558ec3-d684-403c-9afe-f08cb76fb18a" containerName="init" Oct 02 02:00:41 crc kubenswrapper[4775]: E1002 02:00:41.230872 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05cdd83-5605-46ed-88e8-168edbbcf7c1" containerName="barbican-db-sync" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.230877 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05cdd83-5605-46ed-88e8-168edbbcf7c1" containerName="barbican-db-sync" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.234451 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="11558ec3-d684-403c-9afe-f08cb76fb18a" containerName="dnsmasq-dns" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.234514 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05cdd83-5605-46ed-88e8-168edbbcf7c1" containerName="barbican-db-sync" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.235525 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.251081 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.254102 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-cwwrx" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.254171 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.256309 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-547cbfbdcc-nk6pz"] Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.332581 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-config-data-custom\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.332626 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-config-data\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.332678 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/163a6d17-be4c-47fc-8835-e218e8a803f1-logs\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.332730 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64lhb\" (UniqueName: \"kubernetes.io/projected/163a6d17-be4c-47fc-8835-e218e8a803f1-kube-api-access-64lhb\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.332746 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-combined-ca-bundle\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.333688 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-d79dd6fcb-lzscb"] Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.337795 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.341572 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.358432 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-d79dd6fcb-lzscb"] Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.397051 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-npmm7"] Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.398543 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.408002 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-npmm7"] Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434011 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-config-data-custom\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434066 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-config-data-custom\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434089 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-config-data\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434666 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/163a6d17-be4c-47fc-8835-e218e8a803f1-logs\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434691 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434749 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af16c5f4-fcd0-4450-a8b7-789ff104e477-logs\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434779 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64lhb\" (UniqueName: \"kubernetes.io/projected/163a6d17-be4c-47fc-8835-e218e8a803f1-kube-api-access-64lhb\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434799 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-config-data\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434818 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-combined-ca-bundle\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434855 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434882 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434909 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-config\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434928 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqjt7\" (UniqueName: \"kubernetes.io/projected/af16c5f4-fcd0-4450-a8b7-789ff104e477-kube-api-access-dqjt7\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434962 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-combined-ca-bundle\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.434987 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.435009 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdfjz\" (UniqueName: \"kubernetes.io/projected/1f4eab32-163d-44a4-8cbd-66ca8968b869-kube-api-access-gdfjz\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.435254 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/163a6d17-be4c-47fc-8835-e218e8a803f1-logs\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.443966 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-config-data\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.453095 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-config-data-custom\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.453746 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-combined-ca-bundle\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.460767 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64lhb\" (UniqueName: \"kubernetes.io/projected/163a6d17-be4c-47fc-8835-e218e8a803f1-kube-api-access-64lhb\") pod \"barbican-worker-547cbfbdcc-nk6pz\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.463714 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.463775 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.512115 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5698978bb-cdg88"] Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.529611 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.535658 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.535813 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.536373 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-config-data-custom\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.536423 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.536484 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af16c5f4-fcd0-4450-a8b7-789ff104e477-logs\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.536514 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-config-data\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.536554 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.536580 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.536606 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-config\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.536622 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqjt7\" (UniqueName: \"kubernetes.io/projected/af16c5f4-fcd0-4450-a8b7-789ff104e477-kube-api-access-dqjt7\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.536658 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-combined-ca-bundle\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.536680 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.536706 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdfjz\" (UniqueName: \"kubernetes.io/projected/1f4eab32-163d-44a4-8cbd-66ca8968b869-kube-api-access-gdfjz\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.537722 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-ovsdbserver-nb\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.538594 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-dns-swift-storage-0\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.538855 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af16c5f4-fcd0-4450-a8b7-789ff104e477-logs\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.542481 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-config-data-custom\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.543242 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-dns-svc\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.543482 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-ovsdbserver-sb\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.545875 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-config\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.553775 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.560352 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-combined-ca-bundle\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.561554 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-config-data\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.562781 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.573705 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqjt7\" (UniqueName: \"kubernetes.io/projected/af16c5f4-fcd0-4450-a8b7-789ff104e477-kube-api-access-dqjt7\") pod \"barbican-keystone-listener-d79dd6fcb-lzscb\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.574612 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdfjz\" (UniqueName: \"kubernetes.io/projected/1f4eab32-163d-44a4-8cbd-66ca8968b869-kube-api-access-gdfjz\") pod \"dnsmasq-dns-59d5ff467f-npmm7\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.577102 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5698978bb-cdg88"] Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.638427 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-config-data\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.638469 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjd2t\" (UniqueName: \"kubernetes.io/projected/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-kube-api-access-pjd2t\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.638490 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-combined-ca-bundle\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.638509 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-config-data-custom\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.638557 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-logs\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.672675 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.718477 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.740482 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-config-data\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.740538 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjd2t\" (UniqueName: \"kubernetes.io/projected/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-kube-api-access-pjd2t\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.740566 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-combined-ca-bundle\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.740593 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-config-data-custom\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.740656 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-logs\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.741370 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-logs\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.745552 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-config-data\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.748118 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-config-data-custom\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.752608 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-combined-ca-bundle\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.757240 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjd2t\" (UniqueName: \"kubernetes.io/projected/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-kube-api-access-pjd2t\") pod \"barbican-api-5698978bb-cdg88\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:41 crc kubenswrapper[4775]: I1002 02:00:41.973500 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:42 crc kubenswrapper[4775]: I1002 02:00:42.031724 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:42 crc kubenswrapper[4775]: I1002 02:00:42.032184 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:42 crc kubenswrapper[4775]: I1002 02:00:42.062488 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-547cbfbdcc-nk6pz"] Oct 02 02:00:42 crc kubenswrapper[4775]: I1002 02:00:42.253843 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-d79dd6fcb-lzscb"] Oct 02 02:00:42 crc kubenswrapper[4775]: I1002 02:00:42.262326 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5698978bb-cdg88"] Oct 02 02:00:42 crc kubenswrapper[4775]: W1002 02:00:42.264126 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f876ea1_9d84_4dd0_a2f6_2bc742bab66e.slice/crio-b0ffa726f8ffaaafd74e169c0f0ff5960db52a4dd04a8271aec417e9b2fae302 WatchSource:0}: Error finding container b0ffa726f8ffaaafd74e169c0f0ff5960db52a4dd04a8271aec417e9b2fae302: Status 404 returned error can't find the container with id b0ffa726f8ffaaafd74e169c0f0ff5960db52a4dd04a8271aec417e9b2fae302 Oct 02 02:00:42 crc kubenswrapper[4775]: I1002 02:00:42.313893 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-npmm7"] Oct 02 02:00:42 crc kubenswrapper[4775]: W1002 02:00:42.321892 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f4eab32_163d_44a4_8cbd_66ca8968b869.slice/crio-648196906c795b5040126a3ae0667a890cb0eb19ce2eb000ca9e85208bc8396f WatchSource:0}: Error finding container 648196906c795b5040126a3ae0667a890cb0eb19ce2eb000ca9e85208bc8396f: Status 404 returned error can't find the container with id 648196906c795b5040126a3ae0667a890cb0eb19ce2eb000ca9e85208bc8396f Oct 02 02:00:42 crc kubenswrapper[4775]: I1002 02:00:42.967545 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:43 crc kubenswrapper[4775]: I1002 02:00:43.045380 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5698978bb-cdg88" event={"ID":"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e","Type":"ContainerStarted","Data":"250bae0449846fc4dda5adff85b5ed10a3b0e5cffbbb6251ffe6503a2a43110e"} Oct 02 02:00:43 crc kubenswrapper[4775]: I1002 02:00:43.045425 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5698978bb-cdg88" event={"ID":"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e","Type":"ContainerStarted","Data":"b0ffa726f8ffaaafd74e169c0f0ff5960db52a4dd04a8271aec417e9b2fae302"} Oct 02 02:00:43 crc kubenswrapper[4775]: I1002 02:00:43.046863 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" event={"ID":"af16c5f4-fcd0-4450-a8b7-789ff104e477","Type":"ContainerStarted","Data":"4c3c3ac65982b5585d40b0110d0e93c428516101d184f5a470d6b08eb9b85bfb"} Oct 02 02:00:43 crc kubenswrapper[4775]: I1002 02:00:43.048749 4775 generic.go:334] "Generic (PLEG): container finished" podID="1f4eab32-163d-44a4-8cbd-66ca8968b869" containerID="601abb7448e51003fa2c88bcae3cac775671f4277db926661d05d971e6098dfa" exitCode=0 Oct 02 02:00:43 crc kubenswrapper[4775]: I1002 02:00:43.048799 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" event={"ID":"1f4eab32-163d-44a4-8cbd-66ca8968b869","Type":"ContainerDied","Data":"601abb7448e51003fa2c88bcae3cac775671f4277db926661d05d971e6098dfa"} Oct 02 02:00:43 crc kubenswrapper[4775]: I1002 02:00:43.048814 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" event={"ID":"1f4eab32-163d-44a4-8cbd-66ca8968b869","Type":"ContainerStarted","Data":"648196906c795b5040126a3ae0667a890cb0eb19ce2eb000ca9e85208bc8396f"} Oct 02 02:00:43 crc kubenswrapper[4775]: I1002 02:00:43.052422 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-547cbfbdcc-nk6pz" event={"ID":"163a6d17-be4c-47fc-8835-e218e8a803f1","Type":"ContainerStarted","Data":"3ce9875a5a256a84090325bfbb15e5ce36ac47673c839e8cef7c8c64a9706094"} Oct 02 02:00:43 crc kubenswrapper[4775]: I1002 02:00:43.242593 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.017410 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.759278 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-79594d69d8-x75fk"] Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.760846 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.765872 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.766060 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.782647 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-79594d69d8-x75fk"] Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.815624 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-combined-ca-bundle\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.815668 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-internal-tls-certs\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.815748 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-config-data\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.815775 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d6bb525-8c94-461f-84c8-9e466b789706-logs\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.815791 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-config-data-custom\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.815811 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-public-tls-certs\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.815847 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmg46\" (UniqueName: \"kubernetes.io/projected/8d6bb525-8c94-461f-84c8-9e466b789706-kube-api-access-qmg46\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.917042 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-config-data\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.917096 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d6bb525-8c94-461f-84c8-9e466b789706-logs\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.917116 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-config-data-custom\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.917154 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-public-tls-certs\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.917202 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmg46\" (UniqueName: \"kubernetes.io/projected/8d6bb525-8c94-461f-84c8-9e466b789706-kube-api-access-qmg46\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.917234 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-combined-ca-bundle\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.917252 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-internal-tls-certs\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.920129 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d6bb525-8c94-461f-84c8-9e466b789706-logs\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.923622 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-public-tls-certs\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.923849 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-config-data\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.935185 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-internal-tls-certs\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.935310 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-combined-ca-bundle\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.935727 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmg46\" (UniqueName: \"kubernetes.io/projected/8d6bb525-8c94-461f-84c8-9e466b789706-kube-api-access-qmg46\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:44 crc kubenswrapper[4775]: I1002 02:00:44.936053 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-config-data-custom\") pod \"barbican-api-79594d69d8-x75fk\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:45 crc kubenswrapper[4775]: I1002 02:00:45.090320 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:45 crc kubenswrapper[4775]: I1002 02:00:45.239942 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:45 crc kubenswrapper[4775]: I1002 02:00:45.240308 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 02:00:45 crc kubenswrapper[4775]: I1002 02:00:45.285012 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 02:00:45 crc kubenswrapper[4775]: I1002 02:00:45.480177 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 02:00:46 crc kubenswrapper[4775]: I1002 02:00:46.091119 4775 generic.go:334] "Generic (PLEG): container finished" podID="9a3f9c1f-1298-4835-9935-56b808aae84d" containerID="b2061818493c7906d4c6165a2929e0f522ff078e37ced6c3df22b374b06afbf0" exitCode=0 Oct 02 02:00:46 crc kubenswrapper[4775]: I1002 02:00:46.091206 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-x8pvr" event={"ID":"9a3f9c1f-1298-4835-9935-56b808aae84d","Type":"ContainerDied","Data":"b2061818493c7906d4c6165a2929e0f522ff078e37ced6c3df22b374b06afbf0"} Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.746489 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.869296 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-combined-ca-bundle\") pod \"9a3f9c1f-1298-4835-9935-56b808aae84d\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.869385 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-scripts\") pod \"9a3f9c1f-1298-4835-9935-56b808aae84d\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.869460 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a3f9c1f-1298-4835-9935-56b808aae84d-etc-machine-id\") pod \"9a3f9c1f-1298-4835-9935-56b808aae84d\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.869534 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-db-sync-config-data\") pod \"9a3f9c1f-1298-4835-9935-56b808aae84d\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.869580 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzddh\" (UniqueName: \"kubernetes.io/projected/9a3f9c1f-1298-4835-9935-56b808aae84d-kube-api-access-xzddh\") pod \"9a3f9c1f-1298-4835-9935-56b808aae84d\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.869631 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a3f9c1f-1298-4835-9935-56b808aae84d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9a3f9c1f-1298-4835-9935-56b808aae84d" (UID: "9a3f9c1f-1298-4835-9935-56b808aae84d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.869803 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-config-data\") pod \"9a3f9c1f-1298-4835-9935-56b808aae84d\" (UID: \"9a3f9c1f-1298-4835-9935-56b808aae84d\") " Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.872074 4775 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a3f9c1f-1298-4835-9935-56b808aae84d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.875396 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a3f9c1f-1298-4835-9935-56b808aae84d-kube-api-access-xzddh" (OuterVolumeSpecName: "kube-api-access-xzddh") pod "9a3f9c1f-1298-4835-9935-56b808aae84d" (UID: "9a3f9c1f-1298-4835-9935-56b808aae84d"). InnerVolumeSpecName "kube-api-access-xzddh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.876105 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-scripts" (OuterVolumeSpecName: "scripts") pod "9a3f9c1f-1298-4835-9935-56b808aae84d" (UID: "9a3f9c1f-1298-4835-9935-56b808aae84d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.876789 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9a3f9c1f-1298-4835-9935-56b808aae84d" (UID: "9a3f9c1f-1298-4835-9935-56b808aae84d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.901576 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a3f9c1f-1298-4835-9935-56b808aae84d" (UID: "9a3f9c1f-1298-4835-9935-56b808aae84d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.947446 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-config-data" (OuterVolumeSpecName: "config-data") pod "9a3f9c1f-1298-4835-9935-56b808aae84d" (UID: "9a3f9c1f-1298-4835-9935-56b808aae84d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.976278 4775 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.976321 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzddh\" (UniqueName: \"kubernetes.io/projected/9a3f9c1f-1298-4835-9935-56b808aae84d-kube-api-access-xzddh\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.976334 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.976348 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:47 crc kubenswrapper[4775]: I1002 02:00:47.976360 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a3f9c1f-1298-4835-9935-56b808aae84d-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.116658 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-x8pvr" event={"ID":"9a3f9c1f-1298-4835-9935-56b808aae84d","Type":"ContainerDied","Data":"003ee24629e2b9cd3dd9930345a1f20f8ff5431e535892d9895a6bb4aa50c018"} Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.116691 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="003ee24629e2b9cd3dd9930345a1f20f8ff5431e535892d9895a6bb4aa50c018" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.116746 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-x8pvr" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.382475 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:00:48 crc kubenswrapper[4775]: E1002 02:00:48.383203 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a3f9c1f-1298-4835-9935-56b808aae84d" containerName="cinder-db-sync" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.383219 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a3f9c1f-1298-4835-9935-56b808aae84d" containerName="cinder-db-sync" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.383394 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a3f9c1f-1298-4835-9935-56b808aae84d" containerName="cinder-db-sync" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.384224 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.391430 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-w7wmc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.391667 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.391690 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.391827 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.408635 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.494678 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.494809 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.494852 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.494883 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6prjn\" (UniqueName: \"kubernetes.io/projected/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-kube-api-access-6prjn\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.495115 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-scripts\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.495186 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-config-data\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.497697 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-npmm7"] Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.545583 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-8gllc"] Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.547057 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.552045 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-8gllc"] Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.596755 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-ovsdbserver-nb\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.596803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.596828 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-ovsdbserver-sb\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.596853 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6prjn\" (UniqueName: \"kubernetes.io/projected/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-kube-api-access-6prjn\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.596877 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-config\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.597070 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-scripts\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.597125 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-dns-swift-storage-0\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.597152 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-config-data\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.597190 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fns7m\" (UniqueName: \"kubernetes.io/projected/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-kube-api-access-fns7m\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.597231 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-dns-svc\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.597375 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.597545 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.597646 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.601485 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.602478 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.605121 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-scripts\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.611727 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-config-data\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.620329 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6prjn\" (UniqueName: \"kubernetes.io/projected/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-kube-api-access-6prjn\") pod \"cinder-scheduler-0\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.656312 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.658637 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.660390 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.668256 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.699313 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.699753 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j46d7\" (UniqueName: \"kubernetes.io/projected/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-kube-api-access-j46d7\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.699791 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-config-data\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.699823 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-ovsdbserver-nb\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.699904 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-ovsdbserver-sb\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.699932 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-config\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.699990 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-dns-swift-storage-0\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.700019 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fns7m\" (UniqueName: \"kubernetes.io/projected/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-kube-api-access-fns7m\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.700040 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-dns-svc\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.700060 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-config-data-custom\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.700077 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.700102 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-scripts\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.700118 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-logs\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.702967 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-ovsdbserver-nb\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.704003 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-ovsdbserver-sb\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.704763 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-dns-svc\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.705026 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-dns-swift-storage-0\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.706215 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-config\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.726379 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fns7m\" (UniqueName: \"kubernetes.io/projected/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-kube-api-access-fns7m\") pod \"dnsmasq-dns-69c986f6d7-8gllc\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.745161 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.801057 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j46d7\" (UniqueName: \"kubernetes.io/projected/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-kube-api-access-j46d7\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.801109 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-config-data\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.801191 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-config-data-custom\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.801247 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.801342 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.801825 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-scripts\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.801861 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-logs\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.801919 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.802261 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-logs\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.804402 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-scripts\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.806416 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.806460 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-config-data-custom\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.808336 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-config-data\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.819645 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j46d7\" (UniqueName: \"kubernetes.io/projected/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-kube-api-access-j46d7\") pod \"cinder-api-0\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " pod="openstack/cinder-api-0" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.874547 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:48 crc kubenswrapper[4775]: I1002 02:00:48.984042 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 02:00:49 crc kubenswrapper[4775]: I1002 02:00:49.310570 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-79594d69d8-x75fk"] Oct 02 02:00:49 crc kubenswrapper[4775]: W1002 02:00:49.386788 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d6bb525_8c94_461f_84c8_9e466b789706.slice/crio-107e97586776a360c4d800e0a787832481a9cc510a49a9cdb3db505fa31fdb98 WatchSource:0}: Error finding container 107e97586776a360c4d800e0a787832481a9cc510a49a9cdb3db505fa31fdb98: Status 404 returned error can't find the container with id 107e97586776a360c4d800e0a787832481a9cc510a49a9cdb3db505fa31fdb98 Oct 02 02:00:49 crc kubenswrapper[4775]: I1002 02:00:49.909817 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:00:49 crc kubenswrapper[4775]: I1002 02:00:49.916733 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-8gllc"] Oct 02 02:00:49 crc kubenswrapper[4775]: W1002 02:00:49.917887 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3b8d4b2f_17d1_4c47_bcc7_97f3fd22220c.slice/crio-b85be8a505af5b62a8498133fbe804c67a52ee5457e0cb7ba63349c7dfb1e10c WatchSource:0}: Error finding container b85be8a505af5b62a8498133fbe804c67a52ee5457e0cb7ba63349c7dfb1e10c: Status 404 returned error can't find the container with id b85be8a505af5b62a8498133fbe804c67a52ee5457e0cb7ba63349c7dfb1e10c Oct 02 02:00:49 crc kubenswrapper[4775]: I1002 02:00:49.923530 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.143018 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb","Type":"ContainerStarted","Data":"de4b034df51edadffee7c77be51543d58ed4bdf255cbf23cfcab7d1b11981917"} Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.146446 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12864b75-dbdc-4075-aea7-9d3cc2a6ca59","Type":"ContainerStarted","Data":"21eafeea4db54d3a522d2c5f0ab85cd2f58d653150e09d88c4504fb18084f5ba"} Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.146574 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="ceilometer-central-agent" containerID="cri-o://2d8a19e074a59869fa182becc84c8c6153a9d08b9d345d4f7d9e637c21100594" gracePeriod=30 Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.146649 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="sg-core" containerID="cri-o://5b0b554f9e776b0b4e1d869b769c0e6122573dae6943b653687ab9c3b73cafa6" gracePeriod=30 Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.146746 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="proxy-httpd" containerID="cri-o://21eafeea4db54d3a522d2c5f0ab85cd2f58d653150e09d88c4504fb18084f5ba" gracePeriod=30 Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.146748 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="ceilometer-notification-agent" containerID="cri-o://89a042d772657ba277e82d383099114e18ac75f765966f3219b0c0ca7ba4fc91" gracePeriod=30 Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.146599 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.147846 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" event={"ID":"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b","Type":"ContainerStarted","Data":"34793f825a3c40a0ff7049782635751392f23e10053596c1325c51cb316d8a20"} Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.149143 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" event={"ID":"1f4eab32-163d-44a4-8cbd-66ca8968b869","Type":"ContainerStarted","Data":"7efddf689328bfd9c16bd39edd4e72f43dca2d239a2d1b5794ff0ed39eb5a8b5"} Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.149328 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" podUID="1f4eab32-163d-44a4-8cbd-66ca8968b869" containerName="dnsmasq-dns" containerID="cri-o://7efddf689328bfd9c16bd39edd4e72f43dca2d239a2d1b5794ff0ed39eb5a8b5" gracePeriod=10 Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.149427 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.157251 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5698978bb-cdg88" event={"ID":"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e","Type":"ContainerStarted","Data":"169652bfa5bde2f7471ba1a8910ef01f1640b2f0a9daf2c6c26ac42892314dbf"} Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.157478 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.157836 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.161510 4775 generic.go:334] "Generic (PLEG): container finished" podID="c34b79a8-c071-4fff-8dbc-cab3601cc3f8" containerID="3fc5ef432fab421b40f1e30656d3ab99d83e495569dde9d28576177506951a8b" exitCode=0 Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.161585 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hjvr2" event={"ID":"c34b79a8-c071-4fff-8dbc-cab3601cc3f8","Type":"ContainerDied","Data":"3fc5ef432fab421b40f1e30656d3ab99d83e495569dde9d28576177506951a8b"} Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.167131 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c","Type":"ContainerStarted","Data":"b85be8a505af5b62a8498133fbe804c67a52ee5457e0cb7ba63349c7dfb1e10c"} Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.169818 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.342691307 podStartE2EDuration="48.169806093s" podCreationTimestamp="2025-10-02 02:00:02 +0000 UTC" firstStartedPulling="2025-10-02 02:00:03.558790821 +0000 UTC m=+1140.725534861" lastFinishedPulling="2025-10-02 02:00:49.385905597 +0000 UTC m=+1186.552649647" observedRunningTime="2025-10-02 02:00:50.167460813 +0000 UTC m=+1187.334204853" watchObservedRunningTime="2025-10-02 02:00:50.169806093 +0000 UTC m=+1187.336550133" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.173364 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79594d69d8-x75fk" event={"ID":"8d6bb525-8c94-461f-84c8-9e466b789706","Type":"ContainerStarted","Data":"ddc807916bf2152781e0b263b81d2ba2f8ff95a514f51a7d7918774142370f47"} Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.173406 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79594d69d8-x75fk" event={"ID":"8d6bb525-8c94-461f-84c8-9e466b789706","Type":"ContainerStarted","Data":"9f162e9f65c919c062e4c98961fb915d6d73a9e2d76b98f4844b90766369c972"} Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.173421 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79594d69d8-x75fk" event={"ID":"8d6bb525-8c94-461f-84c8-9e466b789706","Type":"ContainerStarted","Data":"107e97586776a360c4d800e0a787832481a9cc510a49a9cdb3db505fa31fdb98"} Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.173863 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.174007 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.177265 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" event={"ID":"af16c5f4-fcd0-4450-a8b7-789ff104e477","Type":"ContainerStarted","Data":"ae9bd041c2c057dee311d85415e1bfd222d0109f3f01cb2fcd2149362f273f6e"} Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.177300 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" event={"ID":"af16c5f4-fcd0-4450-a8b7-789ff104e477","Type":"ContainerStarted","Data":"3aff4b0f97413fafd4a12944ee3201d623169d99e066f7bf1d2aa5a8977df17f"} Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.180608 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-547cbfbdcc-nk6pz" event={"ID":"163a6d17-be4c-47fc-8835-e218e8a803f1","Type":"ContainerStarted","Data":"979478a6b7d6bf80d7823b67cd948eb55c9e92647ad8b39e7e0c75474a16acc7"} Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.180647 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-547cbfbdcc-nk6pz" event={"ID":"163a6d17-be4c-47fc-8835-e218e8a803f1","Type":"ContainerStarted","Data":"9d74e2868485fd65eb7dd4d89c912bc745152afc0102e87eaff58d02c1858850"} Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.191299 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5698978bb-cdg88" podStartSLOduration=9.191285856 podStartE2EDuration="9.191285856s" podCreationTimestamp="2025-10-02 02:00:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:50.189936871 +0000 UTC m=+1187.356680911" watchObservedRunningTime="2025-10-02 02:00:50.191285856 +0000 UTC m=+1187.358029896" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.242351 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" podStartSLOduration=9.242333178 podStartE2EDuration="9.242333178s" podCreationTimestamp="2025-10-02 02:00:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:50.220198768 +0000 UTC m=+1187.386942808" watchObservedRunningTime="2025-10-02 02:00:50.242333178 +0000 UTC m=+1187.409077218" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.294175 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" podStartSLOduration=2.150630422 podStartE2EDuration="9.294154911s" podCreationTimestamp="2025-10-02 02:00:41 +0000 UTC" firstStartedPulling="2025-10-02 02:00:42.276574757 +0000 UTC m=+1179.443318807" lastFinishedPulling="2025-10-02 02:00:49.420099236 +0000 UTC m=+1186.586843296" observedRunningTime="2025-10-02 02:00:50.274405913 +0000 UTC m=+1187.441149953" watchObservedRunningTime="2025-10-02 02:00:50.294154911 +0000 UTC m=+1187.460898951" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.312477 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-547cbfbdcc-nk6pz" podStartSLOduration=1.959899855 podStartE2EDuration="9.312456052s" podCreationTimestamp="2025-10-02 02:00:41 +0000 UTC" firstStartedPulling="2025-10-02 02:00:42.083617093 +0000 UTC m=+1179.250361133" lastFinishedPulling="2025-10-02 02:00:49.43617328 +0000 UTC m=+1186.602917330" observedRunningTime="2025-10-02 02:00:50.302651089 +0000 UTC m=+1187.469395129" watchObservedRunningTime="2025-10-02 02:00:50.312456052 +0000 UTC m=+1187.479200092" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.329888 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-79594d69d8-x75fk" podStartSLOduration=6.329872 podStartE2EDuration="6.329872s" podCreationTimestamp="2025-10-02 02:00:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:50.328202927 +0000 UTC m=+1187.494946967" watchObservedRunningTime="2025-10-02 02:00:50.329872 +0000 UTC m=+1187.496616040" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.623783 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.743597 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-ovsdbserver-sb\") pod \"1f4eab32-163d-44a4-8cbd-66ca8968b869\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.743668 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-config\") pod \"1f4eab32-163d-44a4-8cbd-66ca8968b869\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.743762 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-dns-svc\") pod \"1f4eab32-163d-44a4-8cbd-66ca8968b869\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.743821 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdfjz\" (UniqueName: \"kubernetes.io/projected/1f4eab32-163d-44a4-8cbd-66ca8968b869-kube-api-access-gdfjz\") pod \"1f4eab32-163d-44a4-8cbd-66ca8968b869\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.743901 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-ovsdbserver-nb\") pod \"1f4eab32-163d-44a4-8cbd-66ca8968b869\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.743984 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-dns-swift-storage-0\") pod \"1f4eab32-163d-44a4-8cbd-66ca8968b869\" (UID: \"1f4eab32-163d-44a4-8cbd-66ca8968b869\") " Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.753121 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f4eab32-163d-44a4-8cbd-66ca8968b869-kube-api-access-gdfjz" (OuterVolumeSpecName: "kube-api-access-gdfjz") pod "1f4eab32-163d-44a4-8cbd-66ca8968b869" (UID: "1f4eab32-163d-44a4-8cbd-66ca8968b869"). InnerVolumeSpecName "kube-api-access-gdfjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.795463 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1f4eab32-163d-44a4-8cbd-66ca8968b869" (UID: "1f4eab32-163d-44a4-8cbd-66ca8968b869"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.795618 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1f4eab32-163d-44a4-8cbd-66ca8968b869" (UID: "1f4eab32-163d-44a4-8cbd-66ca8968b869"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.799972 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1f4eab32-163d-44a4-8cbd-66ca8968b869" (UID: "1f4eab32-163d-44a4-8cbd-66ca8968b869"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.801894 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-config" (OuterVolumeSpecName: "config") pod "1f4eab32-163d-44a4-8cbd-66ca8968b869" (UID: "1f4eab32-163d-44a4-8cbd-66ca8968b869"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.805444 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1f4eab32-163d-44a4-8cbd-66ca8968b869" (UID: "1f4eab32-163d-44a4-8cbd-66ca8968b869"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.845509 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.845541 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.845552 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.845561 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.845570 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1f4eab32-163d-44a4-8cbd-66ca8968b869-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:50 crc kubenswrapper[4775]: I1002 02:00:50.845578 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdfjz\" (UniqueName: \"kubernetes.io/projected/1f4eab32-163d-44a4-8cbd-66ca8968b869-kube-api-access-gdfjz\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.038164 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.206945 4775 generic.go:334] "Generic (PLEG): container finished" podID="1f4eab32-163d-44a4-8cbd-66ca8968b869" containerID="7efddf689328bfd9c16bd39edd4e72f43dca2d239a2d1b5794ff0ed39eb5a8b5" exitCode=0 Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.207069 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" event={"ID":"1f4eab32-163d-44a4-8cbd-66ca8968b869","Type":"ContainerDied","Data":"7efddf689328bfd9c16bd39edd4e72f43dca2d239a2d1b5794ff0ed39eb5a8b5"} Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.207101 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" event={"ID":"1f4eab32-163d-44a4-8cbd-66ca8968b869","Type":"ContainerDied","Data":"648196906c795b5040126a3ae0667a890cb0eb19ce2eb000ca9e85208bc8396f"} Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.207123 4775 scope.go:117] "RemoveContainer" containerID="7efddf689328bfd9c16bd39edd4e72f43dca2d239a2d1b5794ff0ed39eb5a8b5" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.207299 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59d5ff467f-npmm7" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.211374 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb","Type":"ContainerStarted","Data":"775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3"} Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.237090 4775 generic.go:334] "Generic (PLEG): container finished" podID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerID="21eafeea4db54d3a522d2c5f0ab85cd2f58d653150e09d88c4504fb18084f5ba" exitCode=0 Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.237128 4775 generic.go:334] "Generic (PLEG): container finished" podID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerID="5b0b554f9e776b0b4e1d869b769c0e6122573dae6943b653687ab9c3b73cafa6" exitCode=2 Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.237136 4775 generic.go:334] "Generic (PLEG): container finished" podID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerID="2d8a19e074a59869fa182becc84c8c6153a9d08b9d345d4f7d9e637c21100594" exitCode=0 Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.237188 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12864b75-dbdc-4075-aea7-9d3cc2a6ca59","Type":"ContainerDied","Data":"21eafeea4db54d3a522d2c5f0ab85cd2f58d653150e09d88c4504fb18084f5ba"} Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.237212 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12864b75-dbdc-4075-aea7-9d3cc2a6ca59","Type":"ContainerDied","Data":"5b0b554f9e776b0b4e1d869b769c0e6122573dae6943b653687ab9c3b73cafa6"} Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.237221 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12864b75-dbdc-4075-aea7-9d3cc2a6ca59","Type":"ContainerDied","Data":"2d8a19e074a59869fa182becc84c8c6153a9d08b9d345d4f7d9e637c21100594"} Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.249861 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" containerID="87a6a9e88e152395b58b4399a93fbcef7b312c11952ed845dc79a14c3e658cd6" exitCode=0 Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.250940 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" event={"ID":"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b","Type":"ContainerDied","Data":"87a6a9e88e152395b58b4399a93fbcef7b312c11952ed845dc79a14c3e658cd6"} Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.250991 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-npmm7"] Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.258649 4775 scope.go:117] "RemoveContainer" containerID="601abb7448e51003fa2c88bcae3cac775671f4277db926661d05d971e6098dfa" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.262791 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59d5ff467f-npmm7"] Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.404578 4775 scope.go:117] "RemoveContainer" containerID="7efddf689328bfd9c16bd39edd4e72f43dca2d239a2d1b5794ff0ed39eb5a8b5" Oct 02 02:00:51 crc kubenswrapper[4775]: E1002 02:00:51.405255 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7efddf689328bfd9c16bd39edd4e72f43dca2d239a2d1b5794ff0ed39eb5a8b5\": container with ID starting with 7efddf689328bfd9c16bd39edd4e72f43dca2d239a2d1b5794ff0ed39eb5a8b5 not found: ID does not exist" containerID="7efddf689328bfd9c16bd39edd4e72f43dca2d239a2d1b5794ff0ed39eb5a8b5" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.405286 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7efddf689328bfd9c16bd39edd4e72f43dca2d239a2d1b5794ff0ed39eb5a8b5"} err="failed to get container status \"7efddf689328bfd9c16bd39edd4e72f43dca2d239a2d1b5794ff0ed39eb5a8b5\": rpc error: code = NotFound desc = could not find container \"7efddf689328bfd9c16bd39edd4e72f43dca2d239a2d1b5794ff0ed39eb5a8b5\": container with ID starting with 7efddf689328bfd9c16bd39edd4e72f43dca2d239a2d1b5794ff0ed39eb5a8b5 not found: ID does not exist" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.405305 4775 scope.go:117] "RemoveContainer" containerID="601abb7448e51003fa2c88bcae3cac775671f4277db926661d05d971e6098dfa" Oct 02 02:00:51 crc kubenswrapper[4775]: E1002 02:00:51.406421 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"601abb7448e51003fa2c88bcae3cac775671f4277db926661d05d971e6098dfa\": container with ID starting with 601abb7448e51003fa2c88bcae3cac775671f4277db926661d05d971e6098dfa not found: ID does not exist" containerID="601abb7448e51003fa2c88bcae3cac775671f4277db926661d05d971e6098dfa" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.406442 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"601abb7448e51003fa2c88bcae3cac775671f4277db926661d05d971e6098dfa"} err="failed to get container status \"601abb7448e51003fa2c88bcae3cac775671f4277db926661d05d971e6098dfa\": rpc error: code = NotFound desc = could not find container \"601abb7448e51003fa2c88bcae3cac775671f4277db926661d05d971e6098dfa\": container with ID starting with 601abb7448e51003fa2c88bcae3cac775671f4277db926661d05d971e6098dfa not found: ID does not exist" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.571662 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hjvr2" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.668034 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-config\") pod \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\" (UID: \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\") " Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.668580 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-696hf\" (UniqueName: \"kubernetes.io/projected/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-kube-api-access-696hf\") pod \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\" (UID: \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\") " Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.668651 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-combined-ca-bundle\") pod \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\" (UID: \"c34b79a8-c071-4fff-8dbc-cab3601cc3f8\") " Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.672700 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-kube-api-access-696hf" (OuterVolumeSpecName: "kube-api-access-696hf") pod "c34b79a8-c071-4fff-8dbc-cab3601cc3f8" (UID: "c34b79a8-c071-4fff-8dbc-cab3601cc3f8"). InnerVolumeSpecName "kube-api-access-696hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.703080 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c34b79a8-c071-4fff-8dbc-cab3601cc3f8" (UID: "c34b79a8-c071-4fff-8dbc-cab3601cc3f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.704188 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-config" (OuterVolumeSpecName: "config") pod "c34b79a8-c071-4fff-8dbc-cab3601cc3f8" (UID: "c34b79a8-c071-4fff-8dbc-cab3601cc3f8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.773330 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-696hf\" (UniqueName: \"kubernetes.io/projected/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-kube-api-access-696hf\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.773360 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.773370 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/c34b79a8-c071-4fff-8dbc-cab3601cc3f8-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:51 crc kubenswrapper[4775]: I1002 02:00:51.795449 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f4eab32-163d-44a4-8cbd-66ca8968b869" path="/var/lib/kubelet/pods/1f4eab32-163d-44a4-8cbd-66ca8968b869/volumes" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.266813 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb","Type":"ContainerStarted","Data":"e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a"} Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.267136 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.267171 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" containerName="cinder-api-log" containerID="cri-o://775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3" gracePeriod=30 Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.267203 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" containerName="cinder-api" containerID="cri-o://e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a" gracePeriod=30 Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.271533 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hjvr2" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.271682 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hjvr2" event={"ID":"c34b79a8-c071-4fff-8dbc-cab3601cc3f8","Type":"ContainerDied","Data":"770927a770e1eae3d2c6eb6cc614d179524138eb7d78c21181fb7a76f2a356c6"} Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.271726 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="770927a770e1eae3d2c6eb6cc614d179524138eb7d78c21181fb7a76f2a356c6" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.279611 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c","Type":"ContainerStarted","Data":"26f726da7ed883ed5e95b1dbab9ff53a265b3c9bd474ea22b22b794a2a0aa4b8"} Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.287243 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" event={"ID":"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b","Type":"ContainerStarted","Data":"77935a60d3a9b8456b1efb4e6ab33d0c777b70d280b01b6905c3da842d902c87"} Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.287442 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.304010 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.303987845 podStartE2EDuration="4.303987845s" podCreationTimestamp="2025-10-02 02:00:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:52.301937182 +0000 UTC m=+1189.468681232" watchObservedRunningTime="2025-10-02 02:00:52.303987845 +0000 UTC m=+1189.470731895" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.326982 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" podStartSLOduration=4.326967396 podStartE2EDuration="4.326967396s" podCreationTimestamp="2025-10-02 02:00:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:52.32052485 +0000 UTC m=+1189.487268890" watchObservedRunningTime="2025-10-02 02:00:52.326967396 +0000 UTC m=+1189.493711436" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.413326 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:52 crc kubenswrapper[4775]: E1002 02:00:52.474645 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5bf2353_cc0f_4450_b237_e0aea4a9e7cb.slice/crio-conmon-775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3.scope\": RecentStats: unable to find data in memory cache]" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.489227 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-8gllc"] Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.525702 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-rf268"] Oct 02 02:00:52 crc kubenswrapper[4775]: E1002 02:00:52.526076 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c34b79a8-c071-4fff-8dbc-cab3601cc3f8" containerName="neutron-db-sync" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.526088 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c34b79a8-c071-4fff-8dbc-cab3601cc3f8" containerName="neutron-db-sync" Oct 02 02:00:52 crc kubenswrapper[4775]: E1002 02:00:52.526103 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f4eab32-163d-44a4-8cbd-66ca8968b869" containerName="init" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.526110 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f4eab32-163d-44a4-8cbd-66ca8968b869" containerName="init" Oct 02 02:00:52 crc kubenswrapper[4775]: E1002 02:00:52.526122 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f4eab32-163d-44a4-8cbd-66ca8968b869" containerName="dnsmasq-dns" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.526127 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f4eab32-163d-44a4-8cbd-66ca8968b869" containerName="dnsmasq-dns" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.526300 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c34b79a8-c071-4fff-8dbc-cab3601cc3f8" containerName="neutron-db-sync" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.526317 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f4eab32-163d-44a4-8cbd-66ca8968b869" containerName="dnsmasq-dns" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.532044 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.571064 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-749868ff64-jpnxz"] Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.576247 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.582263 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-qfnwq" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.582707 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.587138 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.598679 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.598741 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-dns-svc\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.598798 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hzwf\" (UniqueName: \"kubernetes.io/projected/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-kube-api-access-6hzwf\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.598856 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.598902 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.598962 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-config\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.599592 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.612547 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-rf268"] Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.636634 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-749868ff64-jpnxz"] Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.702649 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hzwf\" (UniqueName: \"kubernetes.io/projected/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-kube-api-access-6hzwf\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.702711 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-762zr\" (UniqueName: \"kubernetes.io/projected/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-kube-api-access-762zr\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.702740 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.702773 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-combined-ca-bundle\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.703418 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.703503 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-config\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.703540 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-ovndb-tls-certs\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.703577 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.703660 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-httpd-config\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.703709 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-config\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.703783 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.703835 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-dns-svc\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.704475 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-dns-svc\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.704526 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.705049 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-config\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.709081 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.722504 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hzwf\" (UniqueName: \"kubernetes.io/projected/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-kube-api-access-6hzwf\") pod \"dnsmasq-dns-5784cf869f-rf268\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.806624 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-762zr\" (UniqueName: \"kubernetes.io/projected/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-kube-api-access-762zr\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.806687 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-combined-ca-bundle\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.810709 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-config\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.811000 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-ovndb-tls-certs\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.811199 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-httpd-config\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.814635 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-config\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.817577 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-httpd-config\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.831789 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-762zr\" (UniqueName: \"kubernetes.io/projected/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-kube-api-access-762zr\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.838340 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-combined-ca-bundle\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.843578 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-ovndb-tls-certs\") pod \"neutron-749868ff64-jpnxz\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.881402 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:52 crc kubenswrapper[4775]: I1002 02:00:52.998086 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.102034 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.219642 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-config-data-custom\") pod \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.219735 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-scripts\") pod \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.219757 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-config-data\") pod \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.219809 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j46d7\" (UniqueName: \"kubernetes.io/projected/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-kube-api-access-j46d7\") pod \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.219848 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-logs\") pod \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.219909 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-combined-ca-bundle\") pod \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.219929 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-etc-machine-id\") pod \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\" (UID: \"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb\") " Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.220360 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" (UID: "b5bf2353-cc0f-4450-b237-e0aea4a9e7cb"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.221740 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-logs" (OuterVolumeSpecName: "logs") pod "b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" (UID: "b5bf2353-cc0f-4450-b237-e0aea4a9e7cb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.227330 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-scripts" (OuterVolumeSpecName: "scripts") pod "b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" (UID: "b5bf2353-cc0f-4450-b237-e0aea4a9e7cb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.228370 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-kube-api-access-j46d7" (OuterVolumeSpecName: "kube-api-access-j46d7") pod "b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" (UID: "b5bf2353-cc0f-4450-b237-e0aea4a9e7cb"). InnerVolumeSpecName "kube-api-access-j46d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.231039 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" (UID: "b5bf2353-cc0f-4450-b237-e0aea4a9e7cb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.263519 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" (UID: "b5bf2353-cc0f-4450-b237-e0aea4a9e7cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.299869 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-config-data" (OuterVolumeSpecName: "config-data") pod "b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" (UID: "b5bf2353-cc0f-4450-b237-e0aea4a9e7cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.322136 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.322174 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.322186 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.322198 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j46d7\" (UniqueName: \"kubernetes.io/projected/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-kube-api-access-j46d7\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.322213 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.322227 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.322239 4775 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.337735 4775 generic.go:334] "Generic (PLEG): container finished" podID="b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" containerID="e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a" exitCode=0 Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.337764 4775 generic.go:334] "Generic (PLEG): container finished" podID="b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" containerID="775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3" exitCode=143 Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.337807 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.337814 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb","Type":"ContainerDied","Data":"e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a"} Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.337842 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb","Type":"ContainerDied","Data":"775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3"} Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.337852 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b5bf2353-cc0f-4450-b237-e0aea4a9e7cb","Type":"ContainerDied","Data":"de4b034df51edadffee7c77be51543d58ed4bdf255cbf23cfcab7d1b11981917"} Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.337867 4775 scope.go:117] "RemoveContainer" containerID="e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.361692 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c","Type":"ContainerStarted","Data":"032a92c01d2c434edf83435f1eaf6ace4fd8cdf6d79e46caa4ff3e6b24d459ed"} Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.378681 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.386752 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.395179 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.526709926 podStartE2EDuration="5.395159426s" podCreationTimestamp="2025-10-02 02:00:48 +0000 UTC" firstStartedPulling="2025-10-02 02:00:49.920987692 +0000 UTC m=+1187.087731732" lastFinishedPulling="2025-10-02 02:00:50.789437192 +0000 UTC m=+1187.956181232" observedRunningTime="2025-10-02 02:00:53.393504484 +0000 UTC m=+1190.560248524" watchObservedRunningTime="2025-10-02 02:00:53.395159426 +0000 UTC m=+1190.561903466" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.413231 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:00:53 crc kubenswrapper[4775]: E1002 02:00:53.413714 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" containerName="cinder-api" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.413730 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" containerName="cinder-api" Oct 02 02:00:53 crc kubenswrapper[4775]: E1002 02:00:53.413743 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" containerName="cinder-api-log" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.413750 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" containerName="cinder-api-log" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.413911 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" containerName="cinder-api-log" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.413922 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" containerName="cinder-api" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.414882 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.419253 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.419396 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.419905 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.441059 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-rf268"] Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.441667 4775 scope.go:117] "RemoveContainer" containerID="775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.447038 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.463424 4775 scope.go:117] "RemoveContainer" containerID="e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a" Oct 02 02:00:53 crc kubenswrapper[4775]: E1002 02:00:53.463931 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a\": container with ID starting with e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a not found: ID does not exist" containerID="e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.463978 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a"} err="failed to get container status \"e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a\": rpc error: code = NotFound desc = could not find container \"e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a\": container with ID starting with e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a not found: ID does not exist" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.464002 4775 scope.go:117] "RemoveContainer" containerID="775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3" Oct 02 02:00:53 crc kubenswrapper[4775]: E1002 02:00:53.464266 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3\": container with ID starting with 775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3 not found: ID does not exist" containerID="775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.464316 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3"} err="failed to get container status \"775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3\": rpc error: code = NotFound desc = could not find container \"775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3\": container with ID starting with 775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3 not found: ID does not exist" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.464348 4775 scope.go:117] "RemoveContainer" containerID="e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.464678 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a"} err="failed to get container status \"e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a\": rpc error: code = NotFound desc = could not find container \"e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a\": container with ID starting with e0cf07aa38679fb78c1f7c67d56c289c37956e5cf34d27cc32d3a4f259b41e5a not found: ID does not exist" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.464706 4775 scope.go:117] "RemoveContainer" containerID="775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.465177 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3"} err="failed to get container status \"775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3\": rpc error: code = NotFound desc = could not find container \"775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3\": container with ID starting with 775d5828cf19bd1eb5ff94b009e88badf4cbb65331e6e3e0044e4172465d6cf3 not found: ID does not exist" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.547607 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7cl5\" (UniqueName: \"kubernetes.io/projected/5bca47d0-79b7-4992-867b-d5e05e02cdc8-kube-api-access-x7cl5\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.547711 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.547736 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-scripts\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.547777 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-config-data-custom\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.547816 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-config-data\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.547833 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.547861 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5bca47d0-79b7-4992-867b-d5e05e02cdc8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.548001 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.548090 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bca47d0-79b7-4992-867b-d5e05e02cdc8-logs\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.596645 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-749868ff64-jpnxz"] Oct 02 02:00:53 crc kubenswrapper[4775]: W1002 02:00:53.609244 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb7c56b0_e35f_4fb9_9746_71fe1be1b027.slice/crio-0d9edc310c8f5745221c21900728828e618c061e1c09f2f5eade243c870abd9e WatchSource:0}: Error finding container 0d9edc310c8f5745221c21900728828e618c061e1c09f2f5eade243c870abd9e: Status 404 returned error can't find the container with id 0d9edc310c8f5745221c21900728828e618c061e1c09f2f5eade243c870abd9e Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.649924 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.649984 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-scripts\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.650016 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-config-data-custom\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.650056 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-config-data\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.650072 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.650106 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5bca47d0-79b7-4992-867b-d5e05e02cdc8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.650130 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.650164 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bca47d0-79b7-4992-867b-d5e05e02cdc8-logs\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.650221 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7cl5\" (UniqueName: \"kubernetes.io/projected/5bca47d0-79b7-4992-867b-d5e05e02cdc8-kube-api-access-x7cl5\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.650463 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5bca47d0-79b7-4992-867b-d5e05e02cdc8-etc-machine-id\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.651106 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bca47d0-79b7-4992-867b-d5e05e02cdc8-logs\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.657795 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-public-tls-certs\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.663591 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-scripts\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.665553 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-config-data-custom\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.671726 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.672201 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7cl5\" (UniqueName: \"kubernetes.io/projected/5bca47d0-79b7-4992-867b-d5e05e02cdc8-kube-api-access-x7cl5\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.672236 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-config-data\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.674101 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.746020 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.746663 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 02:00:53 crc kubenswrapper[4775]: I1002 02:00:53.788942 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5bf2353-cc0f-4450-b237-e0aea4a9e7cb" path="/var/lib/kubelet/pods/b5bf2353-cc0f-4450-b237-e0aea4a9e7cb/volumes" Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.054468 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.368970 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5bca47d0-79b7-4992-867b-d5e05e02cdc8","Type":"ContainerStarted","Data":"0bc9445b2e060fd2d8cafb7bc36635d4aca57867fb9172aed3eebad1b3cd6ef2"} Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.371431 4775 generic.go:334] "Generic (PLEG): container finished" podID="9aa78020-05ca-40e3-abd3-d7a960b8b2ad" containerID="9e75c63ad61e752ebf2677a4126de52817646ff1a38139c6459d1e49b6e73721" exitCode=0 Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.371501 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-rf268" event={"ID":"9aa78020-05ca-40e3-abd3-d7a960b8b2ad","Type":"ContainerDied","Data":"9e75c63ad61e752ebf2677a4126de52817646ff1a38139c6459d1e49b6e73721"} Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.371527 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-rf268" event={"ID":"9aa78020-05ca-40e3-abd3-d7a960b8b2ad","Type":"ContainerStarted","Data":"4757c352696a93e19189bf85ac66b6319d45160727f5310ed6fff72f7634afe8"} Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.376284 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-749868ff64-jpnxz" event={"ID":"bb7c56b0-e35f-4fb9-9746-71fe1be1b027","Type":"ContainerStarted","Data":"a9ab79565384ff2e9f02e123a0e766415185a2be063ed2fb25fd5db22c85c64f"} Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.376322 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-749868ff64-jpnxz" event={"ID":"bb7c56b0-e35f-4fb9-9746-71fe1be1b027","Type":"ContainerStarted","Data":"2b17fcefad2d7e332e250dd41af49214922dfec671fef00c94a8e45d9d0d89ed"} Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.376348 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-749868ff64-jpnxz" event={"ID":"bb7c56b0-e35f-4fb9-9746-71fe1be1b027","Type":"ContainerStarted","Data":"0d9edc310c8f5745221c21900728828e618c061e1c09f2f5eade243c870abd9e"} Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.377012 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.380547 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" podUID="e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" containerName="dnsmasq-dns" containerID="cri-o://77935a60d3a9b8456b1efb4e6ab33d0c777b70d280b01b6905c3da842d902c87" gracePeriod=10 Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.415107 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-749868ff64-jpnxz" podStartSLOduration=2.415093643 podStartE2EDuration="2.415093643s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:54.411785218 +0000 UTC m=+1191.578529258" watchObservedRunningTime="2025-10-02 02:00:54.415093643 +0000 UTC m=+1191.581837673" Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.774687 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.869784 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-ovsdbserver-sb\") pod \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.869858 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-ovsdbserver-nb\") pod \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.869896 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-config\") pod \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.869940 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-dns-swift-storage-0\") pod \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.870027 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-dns-svc\") pod \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.870059 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fns7m\" (UniqueName: \"kubernetes.io/projected/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-kube-api-access-fns7m\") pod \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\" (UID: \"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b\") " Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.892765 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-kube-api-access-fns7m" (OuterVolumeSpecName: "kube-api-access-fns7m") pod "e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" (UID: "e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b"). InnerVolumeSpecName "kube-api-access-fns7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.928315 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" (UID: "e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.934133 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" (UID: "e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.936613 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" (UID: "e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.950794 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" (UID: "e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.973399 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.973434 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.973449 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.973458 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fns7m\" (UniqueName: \"kubernetes.io/projected/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-kube-api-access-fns7m\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.973470 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:54 crc kubenswrapper[4775]: I1002 02:00:54.991118 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-config" (OuterVolumeSpecName: "config") pod "e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" (UID: "e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.074975 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.214523 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-546f7b9c49-h6rcs"] Oct 02 02:00:55 crc kubenswrapper[4775]: E1002 02:00:55.215154 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" containerName="init" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.215173 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" containerName="init" Oct 02 02:00:55 crc kubenswrapper[4775]: E1002 02:00:55.215201 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" containerName="dnsmasq-dns" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.215208 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" containerName="dnsmasq-dns" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.215399 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" containerName="dnsmasq-dns" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.216326 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.224572 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.224742 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.255127 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-546f7b9c49-h6rcs"] Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.280480 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-httpd-config\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.280531 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm5nn\" (UniqueName: \"kubernetes.io/projected/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-kube-api-access-lm5nn\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.280562 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-public-tls-certs\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.280599 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-config\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.280628 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-internal-tls-certs\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.280669 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-ovndb-tls-certs\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.280698 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-combined-ca-bundle\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.381974 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-httpd-config\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.382037 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm5nn\" (UniqueName: \"kubernetes.io/projected/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-kube-api-access-lm5nn\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.382070 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-public-tls-certs\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.382103 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-config\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.382159 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-internal-tls-certs\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.382204 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-ovndb-tls-certs\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.382243 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-combined-ca-bundle\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.386444 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-public-tls-certs\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.387835 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-internal-tls-certs\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.399316 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-ovndb-tls-certs\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.404613 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm5nn\" (UniqueName: \"kubernetes.io/projected/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-kube-api-access-lm5nn\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.405528 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-config\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.407705 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-combined-ca-bundle\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.409723 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-rf268" event={"ID":"9aa78020-05ca-40e3-abd3-d7a960b8b2ad","Type":"ContainerStarted","Data":"c7e1a112ab7147ecb551a7b9f6aef7efdc4bb4197ccd00766ebfc02583672984"} Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.410728 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.414183 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-httpd-config\") pod \"neutron-546f7b9c49-h6rcs\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.428455 4775 generic.go:334] "Generic (PLEG): container finished" podID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerID="89a042d772657ba277e82d383099114e18ac75f765966f3219b0c0ca7ba4fc91" exitCode=0 Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.428530 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12864b75-dbdc-4075-aea7-9d3cc2a6ca59","Type":"ContainerDied","Data":"89a042d772657ba277e82d383099114e18ac75f765966f3219b0c0ca7ba4fc91"} Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.437927 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5bca47d0-79b7-4992-867b-d5e05e02cdc8","Type":"ContainerStarted","Data":"04d6bdf03305dd966a2ca95b6a3220b41a5add02a16becda008422ee10720237"} Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.440558 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" containerID="77935a60d3a9b8456b1efb4e6ab33d0c777b70d280b01b6905c3da842d902c87" exitCode=0 Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.441163 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.450260 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-rf268" podStartSLOduration=3.450241333 podStartE2EDuration="3.450241333s" podCreationTimestamp="2025-10-02 02:00:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:55.449485894 +0000 UTC m=+1192.616229934" watchObservedRunningTime="2025-10-02 02:00:55.450241333 +0000 UTC m=+1192.616985383" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.451448 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" event={"ID":"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b","Type":"ContainerDied","Data":"77935a60d3a9b8456b1efb4e6ab33d0c777b70d280b01b6905c3da842d902c87"} Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.451505 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69c986f6d7-8gllc" event={"ID":"e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b","Type":"ContainerDied","Data":"34793f825a3c40a0ff7049782635751392f23e10053596c1325c51cb316d8a20"} Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.451524 4775 scope.go:117] "RemoveContainer" containerID="77935a60d3a9b8456b1efb4e6ab33d0c777b70d280b01b6905c3da842d902c87" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.504051 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-8gllc"] Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.522138 4775 scope.go:117] "RemoveContainer" containerID="87a6a9e88e152395b58b4399a93fbcef7b312c11952ed845dc79a14c3e658cd6" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.528233 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69c986f6d7-8gllc"] Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.548287 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.567515 4775 scope.go:117] "RemoveContainer" containerID="77935a60d3a9b8456b1efb4e6ab33d0c777b70d280b01b6905c3da842d902c87" Oct 02 02:00:55 crc kubenswrapper[4775]: E1002 02:00:55.569947 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77935a60d3a9b8456b1efb4e6ab33d0c777b70d280b01b6905c3da842d902c87\": container with ID starting with 77935a60d3a9b8456b1efb4e6ab33d0c777b70d280b01b6905c3da842d902c87 not found: ID does not exist" containerID="77935a60d3a9b8456b1efb4e6ab33d0c777b70d280b01b6905c3da842d902c87" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.569995 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77935a60d3a9b8456b1efb4e6ab33d0c777b70d280b01b6905c3da842d902c87"} err="failed to get container status \"77935a60d3a9b8456b1efb4e6ab33d0c777b70d280b01b6905c3da842d902c87\": rpc error: code = NotFound desc = could not find container \"77935a60d3a9b8456b1efb4e6ab33d0c777b70d280b01b6905c3da842d902c87\": container with ID starting with 77935a60d3a9b8456b1efb4e6ab33d0c777b70d280b01b6905c3da842d902c87 not found: ID does not exist" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.570016 4775 scope.go:117] "RemoveContainer" containerID="87a6a9e88e152395b58b4399a93fbcef7b312c11952ed845dc79a14c3e658cd6" Oct 02 02:00:55 crc kubenswrapper[4775]: E1002 02:00:55.573578 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87a6a9e88e152395b58b4399a93fbcef7b312c11952ed845dc79a14c3e658cd6\": container with ID starting with 87a6a9e88e152395b58b4399a93fbcef7b312c11952ed845dc79a14c3e658cd6 not found: ID does not exist" containerID="87a6a9e88e152395b58b4399a93fbcef7b312c11952ed845dc79a14c3e658cd6" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.573600 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87a6a9e88e152395b58b4399a93fbcef7b312c11952ed845dc79a14c3e658cd6"} err="failed to get container status \"87a6a9e88e152395b58b4399a93fbcef7b312c11952ed845dc79a14c3e658cd6\": rpc error: code = NotFound desc = could not find container \"87a6a9e88e152395b58b4399a93fbcef7b312c11952ed845dc79a14c3e658cd6\": container with ID starting with 87a6a9e88e152395b58b4399a93fbcef7b312c11952ed845dc79a14c3e658cd6 not found: ID does not exist" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.633212 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.687875 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-log-httpd\") pod \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.687924 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpdgg\" (UniqueName: \"kubernetes.io/projected/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-kube-api-access-wpdgg\") pod \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.687974 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-sg-core-conf-yaml\") pod \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.688034 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-combined-ca-bundle\") pod \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.688063 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-scripts\") pod \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.688124 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-run-httpd\") pod \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.688144 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-config-data\") pod \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\" (UID: \"12864b75-dbdc-4075-aea7-9d3cc2a6ca59\") " Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.696110 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "12864b75-dbdc-4075-aea7-9d3cc2a6ca59" (UID: "12864b75-dbdc-4075-aea7-9d3cc2a6ca59"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.696391 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "12864b75-dbdc-4075-aea7-9d3cc2a6ca59" (UID: "12864b75-dbdc-4075-aea7-9d3cc2a6ca59"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.707320 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-kube-api-access-wpdgg" (OuterVolumeSpecName: "kube-api-access-wpdgg") pod "12864b75-dbdc-4075-aea7-9d3cc2a6ca59" (UID: "12864b75-dbdc-4075-aea7-9d3cc2a6ca59"). InnerVolumeSpecName "kube-api-access-wpdgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.715385 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-scripts" (OuterVolumeSpecName: "scripts") pod "12864b75-dbdc-4075-aea7-9d3cc2a6ca59" (UID: "12864b75-dbdc-4075-aea7-9d3cc2a6ca59"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.759848 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "12864b75-dbdc-4075-aea7-9d3cc2a6ca59" (UID: "12864b75-dbdc-4075-aea7-9d3cc2a6ca59"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.779850 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12864b75-dbdc-4075-aea7-9d3cc2a6ca59" (UID: "12864b75-dbdc-4075-aea7-9d3cc2a6ca59"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.806011 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.806045 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.806064 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.806083 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpdgg\" (UniqueName: \"kubernetes.io/projected/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-kube-api-access-wpdgg\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.806100 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.806108 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.813015 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b" path="/var/lib/kubelet/pods/e3a1f94c-9ace-4f3d-89b0-1ef8f3cd057b/volumes" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.830117 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-config-data" (OuterVolumeSpecName: "config-data") pod "12864b75-dbdc-4075-aea7-9d3cc2a6ca59" (UID: "12864b75-dbdc-4075-aea7-9d3cc2a6ca59"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:00:55 crc kubenswrapper[4775]: I1002 02:00:55.911046 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12864b75-dbdc-4075-aea7-9d3cc2a6ca59-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.165474 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-546f7b9c49-h6rcs"] Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.452032 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-546f7b9c49-h6rcs" event={"ID":"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab","Type":"ContainerStarted","Data":"0e053b46f04d55b9e7591b085c54f9c8a069336ff3dd6b39e07e7d39e1ed955f"} Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.456613 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"12864b75-dbdc-4075-aea7-9d3cc2a6ca59","Type":"ContainerDied","Data":"d8f1d1d00b3490c5599f7fe0eaeb40f3a76aca72ccbf022e9d39d0478cbf05f2"} Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.456637 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.456698 4775 scope.go:117] "RemoveContainer" containerID="21eafeea4db54d3a522d2c5f0ab85cd2f58d653150e09d88c4504fb18084f5ba" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.461085 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5bca47d0-79b7-4992-867b-d5e05e02cdc8","Type":"ContainerStarted","Data":"e38c74d33b3ad7b8f747e53560c787d296c45bf29b3227467260f620932b62a9"} Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.461200 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.498891 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.49886651 podStartE2EDuration="3.49886651s" podCreationTimestamp="2025-10-02 02:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:56.489298424 +0000 UTC m=+1193.656042474" watchObservedRunningTime="2025-10-02 02:00:56.49886651 +0000 UTC m=+1193.665610580" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.517308 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.531066 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.547019 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:00:56 crc kubenswrapper[4775]: E1002 02:00:56.547417 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="proxy-httpd" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.547433 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="proxy-httpd" Oct 02 02:00:56 crc kubenswrapper[4775]: E1002 02:00:56.547447 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="sg-core" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.547453 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="sg-core" Oct 02 02:00:56 crc kubenswrapper[4775]: E1002 02:00:56.547478 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="ceilometer-notification-agent" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.547484 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="ceilometer-notification-agent" Oct 02 02:00:56 crc kubenswrapper[4775]: E1002 02:00:56.547503 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="ceilometer-central-agent" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.547509 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="ceilometer-central-agent" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.547679 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="ceilometer-central-agent" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.547697 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="sg-core" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.547706 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="ceilometer-notification-agent" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.547717 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" containerName="proxy-httpd" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.549228 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.551585 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.551914 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.557938 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.622157 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18a90265-099f-4da5-9d05-32956ab2deb4-run-httpd\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.622215 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-scripts\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.622409 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.622988 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18a90265-099f-4da5-9d05-32956ab2deb4-log-httpd\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.623144 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcbf5\" (UniqueName: \"kubernetes.io/projected/18a90265-099f-4da5-9d05-32956ab2deb4-kube-api-access-wcbf5\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.625094 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-config-data\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.626161 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.732120 4775 scope.go:117] "RemoveContainer" containerID="5b0b554f9e776b0b4e1d869b769c0e6122573dae6943b653687ab9c3b73cafa6" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.732905 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18a90265-099f-4da5-9d05-32956ab2deb4-run-httpd\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.732942 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-scripts\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.733007 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.733042 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18a90265-099f-4da5-9d05-32956ab2deb4-log-httpd\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.733064 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcbf5\" (UniqueName: \"kubernetes.io/projected/18a90265-099f-4da5-9d05-32956ab2deb4-kube-api-access-wcbf5\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.733084 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-config-data\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.733126 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.734194 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18a90265-099f-4da5-9d05-32956ab2deb4-log-httpd\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.734325 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18a90265-099f-4da5-9d05-32956ab2deb4-run-httpd\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.739259 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.740291 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-config-data\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.740335 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.743338 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.751120 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-scripts\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.758681 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcbf5\" (UniqueName: \"kubernetes.io/projected/18a90265-099f-4da5-9d05-32956ab2deb4-kube-api-access-wcbf5\") pod \"ceilometer-0\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.792552 4775 scope.go:117] "RemoveContainer" containerID="89a042d772657ba277e82d383099114e18ac75f765966f3219b0c0ca7ba4fc91" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.815573 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.832471 4775 scope.go:117] "RemoveContainer" containerID="2d8a19e074a59869fa182becc84c8c6153a9d08b9d345d4f7d9e637c21100594" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.878167 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5698978bb-cdg88"] Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.880284 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.880936 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5698978bb-cdg88" podUID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" containerName="barbican-api-log" containerID="cri-o://250bae0449846fc4dda5adff85b5ed10a3b0e5cffbbb6251ffe6503a2a43110e" gracePeriod=30 Oct 02 02:00:56 crc kubenswrapper[4775]: I1002 02:00:56.881609 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-5698978bb-cdg88" podUID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" containerName="barbican-api" containerID="cri-o://169652bfa5bde2f7471ba1a8910ef01f1640b2f0a9daf2c6c26ac42892314dbf" gracePeriod=30 Oct 02 02:00:57 crc kubenswrapper[4775]: I1002 02:00:57.408291 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:00:57 crc kubenswrapper[4775]: W1002 02:00:57.415135 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod18a90265_099f_4da5_9d05_32956ab2deb4.slice/crio-eb40623c790af2cb083402b5d65cfe5b17f1e9e6d8e17b2b8d1f84159b45c20e WatchSource:0}: Error finding container eb40623c790af2cb083402b5d65cfe5b17f1e9e6d8e17b2b8d1f84159b45c20e: Status 404 returned error can't find the container with id eb40623c790af2cb083402b5d65cfe5b17f1e9e6d8e17b2b8d1f84159b45c20e Oct 02 02:00:57 crc kubenswrapper[4775]: I1002 02:00:57.482564 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-546f7b9c49-h6rcs" event={"ID":"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab","Type":"ContainerStarted","Data":"85353a3b8a9a2c508bfbb1d8f15bdea36723aba6e9d1ddda8f826588f795b512"} Oct 02 02:00:57 crc kubenswrapper[4775]: I1002 02:00:57.482607 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-546f7b9c49-h6rcs" event={"ID":"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab","Type":"ContainerStarted","Data":"8c9ca1dbcfc1426fde26f45e447e342f40b762f7d726a5c8194371f054ef7feb"} Oct 02 02:00:57 crc kubenswrapper[4775]: I1002 02:00:57.482639 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:00:57 crc kubenswrapper[4775]: I1002 02:00:57.493538 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18a90265-099f-4da5-9d05-32956ab2deb4","Type":"ContainerStarted","Data":"eb40623c790af2cb083402b5d65cfe5b17f1e9e6d8e17b2b8d1f84159b45c20e"} Oct 02 02:00:57 crc kubenswrapper[4775]: I1002 02:00:57.508731 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-546f7b9c49-h6rcs" podStartSLOduration=2.508717888 podStartE2EDuration="2.508717888s" podCreationTimestamp="2025-10-02 02:00:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:00:57.507423314 +0000 UTC m=+1194.674167354" watchObservedRunningTime="2025-10-02 02:00:57.508717888 +0000 UTC m=+1194.675461928" Oct 02 02:00:57 crc kubenswrapper[4775]: I1002 02:00:57.511273 4775 generic.go:334] "Generic (PLEG): container finished" podID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" containerID="250bae0449846fc4dda5adff85b5ed10a3b0e5cffbbb6251ffe6503a2a43110e" exitCode=143 Oct 02 02:00:57 crc kubenswrapper[4775]: I1002 02:00:57.511363 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5698978bb-cdg88" event={"ID":"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e","Type":"ContainerDied","Data":"250bae0449846fc4dda5adff85b5ed10a3b0e5cffbbb6251ffe6503a2a43110e"} Oct 02 02:00:57 crc kubenswrapper[4775]: I1002 02:00:57.775923 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12864b75-dbdc-4075-aea7-9d3cc2a6ca59" path="/var/lib/kubelet/pods/12864b75-dbdc-4075-aea7-9d3cc2a6ca59/volumes" Oct 02 02:00:58 crc kubenswrapper[4775]: I1002 02:00:58.540972 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18a90265-099f-4da5-9d05-32956ab2deb4","Type":"ContainerStarted","Data":"841211e5dddbb7078cf26acd9820c0e489445c62fa82375fb5b93d954c33ee0b"} Oct 02 02:00:58 crc kubenswrapper[4775]: I1002 02:00:58.591450 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:00:58 crc kubenswrapper[4775]: I1002 02:00:58.939921 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 02:00:58 crc kubenswrapper[4775]: I1002 02:00:58.986050 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:00:59 crc kubenswrapper[4775]: I1002 02:00:59.554985 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18a90265-099f-4da5-9d05-32956ab2deb4","Type":"ContainerStarted","Data":"d62fda876ad3ecd35bd593f1776e6700e2a6b1b84104fa540031723876198146"} Oct 02 02:00:59 crc kubenswrapper[4775]: I1002 02:00:59.555271 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" containerName="cinder-scheduler" containerID="cri-o://26f726da7ed883ed5e95b1dbab9ff53a265b3c9bd474ea22b22b794a2a0aa4b8" gracePeriod=30 Oct 02 02:00:59 crc kubenswrapper[4775]: I1002 02:00:59.555305 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" containerName="probe" containerID="cri-o://032a92c01d2c434edf83435f1eaf6ace4fd8cdf6d79e46caa4ff3e6b24d459ed" gracePeriod=30 Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.163895 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29322841-j7zjm"] Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.165321 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.190040 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29322841-j7zjm"] Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.314685 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-fernet-keys\") pod \"keystone-cron-29322841-j7zjm\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.314778 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-combined-ca-bundle\") pod \"keystone-cron-29322841-j7zjm\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.314834 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-config-data\") pod \"keystone-cron-29322841-j7zjm\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.314879 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7jxd\" (UniqueName: \"kubernetes.io/projected/97a109cd-fde0-409f-a0cc-e2e3c78acb47-kube-api-access-m7jxd\") pod \"keystone-cron-29322841-j7zjm\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.417296 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-fernet-keys\") pod \"keystone-cron-29322841-j7zjm\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.417609 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-combined-ca-bundle\") pod \"keystone-cron-29322841-j7zjm\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.417642 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-config-data\") pod \"keystone-cron-29322841-j7zjm\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.417680 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7jxd\" (UniqueName: \"kubernetes.io/projected/97a109cd-fde0-409f-a0cc-e2e3c78acb47-kube-api-access-m7jxd\") pod \"keystone-cron-29322841-j7zjm\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.423881 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-fernet-keys\") pod \"keystone-cron-29322841-j7zjm\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.424023 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-config-data\") pod \"keystone-cron-29322841-j7zjm\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.424430 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-combined-ca-bundle\") pod \"keystone-cron-29322841-j7zjm\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.434505 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7jxd\" (UniqueName: \"kubernetes.io/projected/97a109cd-fde0-409f-a0cc-e2e3c78acb47-kube-api-access-m7jxd\") pod \"keystone-cron-29322841-j7zjm\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.504786 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.571513 4775 generic.go:334] "Generic (PLEG): container finished" podID="3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" containerID="032a92c01d2c434edf83435f1eaf6ace4fd8cdf6d79e46caa4ff3e6b24d459ed" exitCode=0 Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.571728 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c","Type":"ContainerDied","Data":"032a92c01d2c434edf83435f1eaf6ace4fd8cdf6d79e46caa4ff3e6b24d459ed"} Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.574308 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18a90265-099f-4da5-9d05-32956ab2deb4","Type":"ContainerStarted","Data":"5fae55178786be08f69283be73c0f41e9455f7929793ecfbb705a70df3e98fc4"} Oct 02 02:01:00 crc kubenswrapper[4775]: I1002 02:01:00.954443 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29322841-j7zjm"] Oct 02 02:01:00 crc kubenswrapper[4775]: W1002 02:01:00.960469 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97a109cd_fde0_409f_a0cc_e2e3c78acb47.slice/crio-8b6a0c1861a2eebc67cee3161f50961bcdc51b8c7448716af466c8baf3d44627 WatchSource:0}: Error finding container 8b6a0c1861a2eebc67cee3161f50961bcdc51b8c7448716af466c8baf3d44627: Status 404 returned error can't find the container with id 8b6a0c1861a2eebc67cee3161f50961bcdc51b8c7448716af466c8baf3d44627 Oct 02 02:01:01 crc kubenswrapper[4775]: I1002 02:01:01.587440 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18a90265-099f-4da5-9d05-32956ab2deb4","Type":"ContainerStarted","Data":"f4bc65323b150aac0a5fa8b7190be3cebed294ecdde611749ce8736c77e1c19c"} Oct 02 02:01:01 crc kubenswrapper[4775]: I1002 02:01:01.587864 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:01:01 crc kubenswrapper[4775]: I1002 02:01:01.590912 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322841-j7zjm" event={"ID":"97a109cd-fde0-409f-a0cc-e2e3c78acb47","Type":"ContainerStarted","Data":"11fc048e233fc00c87c2d04d492ad8e05971c6d7aac57380af9bd95f239d9a3d"} Oct 02 02:01:01 crc kubenswrapper[4775]: I1002 02:01:01.591012 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322841-j7zjm" event={"ID":"97a109cd-fde0-409f-a0cc-e2e3c78acb47","Type":"ContainerStarted","Data":"8b6a0c1861a2eebc67cee3161f50961bcdc51b8c7448716af466c8baf3d44627"} Oct 02 02:01:01 crc kubenswrapper[4775]: I1002 02:01:01.610352 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.138840966 podStartE2EDuration="5.610337293s" podCreationTimestamp="2025-10-02 02:00:56 +0000 UTC" firstStartedPulling="2025-10-02 02:00:57.417542952 +0000 UTC m=+1194.584286982" lastFinishedPulling="2025-10-02 02:01:00.889039269 +0000 UTC m=+1198.055783309" observedRunningTime="2025-10-02 02:01:01.60401148 +0000 UTC m=+1198.770755520" watchObservedRunningTime="2025-10-02 02:01:01.610337293 +0000 UTC m=+1198.777081333" Oct 02 02:01:01 crc kubenswrapper[4775]: I1002 02:01:01.619719 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29322841-j7zjm" podStartSLOduration=1.6197083939999999 podStartE2EDuration="1.619708394s" podCreationTimestamp="2025-10-02 02:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:01:01.617470417 +0000 UTC m=+1198.784214497" watchObservedRunningTime="2025-10-02 02:01:01.619708394 +0000 UTC m=+1198.786452434" Oct 02 02:01:01 crc kubenswrapper[4775]: I1002 02:01:01.720893 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-564487c87-zppvw" Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.016425 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5698978bb-cdg88" podUID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.294502 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5698978bb-cdg88" podUID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:55024->10.217.0.156:9311: read: connection reset by peer" Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.294551 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5698978bb-cdg88" podUID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:49910->10.217.0.156:9311: read: connection reset by peer" Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.295050 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-5698978bb-cdg88" podUID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": dial tcp 10.217.0.156:9311: connect: connection refused" Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.628265 4775 generic.go:334] "Generic (PLEG): container finished" podID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" containerID="169652bfa5bde2f7471ba1a8910ef01f1640b2f0a9daf2c6c26ac42892314dbf" exitCode=0 Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.631019 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5698978bb-cdg88" event={"ID":"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e","Type":"ContainerDied","Data":"169652bfa5bde2f7471ba1a8910ef01f1640b2f0a9daf2c6c26ac42892314dbf"} Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.828543 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.884073 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.954034 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-bczkw"] Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.954286 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" podUID="9e78aa9b-9107-461c-b94f-c0f166b489b4" containerName="dnsmasq-dns" containerID="cri-o://10a348acd950eaad46c525129402d791195a91d90d80d01aa576064a8aeec468" gracePeriod=10 Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.962557 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-config-data\") pod \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.962623 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-combined-ca-bundle\") pod \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.962759 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-config-data-custom\") pod \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.963593 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-logs\") pod \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.963643 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjd2t\" (UniqueName: \"kubernetes.io/projected/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-kube-api-access-pjd2t\") pod \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\" (UID: \"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e\") " Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.964755 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-logs" (OuterVolumeSpecName: "logs") pod "4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" (UID: "4f876ea1-9d84-4dd0-a2f6-2bc742bab66e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.979176 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-kube-api-access-pjd2t" (OuterVolumeSpecName: "kube-api-access-pjd2t") pod "4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" (UID: "4f876ea1-9d84-4dd0-a2f6-2bc742bab66e"). InnerVolumeSpecName "kube-api-access-pjd2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:02 crc kubenswrapper[4775]: I1002 02:01:02.979202 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" (UID: "4f876ea1-9d84-4dd0-a2f6-2bc742bab66e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.003274 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" (UID: "4f876ea1-9d84-4dd0-a2f6-2bc742bab66e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.060101 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-config-data" (OuterVolumeSpecName: "config-data") pod "4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" (UID: "4f876ea1-9d84-4dd0-a2f6-2bc742bab66e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.068239 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.068287 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.068297 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjd2t\" (UniqueName: \"kubernetes.io/projected/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-kube-api-access-pjd2t\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.068308 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.068317 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.350600 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.474624 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6prjn\" (UniqueName: \"kubernetes.io/projected/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-kube-api-access-6prjn\") pod \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.474708 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-combined-ca-bundle\") pod \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.474736 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-config-data\") pod \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.474760 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-config-data-custom\") pod \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.474829 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-scripts\") pod \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.474905 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-etc-machine-id\") pod \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\" (UID: \"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c\") " Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.475383 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" (UID: "3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.475713 4775 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.480285 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-scripts" (OuterVolumeSpecName: "scripts") pod "3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" (UID: "3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.481545 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-kube-api-access-6prjn" (OuterVolumeSpecName: "kube-api-access-6prjn") pod "3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" (UID: "3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c"). InnerVolumeSpecName "kube-api-access-6prjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.481703 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" (UID: "3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.577264 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6prjn\" (UniqueName: \"kubernetes.io/projected/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-kube-api-access-6prjn\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.577301 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.577312 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.589083 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" (UID: "3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.622145 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-config-data" (OuterVolumeSpecName: "config-data") pod "3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" (UID: "3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.647353 4775 generic.go:334] "Generic (PLEG): container finished" podID="9e78aa9b-9107-461c-b94f-c0f166b489b4" containerID="10a348acd950eaad46c525129402d791195a91d90d80d01aa576064a8aeec468" exitCode=0 Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.647403 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" event={"ID":"9e78aa9b-9107-461c-b94f-c0f166b489b4","Type":"ContainerDied","Data":"10a348acd950eaad46c525129402d791195a91d90d80d01aa576064a8aeec468"} Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.647426 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" event={"ID":"9e78aa9b-9107-461c-b94f-c0f166b489b4","Type":"ContainerDied","Data":"1cde4bf949e3ada68147efe3390db54f06c28389d308e1852c50298736a76357"} Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.647438 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cde4bf949e3ada68147efe3390db54f06c28389d308e1852c50298736a76357" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.647516 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.654427 4775 generic.go:334] "Generic (PLEG): container finished" podID="97a109cd-fde0-409f-a0cc-e2e3c78acb47" containerID="11fc048e233fc00c87c2d04d492ad8e05971c6d7aac57380af9bd95f239d9a3d" exitCode=0 Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.654474 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322841-j7zjm" event={"ID":"97a109cd-fde0-409f-a0cc-e2e3c78acb47","Type":"ContainerDied","Data":"11fc048e233fc00c87c2d04d492ad8e05971c6d7aac57380af9bd95f239d9a3d"} Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.655635 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5698978bb-cdg88" event={"ID":"4f876ea1-9d84-4dd0-a2f6-2bc742bab66e","Type":"ContainerDied","Data":"b0ffa726f8ffaaafd74e169c0f0ff5960db52a4dd04a8271aec417e9b2fae302"} Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.655665 4775 scope.go:117] "RemoveContainer" containerID="169652bfa5bde2f7471ba1a8910ef01f1640b2f0a9daf2c6c26ac42892314dbf" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.655754 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5698978bb-cdg88" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.679215 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.679259 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.702083 4775 generic.go:334] "Generic (PLEG): container finished" podID="3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" containerID="26f726da7ed883ed5e95b1dbab9ff53a265b3c9bd474ea22b22b794a2a0aa4b8" exitCode=0 Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.702407 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c","Type":"ContainerDied","Data":"26f726da7ed883ed5e95b1dbab9ff53a265b3c9bd474ea22b22b794a2a0aa4b8"} Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.702434 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c","Type":"ContainerDied","Data":"b85be8a505af5b62a8498133fbe804c67a52ee5457e0cb7ba63349c7dfb1e10c"} Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.702492 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.715841 4775 scope.go:117] "RemoveContainer" containerID="250bae0449846fc4dda5adff85b5ed10a3b0e5cffbbb6251ffe6503a2a43110e" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.739132 4775 scope.go:117] "RemoveContainer" containerID="032a92c01d2c434edf83435f1eaf6ace4fd8cdf6d79e46caa4ff3e6b24d459ed" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.749644 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-5698978bb-cdg88"] Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.763082 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-5698978bb-cdg88"] Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.765046 4775 scope.go:117] "RemoveContainer" containerID="26f726da7ed883ed5e95b1dbab9ff53a265b3c9bd474ea22b22b794a2a0aa4b8" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.781190 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-dns-swift-storage-0\") pod \"9e78aa9b-9107-461c-b94f-c0f166b489b4\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.781846 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mchn\" (UniqueName: \"kubernetes.io/projected/9e78aa9b-9107-461c-b94f-c0f166b489b4-kube-api-access-4mchn\") pod \"9e78aa9b-9107-461c-b94f-c0f166b489b4\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.782005 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-config\") pod \"9e78aa9b-9107-461c-b94f-c0f166b489b4\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.782134 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-dns-svc\") pod \"9e78aa9b-9107-461c-b94f-c0f166b489b4\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.782309 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-ovsdbserver-nb\") pod \"9e78aa9b-9107-461c-b94f-c0f166b489b4\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.782430 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-ovsdbserver-sb\") pod \"9e78aa9b-9107-461c-b94f-c0f166b489b4\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.806491 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" path="/var/lib/kubelet/pods/4f876ea1-9d84-4dd0-a2f6-2bc742bab66e/volumes" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.811584 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e78aa9b-9107-461c-b94f-c0f166b489b4-kube-api-access-4mchn" (OuterVolumeSpecName: "kube-api-access-4mchn") pod "9e78aa9b-9107-461c-b94f-c0f166b489b4" (UID: "9e78aa9b-9107-461c-b94f-c0f166b489b4"). InnerVolumeSpecName "kube-api-access-4mchn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.813019 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.813050 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.833402 4775 scope.go:117] "RemoveContainer" containerID="032a92c01d2c434edf83435f1eaf6ace4fd8cdf6d79e46caa4ff3e6b24d459ed" Oct 02 02:01:03 crc kubenswrapper[4775]: E1002 02:01:03.834336 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"032a92c01d2c434edf83435f1eaf6ace4fd8cdf6d79e46caa4ff3e6b24d459ed\": container with ID starting with 032a92c01d2c434edf83435f1eaf6ace4fd8cdf6d79e46caa4ff3e6b24d459ed not found: ID does not exist" containerID="032a92c01d2c434edf83435f1eaf6ace4fd8cdf6d79e46caa4ff3e6b24d459ed" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.834387 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"032a92c01d2c434edf83435f1eaf6ace4fd8cdf6d79e46caa4ff3e6b24d459ed"} err="failed to get container status \"032a92c01d2c434edf83435f1eaf6ace4fd8cdf6d79e46caa4ff3e6b24d459ed\": rpc error: code = NotFound desc = could not find container \"032a92c01d2c434edf83435f1eaf6ace4fd8cdf6d79e46caa4ff3e6b24d459ed\": container with ID starting with 032a92c01d2c434edf83435f1eaf6ace4fd8cdf6d79e46caa4ff3e6b24d459ed not found: ID does not exist" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.834407 4775 scope.go:117] "RemoveContainer" containerID="26f726da7ed883ed5e95b1dbab9ff53a265b3c9bd474ea22b22b794a2a0aa4b8" Oct 02 02:01:03 crc kubenswrapper[4775]: E1002 02:01:03.834886 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26f726da7ed883ed5e95b1dbab9ff53a265b3c9bd474ea22b22b794a2a0aa4b8\": container with ID starting with 26f726da7ed883ed5e95b1dbab9ff53a265b3c9bd474ea22b22b794a2a0aa4b8 not found: ID does not exist" containerID="26f726da7ed883ed5e95b1dbab9ff53a265b3c9bd474ea22b22b794a2a0aa4b8" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.834906 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26f726da7ed883ed5e95b1dbab9ff53a265b3c9bd474ea22b22b794a2a0aa4b8"} err="failed to get container status \"26f726da7ed883ed5e95b1dbab9ff53a265b3c9bd474ea22b22b794a2a0aa4b8\": rpc error: code = NotFound desc = could not find container \"26f726da7ed883ed5e95b1dbab9ff53a265b3c9bd474ea22b22b794a2a0aa4b8\": container with ID starting with 26f726da7ed883ed5e95b1dbab9ff53a265b3c9bd474ea22b22b794a2a0aa4b8 not found: ID does not exist" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.842654 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9e78aa9b-9107-461c-b94f-c0f166b489b4" (UID: "9e78aa9b-9107-461c-b94f-c0f166b489b4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.844844 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:01:03 crc kubenswrapper[4775]: E1002 02:01:03.845381 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e78aa9b-9107-461c-b94f-c0f166b489b4" containerName="dnsmasq-dns" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.845399 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e78aa9b-9107-461c-b94f-c0f166b489b4" containerName="dnsmasq-dns" Oct 02 02:01:03 crc kubenswrapper[4775]: E1002 02:01:03.845416 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" containerName="barbican-api" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.845422 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" containerName="barbican-api" Oct 02 02:01:03 crc kubenswrapper[4775]: E1002 02:01:03.845436 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" containerName="probe" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.845442 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" containerName="probe" Oct 02 02:01:03 crc kubenswrapper[4775]: E1002 02:01:03.845450 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" containerName="cinder-scheduler" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.845456 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" containerName="cinder-scheduler" Oct 02 02:01:03 crc kubenswrapper[4775]: E1002 02:01:03.845484 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" containerName="barbican-api-log" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.845538 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" containerName="barbican-api-log" Oct 02 02:01:03 crc kubenswrapper[4775]: E1002 02:01:03.845552 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e78aa9b-9107-461c-b94f-c0f166b489b4" containerName="init" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.845558 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e78aa9b-9107-461c-b94f-c0f166b489b4" containerName="init" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.845719 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" containerName="barbican-api-log" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.845731 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" containerName="probe" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.845742 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f876ea1-9d84-4dd0-a2f6-2bc742bab66e" containerName="barbican-api" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.845755 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" containerName="cinder-scheduler" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.845767 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e78aa9b-9107-461c-b94f-c0f166b489b4" containerName="dnsmasq-dns" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.846627 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.848583 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.857089 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.873974 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9e78aa9b-9107-461c-b94f-c0f166b489b4" (UID: "9e78aa9b-9107-461c-b94f-c0f166b489b4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.876742 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9e78aa9b-9107-461c-b94f-c0f166b489b4" (UID: "9e78aa9b-9107-461c-b94f-c0f166b489b4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.887781 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-config" (OuterVolumeSpecName: "config") pod "9e78aa9b-9107-461c-b94f-c0f166b489b4" (UID: "9e78aa9b-9107-461c-b94f-c0f166b489b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.888286 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9e78aa9b-9107-461c-b94f-c0f166b489b4" (UID: "9e78aa9b-9107-461c-b94f-c0f166b489b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.888597 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-config\") pod \"9e78aa9b-9107-461c-b94f-c0f166b489b4\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.888659 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-dns-svc\") pod \"9e78aa9b-9107-461c-b94f-c0f166b489b4\" (UID: \"9e78aa9b-9107-461c-b94f-c0f166b489b4\") " Oct 02 02:01:03 crc kubenswrapper[4775]: W1002 02:01:03.888694 4775 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/9e78aa9b-9107-461c-b94f-c0f166b489b4/volumes/kubernetes.io~configmap/config Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.888703 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-config" (OuterVolumeSpecName: "config") pod "9e78aa9b-9107-461c-b94f-c0f166b489b4" (UID: "9e78aa9b-9107-461c-b94f-c0f166b489b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: W1002 02:01:03.888800 4775 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/9e78aa9b-9107-461c-b94f-c0f166b489b4/volumes/kubernetes.io~configmap/dns-svc Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.888811 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9e78aa9b-9107-461c-b94f-c0f166b489b4" (UID: "9e78aa9b-9107-461c-b94f-c0f166b489b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.889110 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.889128 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.889138 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mchn\" (UniqueName: \"kubernetes.io/projected/9e78aa9b-9107-461c-b94f-c0f166b489b4-kube-api-access-4mchn\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.889148 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.889156 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.889164 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9e78aa9b-9107-461c-b94f-c0f166b489b4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.990227 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-scripts\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.990288 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.990319 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.990357 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.990570 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-config-data\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:03 crc kubenswrapper[4775]: I1002 02:01:03.990659 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs57f\" (UniqueName: \"kubernetes.io/projected/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-kube-api-access-xs57f\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.092201 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-config-data\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.092251 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs57f\" (UniqueName: \"kubernetes.io/projected/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-kube-api-access-xs57f\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.092303 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-scripts\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.092329 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.092353 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.092388 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.093091 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.098465 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.098474 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.099667 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-config-data\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.101719 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-scripts\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.111676 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs57f\" (UniqueName: \"kubernetes.io/projected/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-kube-api-access-xs57f\") pod \"cinder-scheduler-0\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " pod="openstack/cinder-scheduler-0" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.161475 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.592154 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:01:04 crc kubenswrapper[4775]: W1002 02:01:04.604137 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24e08fdf_b02d_4d6b_b451_8aa967ff04f6.slice/crio-4c6de7d12a2ce34841501a709102fa7d8c0d30c95ba74db6eead2bb64802a87f WatchSource:0}: Error finding container 4c6de7d12a2ce34841501a709102fa7d8c0d30c95ba74db6eead2bb64802a87f: Status 404 returned error can't find the container with id 4c6de7d12a2ce34841501a709102fa7d8c0d30c95ba74db6eead2bb64802a87f Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.719987 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"24e08fdf-b02d-4d6b-b451-8aa967ff04f6","Type":"ContainerStarted","Data":"4c6de7d12a2ce34841501a709102fa7d8c0d30c95ba74db6eead2bb64802a87f"} Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.720096 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-bczkw" Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.758994 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-bczkw"] Oct 02 02:01:04 crc kubenswrapper[4775]: I1002 02:01:04.767320 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-bczkw"] Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.035804 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.124273 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-combined-ca-bundle\") pod \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.124432 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-fernet-keys\") pod \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.124462 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7jxd\" (UniqueName: \"kubernetes.io/projected/97a109cd-fde0-409f-a0cc-e2e3c78acb47-kube-api-access-m7jxd\") pod \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.124499 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-config-data\") pod \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\" (UID: \"97a109cd-fde0-409f-a0cc-e2e3c78acb47\") " Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.129166 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "97a109cd-fde0-409f-a0cc-e2e3c78acb47" (UID: "97a109cd-fde0-409f-a0cc-e2e3c78acb47"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.133067 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97a109cd-fde0-409f-a0cc-e2e3c78acb47-kube-api-access-m7jxd" (OuterVolumeSpecName: "kube-api-access-m7jxd") pod "97a109cd-fde0-409f-a0cc-e2e3c78acb47" (UID: "97a109cd-fde0-409f-a0cc-e2e3c78acb47"). InnerVolumeSpecName "kube-api-access-m7jxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.161775 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97a109cd-fde0-409f-a0cc-e2e3c78acb47" (UID: "97a109cd-fde0-409f-a0cc-e2e3c78acb47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.194007 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-config-data" (OuterVolumeSpecName: "config-data") pod "97a109cd-fde0-409f-a0cc-e2e3c78acb47" (UID: "97a109cd-fde0-409f-a0cc-e2e3c78acb47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.226249 4775 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.226280 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7jxd\" (UniqueName: \"kubernetes.io/projected/97a109cd-fde0-409f-a0cc-e2e3c78acb47-kube-api-access-m7jxd\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.226292 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.226300 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97a109cd-fde0-409f-a0cc-e2e3c78acb47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.439162 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 02:01:05 crc kubenswrapper[4775]: E1002 02:01:05.439666 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97a109cd-fde0-409f-a0cc-e2e3c78acb47" containerName="keystone-cron" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.439692 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="97a109cd-fde0-409f-a0cc-e2e3c78acb47" containerName="keystone-cron" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.439985 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="97a109cd-fde0-409f-a0cc-e2e3c78acb47" containerName="keystone-cron" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.440732 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.444143 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-ddgf5" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.444471 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.444644 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.479563 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.531345 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764d109a-a65d-4fa6-ab71-a8ffb49e442e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " pod="openstack/openstackclient" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.531399 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/764d109a-a65d-4fa6-ab71-a8ffb49e442e-openstack-config-secret\") pod \"openstackclient\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " pod="openstack/openstackclient" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.531520 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/764d109a-a65d-4fa6-ab71-a8ffb49e442e-openstack-config\") pod \"openstackclient\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " pod="openstack/openstackclient" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.531558 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcvnx\" (UniqueName: \"kubernetes.io/projected/764d109a-a65d-4fa6-ab71-a8ffb49e442e-kube-api-access-vcvnx\") pod \"openstackclient\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " pod="openstack/openstackclient" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.633073 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/764d109a-a65d-4fa6-ab71-a8ffb49e442e-openstack-config\") pod \"openstackclient\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " pod="openstack/openstackclient" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.633338 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcvnx\" (UniqueName: \"kubernetes.io/projected/764d109a-a65d-4fa6-ab71-a8ffb49e442e-kube-api-access-vcvnx\") pod \"openstackclient\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " pod="openstack/openstackclient" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.633477 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764d109a-a65d-4fa6-ab71-a8ffb49e442e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " pod="openstack/openstackclient" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.633572 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/764d109a-a65d-4fa6-ab71-a8ffb49e442e-openstack-config-secret\") pod \"openstackclient\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " pod="openstack/openstackclient" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.634256 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/764d109a-a65d-4fa6-ab71-a8ffb49e442e-openstack-config\") pod \"openstackclient\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " pod="openstack/openstackclient" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.649464 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/764d109a-a65d-4fa6-ab71-a8ffb49e442e-openstack-config-secret\") pod \"openstackclient\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " pod="openstack/openstackclient" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.649819 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764d109a-a65d-4fa6-ab71-a8ffb49e442e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " pod="openstack/openstackclient" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.654744 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcvnx\" (UniqueName: \"kubernetes.io/projected/764d109a-a65d-4fa6-ab71-a8ffb49e442e-kube-api-access-vcvnx\") pod \"openstackclient\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " pod="openstack/openstackclient" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.708834 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.757754 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322841-j7zjm" event={"ID":"97a109cd-fde0-409f-a0cc-e2e3c78acb47","Type":"ContainerDied","Data":"8b6a0c1861a2eebc67cee3161f50961bcdc51b8c7448716af466c8baf3d44627"} Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.757791 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b6a0c1861a2eebc67cee3161f50961bcdc51b8c7448716af466c8baf3d44627" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.757779 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322841-j7zjm" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.761448 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"24e08fdf-b02d-4d6b-b451-8aa967ff04f6","Type":"ContainerStarted","Data":"86650058d32beec95536f3008ef303ebaebff3385b901bf98a315478321c1bce"} Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.783283 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c" path="/var/lib/kubelet/pods/3b8d4b2f-17d1-4c47-bcc7-97f3fd22220c/volumes" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.785163 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e78aa9b-9107-461c-b94f-c0f166b489b4" path="/var/lib/kubelet/pods/9e78aa9b-9107-461c-b94f-c0f166b489b4/volumes" Oct 02 02:01:05 crc kubenswrapper[4775]: I1002 02:01:05.785244 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 02:01:06 crc kubenswrapper[4775]: I1002 02:01:06.321692 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 02:01:06 crc kubenswrapper[4775]: W1002 02:01:06.323659 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod764d109a_a65d_4fa6_ab71_a8ffb49e442e.slice/crio-9935899c8ea57c04dd3323283b4595948225cf06d8dc3c4b280ffb5b2632bae4 WatchSource:0}: Error finding container 9935899c8ea57c04dd3323283b4595948225cf06d8dc3c4b280ffb5b2632bae4: Status 404 returned error can't find the container with id 9935899c8ea57c04dd3323283b4595948225cf06d8dc3c4b280ffb5b2632bae4 Oct 02 02:01:06 crc kubenswrapper[4775]: I1002 02:01:06.769062 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"764d109a-a65d-4fa6-ab71-a8ffb49e442e","Type":"ContainerStarted","Data":"9935899c8ea57c04dd3323283b4595948225cf06d8dc3c4b280ffb5b2632bae4"} Oct 02 02:01:06 crc kubenswrapper[4775]: I1002 02:01:06.771559 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"24e08fdf-b02d-4d6b-b451-8aa967ff04f6","Type":"ContainerStarted","Data":"ecfa4d3c77b7cfaba81f2e5e014ad3a08c3704a721e8b2bcc3173012b42a0aff"} Oct 02 02:01:06 crc kubenswrapper[4775]: I1002 02:01:06.792739 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.792723092 podStartE2EDuration="3.792723092s" podCreationTimestamp="2025-10-02 02:01:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:01:06.787132698 +0000 UTC m=+1203.953876738" watchObservedRunningTime="2025-10-02 02:01:06.792723092 +0000 UTC m=+1203.959467132" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.162435 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.165983 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-585fd6468c-8gx5n"] Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.167988 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.171631 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.171790 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.173636 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.204129 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-585fd6468c-8gx5n"] Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.259179 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.259445 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="ceilometer-central-agent" containerID="cri-o://841211e5dddbb7078cf26acd9820c0e489445c62fa82375fb5b93d954c33ee0b" gracePeriod=30 Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.259556 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="proxy-httpd" containerID="cri-o://f4bc65323b150aac0a5fa8b7190be3cebed294ecdde611749ce8736c77e1c19c" gracePeriod=30 Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.259597 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="sg-core" containerID="cri-o://5fae55178786be08f69283be73c0f41e9455f7929793ecfbb705a70df3e98fc4" gracePeriod=30 Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.259627 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="ceilometer-notification-agent" containerID="cri-o://d62fda876ad3ecd35bd593f1776e6700e2a6b1b84104fa540031723876198146" gracePeriod=30 Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.308243 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-public-tls-certs\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.308295 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-internal-tls-certs\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.308319 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/446bd539-6bf4-42b1-ac19-40c8c80b2d45-run-httpd\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.308380 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-combined-ca-bundle\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.308406 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-config-data\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.308445 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/446bd539-6bf4-42b1-ac19-40c8c80b2d45-log-httpd\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.308482 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsqxs\" (UniqueName: \"kubernetes.io/projected/446bd539-6bf4-42b1-ac19-40c8c80b2d45-kube-api-access-dsqxs\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.308517 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/446bd539-6bf4-42b1-ac19-40c8c80b2d45-etc-swift\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.414417 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-public-tls-certs\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.414531 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-internal-tls-certs\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.414566 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/446bd539-6bf4-42b1-ac19-40c8c80b2d45-run-httpd\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.414682 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-combined-ca-bundle\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.414742 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-config-data\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.414813 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/446bd539-6bf4-42b1-ac19-40c8c80b2d45-log-httpd\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.414988 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dsqxs\" (UniqueName: \"kubernetes.io/projected/446bd539-6bf4-42b1-ac19-40c8c80b2d45-kube-api-access-dsqxs\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.415045 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/446bd539-6bf4-42b1-ac19-40c8c80b2d45-etc-swift\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.415787 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/446bd539-6bf4-42b1-ac19-40c8c80b2d45-log-httpd\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.416911 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/446bd539-6bf4-42b1-ac19-40c8c80b2d45-run-httpd\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.424210 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-combined-ca-bundle\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.431606 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-public-tls-certs\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.432927 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/446bd539-6bf4-42b1-ac19-40c8c80b2d45-etc-swift\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.436393 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsqxs\" (UniqueName: \"kubernetes.io/projected/446bd539-6bf4-42b1-ac19-40c8c80b2d45-kube-api-access-dsqxs\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.437981 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-config-data\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.438169 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-internal-tls-certs\") pod \"swift-proxy-585fd6468c-8gx5n\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.500588 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.838355 4775 generic.go:334] "Generic (PLEG): container finished" podID="18a90265-099f-4da5-9d05-32956ab2deb4" containerID="f4bc65323b150aac0a5fa8b7190be3cebed294ecdde611749ce8736c77e1c19c" exitCode=0 Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.838559 4775 generic.go:334] "Generic (PLEG): container finished" podID="18a90265-099f-4da5-9d05-32956ab2deb4" containerID="5fae55178786be08f69283be73c0f41e9455f7929793ecfbb705a70df3e98fc4" exitCode=2 Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.838569 4775 generic.go:334] "Generic (PLEG): container finished" podID="18a90265-099f-4da5-9d05-32956ab2deb4" containerID="841211e5dddbb7078cf26acd9820c0e489445c62fa82375fb5b93d954c33ee0b" exitCode=0 Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.838588 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18a90265-099f-4da5-9d05-32956ab2deb4","Type":"ContainerDied","Data":"f4bc65323b150aac0a5fa8b7190be3cebed294ecdde611749ce8736c77e1c19c"} Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.838611 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18a90265-099f-4da5-9d05-32956ab2deb4","Type":"ContainerDied","Data":"5fae55178786be08f69283be73c0f41e9455f7929793ecfbb705a70df3e98fc4"} Oct 02 02:01:09 crc kubenswrapper[4775]: I1002 02:01:09.838620 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18a90265-099f-4da5-9d05-32956ab2deb4","Type":"ContainerDied","Data":"841211e5dddbb7078cf26acd9820c0e489445c62fa82375fb5b93d954c33ee0b"} Oct 02 02:01:10 crc kubenswrapper[4775]: I1002 02:01:10.113369 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-585fd6468c-8gx5n"] Oct 02 02:01:10 crc kubenswrapper[4775]: W1002 02:01:10.126484 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod446bd539_6bf4_42b1_ac19_40c8c80b2d45.slice/crio-66aa2c1759ddb8693183396b39f057b3cb58c4c2ae61f8f748b5bf95acd9d362 WatchSource:0}: Error finding container 66aa2c1759ddb8693183396b39f057b3cb58c4c2ae61f8f748b5bf95acd9d362: Status 404 returned error can't find the container with id 66aa2c1759ddb8693183396b39f057b3cb58c4c2ae61f8f748b5bf95acd9d362 Oct 02 02:01:10 crc kubenswrapper[4775]: I1002 02:01:10.858500 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-585fd6468c-8gx5n" event={"ID":"446bd539-6bf4-42b1-ac19-40c8c80b2d45","Type":"ContainerStarted","Data":"9d2ff242a1e29ad8407f6c796b021ec08d4a487ead665a7032ac8ddbf2177bbc"} Oct 02 02:01:10 crc kubenswrapper[4775]: I1002 02:01:10.858747 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-585fd6468c-8gx5n" event={"ID":"446bd539-6bf4-42b1-ac19-40c8c80b2d45","Type":"ContainerStarted","Data":"66aa2c1759ddb8693183396b39f057b3cb58c4c2ae61f8f748b5bf95acd9d362"} Oct 02 02:01:11 crc kubenswrapper[4775]: I1002 02:01:11.878778 4775 generic.go:334] "Generic (PLEG): container finished" podID="18a90265-099f-4da5-9d05-32956ab2deb4" containerID="d62fda876ad3ecd35bd593f1776e6700e2a6b1b84104fa540031723876198146" exitCode=0 Oct 02 02:01:11 crc kubenswrapper[4775]: I1002 02:01:11.878837 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18a90265-099f-4da5-9d05-32956ab2deb4","Type":"ContainerDied","Data":"d62fda876ad3ecd35bd593f1776e6700e2a6b1b84104fa540031723876198146"} Oct 02 02:01:14 crc kubenswrapper[4775]: I1002 02:01:14.344858 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.108617 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.262352 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-scripts\") pod \"18a90265-099f-4da5-9d05-32956ab2deb4\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.262430 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18a90265-099f-4da5-9d05-32956ab2deb4-log-httpd\") pod \"18a90265-099f-4da5-9d05-32956ab2deb4\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.262452 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18a90265-099f-4da5-9d05-32956ab2deb4-run-httpd\") pod \"18a90265-099f-4da5-9d05-32956ab2deb4\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.262486 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-combined-ca-bundle\") pod \"18a90265-099f-4da5-9d05-32956ab2deb4\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.262534 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-config-data\") pod \"18a90265-099f-4da5-9d05-32956ab2deb4\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.262596 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-sg-core-conf-yaml\") pod \"18a90265-099f-4da5-9d05-32956ab2deb4\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.262696 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcbf5\" (UniqueName: \"kubernetes.io/projected/18a90265-099f-4da5-9d05-32956ab2deb4-kube-api-access-wcbf5\") pod \"18a90265-099f-4da5-9d05-32956ab2deb4\" (UID: \"18a90265-099f-4da5-9d05-32956ab2deb4\") " Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.263041 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18a90265-099f-4da5-9d05-32956ab2deb4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "18a90265-099f-4da5-9d05-32956ab2deb4" (UID: "18a90265-099f-4da5-9d05-32956ab2deb4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.263159 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18a90265-099f-4da5-9d05-32956ab2deb4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "18a90265-099f-4da5-9d05-32956ab2deb4" (UID: "18a90265-099f-4da5-9d05-32956ab2deb4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.271170 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18a90265-099f-4da5-9d05-32956ab2deb4-kube-api-access-wcbf5" (OuterVolumeSpecName: "kube-api-access-wcbf5") pod "18a90265-099f-4da5-9d05-32956ab2deb4" (UID: "18a90265-099f-4da5-9d05-32956ab2deb4"). InnerVolumeSpecName "kube-api-access-wcbf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.275923 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-scripts" (OuterVolumeSpecName: "scripts") pod "18a90265-099f-4da5-9d05-32956ab2deb4" (UID: "18a90265-099f-4da5-9d05-32956ab2deb4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.293006 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "18a90265-099f-4da5-9d05-32956ab2deb4" (UID: "18a90265-099f-4da5-9d05-32956ab2deb4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.333578 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18a90265-099f-4da5-9d05-32956ab2deb4" (UID: "18a90265-099f-4da5-9d05-32956ab2deb4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.358137 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-config-data" (OuterVolumeSpecName: "config-data") pod "18a90265-099f-4da5-9d05-32956ab2deb4" (UID: "18a90265-099f-4da5-9d05-32956ab2deb4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.365332 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcbf5\" (UniqueName: \"kubernetes.io/projected/18a90265-099f-4da5-9d05-32956ab2deb4-kube-api-access-wcbf5\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.365583 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.365673 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18a90265-099f-4da5-9d05-32956ab2deb4-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.365781 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/18a90265-099f-4da5-9d05-32956ab2deb4-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.365895 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.365995 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.366074 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/18a90265-099f-4da5-9d05-32956ab2deb4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.949823 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-585fd6468c-8gx5n" event={"ID":"446bd539-6bf4-42b1-ac19-40c8c80b2d45","Type":"ContainerStarted","Data":"9a86b1d47a3009097a1c9b5e0d10b4d2e4d64194427ec3554b609bf3c481d41c"} Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.950008 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.950054 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.952860 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"18a90265-099f-4da5-9d05-32956ab2deb4","Type":"ContainerDied","Data":"eb40623c790af2cb083402b5d65cfe5b17f1e9e6d8e17b2b8d1f84159b45c20e"} Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.952891 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.952915 4775 scope.go:117] "RemoveContainer" containerID="f4bc65323b150aac0a5fa8b7190be3cebed294ecdde611749ce8736c77e1c19c" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.954185 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"764d109a-a65d-4fa6-ab71-a8ffb49e442e","Type":"ContainerStarted","Data":"ce4365b06788ec5e2d1ab7f03c97b339cfee51b36bf1b4f570178d210ed939dc"} Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.963667 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-585fd6468c-8gx5n" podUID="446bd539-6bf4-42b1-ac19-40c8c80b2d45" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.981530 4775 scope.go:117] "RemoveContainer" containerID="5fae55178786be08f69283be73c0f41e9455f7929793ecfbb705a70df3e98fc4" Oct 02 02:01:16 crc kubenswrapper[4775]: I1002 02:01:16.992365 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-585fd6468c-8gx5n" podStartSLOduration=7.99233607 podStartE2EDuration="7.99233607s" podCreationTimestamp="2025-10-02 02:01:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:01:16.986598453 +0000 UTC m=+1214.153342493" watchObservedRunningTime="2025-10-02 02:01:16.99233607 +0000 UTC m=+1214.159080140" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.015294 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.020146 4775 scope.go:117] "RemoveContainer" containerID="d62fda876ad3ecd35bd593f1776e6700e2a6b1b84104fa540031723876198146" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.032940 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.045705 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.500815208 podStartE2EDuration="12.045682413s" podCreationTimestamp="2025-10-02 02:01:05 +0000 UTC" firstStartedPulling="2025-10-02 02:01:06.325727998 +0000 UTC m=+1203.492472038" lastFinishedPulling="2025-10-02 02:01:15.870595203 +0000 UTC m=+1213.037339243" observedRunningTime="2025-10-02 02:01:17.021935642 +0000 UTC m=+1214.188679682" watchObservedRunningTime="2025-10-02 02:01:17.045682413 +0000 UTC m=+1214.212426463" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.046580 4775 scope.go:117] "RemoveContainer" containerID="841211e5dddbb7078cf26acd9820c0e489445c62fa82375fb5b93d954c33ee0b" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.062711 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:17 crc kubenswrapper[4775]: E1002 02:01:17.063157 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="proxy-httpd" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.063170 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="proxy-httpd" Oct 02 02:01:17 crc kubenswrapper[4775]: E1002 02:01:17.063195 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="ceilometer-central-agent" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.063201 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="ceilometer-central-agent" Oct 02 02:01:17 crc kubenswrapper[4775]: E1002 02:01:17.063217 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="sg-core" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.063223 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="sg-core" Oct 02 02:01:17 crc kubenswrapper[4775]: E1002 02:01:17.063235 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="ceilometer-notification-agent" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.063242 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="ceilometer-notification-agent" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.063423 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="ceilometer-notification-agent" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.063433 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="ceilometer-central-agent" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.063450 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="proxy-httpd" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.063464 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" containerName="sg-core" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.065256 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.070984 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.071327 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.078488 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.183010 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.183061 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78a04f54-4d65-428e-bac7-099665ceaf38-run-httpd\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.183140 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78a04f54-4d65-428e-bac7-099665ceaf38-log-httpd\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.183173 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-scripts\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.183199 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.183268 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-config-data\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.183286 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grfmt\" (UniqueName: \"kubernetes.io/projected/78a04f54-4d65-428e-bac7-099665ceaf38-kube-api-access-grfmt\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.285322 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78a04f54-4d65-428e-bac7-099665ceaf38-log-httpd\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.285380 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-scripts\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.285405 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.285466 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grfmt\" (UniqueName: \"kubernetes.io/projected/78a04f54-4d65-428e-bac7-099665ceaf38-kube-api-access-grfmt\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.285483 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-config-data\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.285530 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.285544 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78a04f54-4d65-428e-bac7-099665ceaf38-run-httpd\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.285920 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78a04f54-4d65-428e-bac7-099665ceaf38-log-httpd\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.286023 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78a04f54-4d65-428e-bac7-099665ceaf38-run-httpd\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.290569 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-scripts\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.293653 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.302436 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-config-data\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.303028 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grfmt\" (UniqueName: \"kubernetes.io/projected/78a04f54-4d65-428e-bac7-099665ceaf38-kube-api-access-grfmt\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.316431 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.397231 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.781239 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18a90265-099f-4da5-9d05-32956ab2deb4" path="/var/lib/kubelet/pods/18a90265-099f-4da5-9d05-32956ab2deb4/volumes" Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.855860 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:17 crc kubenswrapper[4775]: W1002 02:01:17.860212 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78a04f54_4d65_428e_bac7_099665ceaf38.slice/crio-e0adffd259ec9c5675db160fcdfe6757e5b06e096ca818521b0faedcaf4aa368 WatchSource:0}: Error finding container e0adffd259ec9c5675db160fcdfe6757e5b06e096ca818521b0faedcaf4aa368: Status 404 returned error can't find the container with id e0adffd259ec9c5675db160fcdfe6757e5b06e096ca818521b0faedcaf4aa368 Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.975092 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78a04f54-4d65-428e-bac7-099665ceaf38","Type":"ContainerStarted","Data":"e0adffd259ec9c5675db160fcdfe6757e5b06e096ca818521b0faedcaf4aa368"} Oct 02 02:01:17 crc kubenswrapper[4775]: I1002 02:01:17.986055 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:19 crc kubenswrapper[4775]: I1002 02:01:19.997556 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78a04f54-4d65-428e-bac7-099665ceaf38","Type":"ContainerStarted","Data":"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044"} Oct 02 02:01:20 crc kubenswrapper[4775]: I1002 02:01:20.518849 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:21 crc kubenswrapper[4775]: I1002 02:01:21.007998 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78a04f54-4d65-428e-bac7-099665ceaf38","Type":"ContainerStarted","Data":"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf"} Oct 02 02:01:21 crc kubenswrapper[4775]: I1002 02:01:21.008329 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78a04f54-4d65-428e-bac7-099665ceaf38","Type":"ContainerStarted","Data":"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52"} Oct 02 02:01:22 crc kubenswrapper[4775]: I1002 02:01:22.770658 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-b8mb2"] Oct 02 02:01:22 crc kubenswrapper[4775]: I1002 02:01:22.771985 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-b8mb2" Oct 02 02:01:22 crc kubenswrapper[4775]: I1002 02:01:22.783454 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-b8mb2"] Oct 02 02:01:22 crc kubenswrapper[4775]: I1002 02:01:22.858807 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-rjs9s"] Oct 02 02:01:22 crc kubenswrapper[4775]: I1002 02:01:22.859810 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rjs9s" Oct 02 02:01:22 crc kubenswrapper[4775]: I1002 02:01:22.884244 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7gj7\" (UniqueName: \"kubernetes.io/projected/b01ed253-a630-4b6c-9282-929b34b69b6d-kube-api-access-l7gj7\") pod \"nova-api-db-create-b8mb2\" (UID: \"b01ed253-a630-4b6c-9282-929b34b69b6d\") " pod="openstack/nova-api-db-create-b8mb2" Oct 02 02:01:22 crc kubenswrapper[4775]: I1002 02:01:22.896098 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-rjs9s"] Oct 02 02:01:22 crc kubenswrapper[4775]: I1002 02:01:22.969147 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-ddr42"] Oct 02 02:01:22 crc kubenswrapper[4775]: I1002 02:01:22.970438 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ddr42" Oct 02 02:01:22 crc kubenswrapper[4775]: I1002 02:01:22.986226 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7gj7\" (UniqueName: \"kubernetes.io/projected/b01ed253-a630-4b6c-9282-929b34b69b6d-kube-api-access-l7gj7\") pod \"nova-api-db-create-b8mb2\" (UID: \"b01ed253-a630-4b6c-9282-929b34b69b6d\") " pod="openstack/nova-api-db-create-b8mb2" Oct 02 02:01:22 crc kubenswrapper[4775]: I1002 02:01:22.986394 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22hh2\" (UniqueName: \"kubernetes.io/projected/8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9-kube-api-access-22hh2\") pod \"nova-cell0-db-create-rjs9s\" (UID: \"8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9\") " pod="openstack/nova-cell0-db-create-rjs9s" Oct 02 02:01:22 crc kubenswrapper[4775]: I1002 02:01:22.991399 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ddr42"] Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.006864 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7gj7\" (UniqueName: \"kubernetes.io/projected/b01ed253-a630-4b6c-9282-929b34b69b6d-kube-api-access-l7gj7\") pod \"nova-api-db-create-b8mb2\" (UID: \"b01ed253-a630-4b6c-9282-929b34b69b6d\") " pod="openstack/nova-api-db-create-b8mb2" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.007324 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.039155 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78a04f54-4d65-428e-bac7-099665ceaf38","Type":"ContainerStarted","Data":"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201"} Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.039309 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="ceilometer-central-agent" containerID="cri-o://06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044" gracePeriod=30 Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.040081 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="sg-core" containerID="cri-o://312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf" gracePeriod=30 Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.040189 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="proxy-httpd" containerID="cri-o://448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201" gracePeriod=30 Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.040242 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="ceilometer-notification-agent" containerID="cri-o://2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52" gracePeriod=30 Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.040273 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.087773 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2frlb\" (UniqueName: \"kubernetes.io/projected/c2d22306-6f52-462a-a980-210c208c595e-kube-api-access-2frlb\") pod \"nova-cell1-db-create-ddr42\" (UID: \"c2d22306-6f52-462a-a980-210c208c595e\") " pod="openstack/nova-cell1-db-create-ddr42" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.087917 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-b8mb2" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.089043 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22hh2\" (UniqueName: \"kubernetes.io/projected/8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9-kube-api-access-22hh2\") pod \"nova-cell0-db-create-rjs9s\" (UID: \"8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9\") " pod="openstack/nova-cell0-db-create-rjs9s" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.112826 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22hh2\" (UniqueName: \"kubernetes.io/projected/8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9-kube-api-access-22hh2\") pod \"nova-cell0-db-create-rjs9s\" (UID: \"8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9\") " pod="openstack/nova-cell0-db-create-rjs9s" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.174451 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rjs9s" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.193838 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2frlb\" (UniqueName: \"kubernetes.io/projected/c2d22306-6f52-462a-a980-210c208c595e-kube-api-access-2frlb\") pod \"nova-cell1-db-create-ddr42\" (UID: \"c2d22306-6f52-462a-a980-210c208c595e\") " pod="openstack/nova-cell1-db-create-ddr42" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.225910 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2frlb\" (UniqueName: \"kubernetes.io/projected/c2d22306-6f52-462a-a980-210c208c595e-kube-api-access-2frlb\") pod \"nova-cell1-db-create-ddr42\" (UID: \"c2d22306-6f52-462a-a980-210c208c595e\") " pod="openstack/nova-cell1-db-create-ddr42" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.291304 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ddr42" Oct 02 02:01:23 crc kubenswrapper[4775]: E1002 02:01:23.419478 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78a04f54_4d65_428e_bac7_099665ceaf38.slice/crio-conmon-448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78a04f54_4d65_428e_bac7_099665ceaf38.slice/crio-448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201.scope\": RecentStats: unable to find data in memory cache]" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.576595 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.478448206 podStartE2EDuration="6.576578932s" podCreationTimestamp="2025-10-02 02:01:17 +0000 UTC" firstStartedPulling="2025-10-02 02:01:17.86265342 +0000 UTC m=+1215.029397480" lastFinishedPulling="2025-10-02 02:01:21.960784166 +0000 UTC m=+1219.127528206" observedRunningTime="2025-10-02 02:01:23.071351095 +0000 UTC m=+1220.238095135" watchObservedRunningTime="2025-10-02 02:01:23.576578932 +0000 UTC m=+1220.743322972" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.584493 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-b8mb2"] Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.734738 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-rjs9s"] Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.824873 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.831829 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-ddr42"] Oct 02 02:01:23 crc kubenswrapper[4775]: W1002 02:01:23.887970 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2d22306_6f52_462a_a980_210c208c595e.slice/crio-faf1666928cea68fd0e07129eb445fa421edc7ca960485b5baf5da00acff475a WatchSource:0}: Error finding container faf1666928cea68fd0e07129eb445fa421edc7ca960485b5baf5da00acff475a: Status 404 returned error can't find the container with id faf1666928cea68fd0e07129eb445fa421edc7ca960485b5baf5da00acff475a Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.919052 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-config-data\") pod \"78a04f54-4d65-428e-bac7-099665ceaf38\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.919199 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grfmt\" (UniqueName: \"kubernetes.io/projected/78a04f54-4d65-428e-bac7-099665ceaf38-kube-api-access-grfmt\") pod \"78a04f54-4d65-428e-bac7-099665ceaf38\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.919260 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78a04f54-4d65-428e-bac7-099665ceaf38-run-httpd\") pod \"78a04f54-4d65-428e-bac7-099665ceaf38\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.919283 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-scripts\") pod \"78a04f54-4d65-428e-bac7-099665ceaf38\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.919351 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-combined-ca-bundle\") pod \"78a04f54-4d65-428e-bac7-099665ceaf38\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.919404 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-sg-core-conf-yaml\") pod \"78a04f54-4d65-428e-bac7-099665ceaf38\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.919419 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78a04f54-4d65-428e-bac7-099665ceaf38-log-httpd\") pod \"78a04f54-4d65-428e-bac7-099665ceaf38\" (UID: \"78a04f54-4d65-428e-bac7-099665ceaf38\") " Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.921707 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78a04f54-4d65-428e-bac7-099665ceaf38-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "78a04f54-4d65-428e-bac7-099665ceaf38" (UID: "78a04f54-4d65-428e-bac7-099665ceaf38"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.922385 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78a04f54-4d65-428e-bac7-099665ceaf38-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "78a04f54-4d65-428e-bac7-099665ceaf38" (UID: "78a04f54-4d65-428e-bac7-099665ceaf38"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.931665 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78a04f54-4d65-428e-bac7-099665ceaf38-kube-api-access-grfmt" (OuterVolumeSpecName: "kube-api-access-grfmt") pod "78a04f54-4d65-428e-bac7-099665ceaf38" (UID: "78a04f54-4d65-428e-bac7-099665ceaf38"). InnerVolumeSpecName "kube-api-access-grfmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.933690 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-scripts" (OuterVolumeSpecName: "scripts") pod "78a04f54-4d65-428e-bac7-099665ceaf38" (UID: "78a04f54-4d65-428e-bac7-099665ceaf38"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:23 crc kubenswrapper[4775]: I1002 02:01:23.957620 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "78a04f54-4d65-428e-bac7-099665ceaf38" (UID: "78a04f54-4d65-428e-bac7-099665ceaf38"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.002698 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78a04f54-4d65-428e-bac7-099665ceaf38" (UID: "78a04f54-4d65-428e-bac7-099665ceaf38"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.021798 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grfmt\" (UniqueName: \"kubernetes.io/projected/78a04f54-4d65-428e-bac7-099665ceaf38-kube-api-access-grfmt\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.021830 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78a04f54-4d65-428e-bac7-099665ceaf38-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.021839 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.021850 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.021863 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.021874 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/78a04f54-4d65-428e-bac7-099665ceaf38-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.043867 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-config-data" (OuterVolumeSpecName: "config-data") pod "78a04f54-4d65-428e-bac7-099665ceaf38" (UID: "78a04f54-4d65-428e-bac7-099665ceaf38"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.047529 4775 generic.go:334] "Generic (PLEG): container finished" podID="b01ed253-a630-4b6c-9282-929b34b69b6d" containerID="54e4763076c1e59bd43a8d1b7a00c8dc77ba9c288b69de634c971e266ee9e637" exitCode=0 Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.047912 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-b8mb2" event={"ID":"b01ed253-a630-4b6c-9282-929b34b69b6d","Type":"ContainerDied","Data":"54e4763076c1e59bd43a8d1b7a00c8dc77ba9c288b69de634c971e266ee9e637"} Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.047939 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-b8mb2" event={"ID":"b01ed253-a630-4b6c-9282-929b34b69b6d","Type":"ContainerStarted","Data":"910ec8ef8d00e72be83c879fffa937feb04eb691344a5739516a90e025cfdb28"} Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.049407 4775 generic.go:334] "Generic (PLEG): container finished" podID="8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9" containerID="e93c0d6ee8dbff552b995e91001193f38793e530b0cc3fab57aace81d77b326e" exitCode=0 Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.049446 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rjs9s" event={"ID":"8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9","Type":"ContainerDied","Data":"e93c0d6ee8dbff552b995e91001193f38793e530b0cc3fab57aace81d77b326e"} Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.049460 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rjs9s" event={"ID":"8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9","Type":"ContainerStarted","Data":"e2cd6dba0d5cfa556e0880b9773e4da47c7f578495d208b97fca9a70948a7cf4"} Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.056829 4775 generic.go:334] "Generic (PLEG): container finished" podID="78a04f54-4d65-428e-bac7-099665ceaf38" containerID="448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201" exitCode=0 Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.056854 4775 generic.go:334] "Generic (PLEG): container finished" podID="78a04f54-4d65-428e-bac7-099665ceaf38" containerID="312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf" exitCode=2 Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.056862 4775 generic.go:334] "Generic (PLEG): container finished" podID="78a04f54-4d65-428e-bac7-099665ceaf38" containerID="2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52" exitCode=0 Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.056869 4775 generic.go:334] "Generic (PLEG): container finished" podID="78a04f54-4d65-428e-bac7-099665ceaf38" containerID="06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044" exitCode=0 Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.056904 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78a04f54-4d65-428e-bac7-099665ceaf38","Type":"ContainerDied","Data":"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201"} Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.056926 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78a04f54-4d65-428e-bac7-099665ceaf38","Type":"ContainerDied","Data":"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf"} Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.056938 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78a04f54-4d65-428e-bac7-099665ceaf38","Type":"ContainerDied","Data":"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52"} Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.056966 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78a04f54-4d65-428e-bac7-099665ceaf38","Type":"ContainerDied","Data":"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044"} Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.056976 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"78a04f54-4d65-428e-bac7-099665ceaf38","Type":"ContainerDied","Data":"e0adffd259ec9c5675db160fcdfe6757e5b06e096ca818521b0faedcaf4aa368"} Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.056990 4775 scope.go:117] "RemoveContainer" containerID="448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.057151 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.061863 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ddr42" event={"ID":"c2d22306-6f52-462a-a980-210c208c595e","Type":"ContainerStarted","Data":"faf1666928cea68fd0e07129eb445fa421edc7ca960485b5baf5da00acff475a"} Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.124067 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78a04f54-4d65-428e-bac7-099665ceaf38-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.142013 4775 scope.go:117] "RemoveContainer" containerID="312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.143670 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.151052 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.160419 4775 scope.go:117] "RemoveContainer" containerID="2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.170579 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:24 crc kubenswrapper[4775]: E1002 02:01:24.170984 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="ceilometer-central-agent" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.171000 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="ceilometer-central-agent" Oct 02 02:01:24 crc kubenswrapper[4775]: E1002 02:01:24.171012 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="sg-core" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.171018 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="sg-core" Oct 02 02:01:24 crc kubenswrapper[4775]: E1002 02:01:24.171027 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="proxy-httpd" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.171033 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="proxy-httpd" Oct 02 02:01:24 crc kubenswrapper[4775]: E1002 02:01:24.171054 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="ceilometer-notification-agent" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.171059 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="ceilometer-notification-agent" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.171256 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="sg-core" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.171276 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="proxy-httpd" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.171284 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="ceilometer-central-agent" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.171293 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" containerName="ceilometer-notification-agent" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.174796 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.183535 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.183889 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.184913 4775 scope.go:117] "RemoveContainer" containerID="06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.198789 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.216387 4775 scope.go:117] "RemoveContainer" containerID="448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201" Oct 02 02:01:24 crc kubenswrapper[4775]: E1002 02:01:24.217273 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201\": container with ID starting with 448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201 not found: ID does not exist" containerID="448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.217311 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201"} err="failed to get container status \"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201\": rpc error: code = NotFound desc = could not find container \"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201\": container with ID starting with 448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201 not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.217335 4775 scope.go:117] "RemoveContainer" containerID="312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf" Oct 02 02:01:24 crc kubenswrapper[4775]: E1002 02:01:24.217604 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf\": container with ID starting with 312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf not found: ID does not exist" containerID="312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.217620 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf"} err="failed to get container status \"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf\": rpc error: code = NotFound desc = could not find container \"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf\": container with ID starting with 312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.217631 4775 scope.go:117] "RemoveContainer" containerID="2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52" Oct 02 02:01:24 crc kubenswrapper[4775]: E1002 02:01:24.217872 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52\": container with ID starting with 2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52 not found: ID does not exist" containerID="2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.217888 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52"} err="failed to get container status \"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52\": rpc error: code = NotFound desc = could not find container \"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52\": container with ID starting with 2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52 not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.217902 4775 scope.go:117] "RemoveContainer" containerID="06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044" Oct 02 02:01:24 crc kubenswrapper[4775]: E1002 02:01:24.218076 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044\": container with ID starting with 06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044 not found: ID does not exist" containerID="06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.218095 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044"} err="failed to get container status \"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044\": rpc error: code = NotFound desc = could not find container \"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044\": container with ID starting with 06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044 not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.218106 4775 scope.go:117] "RemoveContainer" containerID="448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.218287 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201"} err="failed to get container status \"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201\": rpc error: code = NotFound desc = could not find container \"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201\": container with ID starting with 448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201 not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.218299 4775 scope.go:117] "RemoveContainer" containerID="312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.218599 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf"} err="failed to get container status \"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf\": rpc error: code = NotFound desc = could not find container \"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf\": container with ID starting with 312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.218612 4775 scope.go:117] "RemoveContainer" containerID="2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.218846 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52"} err="failed to get container status \"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52\": rpc error: code = NotFound desc = could not find container \"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52\": container with ID starting with 2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52 not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.218858 4775 scope.go:117] "RemoveContainer" containerID="06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.221586 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044"} err="failed to get container status \"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044\": rpc error: code = NotFound desc = could not find container \"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044\": container with ID starting with 06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044 not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.221651 4775 scope.go:117] "RemoveContainer" containerID="448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.224520 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201"} err="failed to get container status \"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201\": rpc error: code = NotFound desc = could not find container \"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201\": container with ID starting with 448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201 not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.224558 4775 scope.go:117] "RemoveContainer" containerID="312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.226368 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf"} err="failed to get container status \"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf\": rpc error: code = NotFound desc = could not find container \"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf\": container with ID starting with 312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.226438 4775 scope.go:117] "RemoveContainer" containerID="2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.226766 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52"} err="failed to get container status \"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52\": rpc error: code = NotFound desc = could not find container \"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52\": container with ID starting with 2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52 not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.226793 4775 scope.go:117] "RemoveContainer" containerID="06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.229169 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044"} err="failed to get container status \"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044\": rpc error: code = NotFound desc = could not find container \"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044\": container with ID starting with 06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044 not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.229248 4775 scope.go:117] "RemoveContainer" containerID="448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.233257 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201"} err="failed to get container status \"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201\": rpc error: code = NotFound desc = could not find container \"448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201\": container with ID starting with 448f50d16ec5cd843e93fc3acb78078c013967369eaa9952b49fcb3e7e9b9201 not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.233281 4775 scope.go:117] "RemoveContainer" containerID="312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.233755 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf"} err="failed to get container status \"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf\": rpc error: code = NotFound desc = could not find container \"312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf\": container with ID starting with 312c3e5d2f5de07ebdd58d64a0e716962dac0c6ffd3474be4d3f8427c88730bf not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.233796 4775 scope.go:117] "RemoveContainer" containerID="2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.234205 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52"} err="failed to get container status \"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52\": rpc error: code = NotFound desc = could not find container \"2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52\": container with ID starting with 2b9420dd397ac768e424ad2bfa6bd7b82dd95ae7f3de5eb2efb85965f9ca0d52 not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.234226 4775 scope.go:117] "RemoveContainer" containerID="06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.234428 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044"} err="failed to get container status \"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044\": rpc error: code = NotFound desc = could not find container \"06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044\": container with ID starting with 06045667bbcf02754d64a2615eea451182989fe448d446d7873b0ecd6d994044 not found: ID does not exist" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.326532 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xhqg\" (UniqueName: \"kubernetes.io/projected/b05c13f8-c390-4c14-b405-312f24c169a7-kube-api-access-2xhqg\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.326577 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.326621 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-config-data\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.326688 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-scripts\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.326847 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b05c13f8-c390-4c14-b405-312f24c169a7-run-httpd\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.327017 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.327239 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b05c13f8-c390-4c14-b405-312f24c169a7-log-httpd\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.428902 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b05c13f8-c390-4c14-b405-312f24c169a7-log-httpd\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.428992 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xhqg\" (UniqueName: \"kubernetes.io/projected/b05c13f8-c390-4c14-b405-312f24c169a7-kube-api-access-2xhqg\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.429028 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.429066 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-config-data\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.429084 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-scripts\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.429152 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b05c13f8-c390-4c14-b405-312f24c169a7-run-httpd\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.429208 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.429470 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b05c13f8-c390-4c14-b405-312f24c169a7-log-httpd\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.430069 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b05c13f8-c390-4c14-b405-312f24c169a7-run-httpd\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.433622 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-scripts\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.433879 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-config-data\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.435408 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.440800 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.448631 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xhqg\" (UniqueName: \"kubernetes.io/projected/b05c13f8-c390-4c14-b405-312f24c169a7-kube-api-access-2xhqg\") pod \"ceilometer-0\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.498299 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.516149 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:01:24 crc kubenswrapper[4775]: I1002 02:01:24.955699 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.071700 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b05c13f8-c390-4c14-b405-312f24c169a7","Type":"ContainerStarted","Data":"5cd0141732377a65cd43805f01ae31a4b93755faca90868fc4d07fa40781c09d"} Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.074533 4775 generic.go:334] "Generic (PLEG): container finished" podID="c2d22306-6f52-462a-a980-210c208c595e" containerID="4970d9325d20fcb9a2e4c4c76b0503d72a5a3a440e5f7bc4e1e03c493386fb84" exitCode=0 Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.074585 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ddr42" event={"ID":"c2d22306-6f52-462a-a980-210c208c595e","Type":"ContainerDied","Data":"4970d9325d20fcb9a2e4c4c76b0503d72a5a3a440e5f7bc4e1e03c493386fb84"} Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.576212 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.619985 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-b8mb2" Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.621997 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-749868ff64-jpnxz"] Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.622223 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-749868ff64-jpnxz" podUID="bb7c56b0-e35f-4fb9-9746-71fe1be1b027" containerName="neutron-api" containerID="cri-o://2b17fcefad2d7e332e250dd41af49214922dfec671fef00c94a8e45d9d0d89ed" gracePeriod=30 Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.622358 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-749868ff64-jpnxz" podUID="bb7c56b0-e35f-4fb9-9746-71fe1be1b027" containerName="neutron-httpd" containerID="cri-o://a9ab79565384ff2e9f02e123a0e766415185a2be063ed2fb25fd5db22c85c64f" gracePeriod=30 Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.632244 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rjs9s" Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.763832 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7gj7\" (UniqueName: \"kubernetes.io/projected/b01ed253-a630-4b6c-9282-929b34b69b6d-kube-api-access-l7gj7\") pod \"b01ed253-a630-4b6c-9282-929b34b69b6d\" (UID: \"b01ed253-a630-4b6c-9282-929b34b69b6d\") " Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.763945 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22hh2\" (UniqueName: \"kubernetes.io/projected/8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9-kube-api-access-22hh2\") pod \"8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9\" (UID: \"8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9\") " Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.768130 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b01ed253-a630-4b6c-9282-929b34b69b6d-kube-api-access-l7gj7" (OuterVolumeSpecName: "kube-api-access-l7gj7") pod "b01ed253-a630-4b6c-9282-929b34b69b6d" (UID: "b01ed253-a630-4b6c-9282-929b34b69b6d"). InnerVolumeSpecName "kube-api-access-l7gj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.769305 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9-kube-api-access-22hh2" (OuterVolumeSpecName: "kube-api-access-22hh2") pod "8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9" (UID: "8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9"). InnerVolumeSpecName "kube-api-access-22hh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.781715 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78a04f54-4d65-428e-bac7-099665ceaf38" path="/var/lib/kubelet/pods/78a04f54-4d65-428e-bac7-099665ceaf38/volumes" Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.869372 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7gj7\" (UniqueName: \"kubernetes.io/projected/b01ed253-a630-4b6c-9282-929b34b69b6d-kube-api-access-l7gj7\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:25 crc kubenswrapper[4775]: I1002 02:01:25.869410 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22hh2\" (UniqueName: \"kubernetes.io/projected/8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9-kube-api-access-22hh2\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:26 crc kubenswrapper[4775]: I1002 02:01:26.085975 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-b8mb2" event={"ID":"b01ed253-a630-4b6c-9282-929b34b69b6d","Type":"ContainerDied","Data":"910ec8ef8d00e72be83c879fffa937feb04eb691344a5739516a90e025cfdb28"} Oct 02 02:01:26 crc kubenswrapper[4775]: I1002 02:01:26.086013 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="910ec8ef8d00e72be83c879fffa937feb04eb691344a5739516a90e025cfdb28" Oct 02 02:01:26 crc kubenswrapper[4775]: I1002 02:01:26.086097 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-b8mb2" Oct 02 02:01:26 crc kubenswrapper[4775]: I1002 02:01:26.089218 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b05c13f8-c390-4c14-b405-312f24c169a7","Type":"ContainerStarted","Data":"36edc36220dc490ccd56619630b355363f14702467e8e728ba5b532fa5d79dc2"} Oct 02 02:01:26 crc kubenswrapper[4775]: I1002 02:01:26.092814 4775 generic.go:334] "Generic (PLEG): container finished" podID="bb7c56b0-e35f-4fb9-9746-71fe1be1b027" containerID="a9ab79565384ff2e9f02e123a0e766415185a2be063ed2fb25fd5db22c85c64f" exitCode=0 Oct 02 02:01:26 crc kubenswrapper[4775]: I1002 02:01:26.092869 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-749868ff64-jpnxz" event={"ID":"bb7c56b0-e35f-4fb9-9746-71fe1be1b027","Type":"ContainerDied","Data":"a9ab79565384ff2e9f02e123a0e766415185a2be063ed2fb25fd5db22c85c64f"} Oct 02 02:01:26 crc kubenswrapper[4775]: I1002 02:01:26.095306 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-rjs9s" event={"ID":"8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9","Type":"ContainerDied","Data":"e2cd6dba0d5cfa556e0880b9773e4da47c7f578495d208b97fca9a70948a7cf4"} Oct 02 02:01:26 crc kubenswrapper[4775]: I1002 02:01:26.095343 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2cd6dba0d5cfa556e0880b9773e4da47c7f578495d208b97fca9a70948a7cf4" Oct 02 02:01:26 crc kubenswrapper[4775]: I1002 02:01:26.095459 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-rjs9s" Oct 02 02:01:26 crc kubenswrapper[4775]: I1002 02:01:26.364997 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ddr42" Oct 02 02:01:26 crc kubenswrapper[4775]: I1002 02:01:26.479509 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2frlb\" (UniqueName: \"kubernetes.io/projected/c2d22306-6f52-462a-a980-210c208c595e-kube-api-access-2frlb\") pod \"c2d22306-6f52-462a-a980-210c208c595e\" (UID: \"c2d22306-6f52-462a-a980-210c208c595e\") " Oct 02 02:01:26 crc kubenswrapper[4775]: I1002 02:01:26.483235 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2d22306-6f52-462a-a980-210c208c595e-kube-api-access-2frlb" (OuterVolumeSpecName: "kube-api-access-2frlb") pod "c2d22306-6f52-462a-a980-210c208c595e" (UID: "c2d22306-6f52-462a-a980-210c208c595e"). InnerVolumeSpecName "kube-api-access-2frlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:26 crc kubenswrapper[4775]: I1002 02:01:26.581782 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2frlb\" (UniqueName: \"kubernetes.io/projected/c2d22306-6f52-462a-a980-210c208c595e-kube-api-access-2frlb\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:26 crc kubenswrapper[4775]: I1002 02:01:26.753831 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:27 crc kubenswrapper[4775]: I1002 02:01:27.103975 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-ddr42" event={"ID":"c2d22306-6f52-462a-a980-210c208c595e","Type":"ContainerDied","Data":"faf1666928cea68fd0e07129eb445fa421edc7ca960485b5baf5da00acff475a"} Oct 02 02:01:27 crc kubenswrapper[4775]: I1002 02:01:27.104012 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="faf1666928cea68fd0e07129eb445fa421edc7ca960485b5baf5da00acff475a" Oct 02 02:01:27 crc kubenswrapper[4775]: I1002 02:01:27.104067 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-ddr42" Oct 02 02:01:27 crc kubenswrapper[4775]: I1002 02:01:27.110259 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b05c13f8-c390-4c14-b405-312f24c169a7","Type":"ContainerStarted","Data":"f465f468da2e96ad167ecb5889ef0b393c487f154b84cd8a569ea587e0052914"} Oct 02 02:01:27 crc kubenswrapper[4775]: I1002 02:01:27.110316 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b05c13f8-c390-4c14-b405-312f24c169a7","Type":"ContainerStarted","Data":"f66f06cabe0051182dc6470ed27b096c58fda0d8560264dc26d62abaa44f98f8"} Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.143492 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b05c13f8-c390-4c14-b405-312f24c169a7","Type":"ContainerStarted","Data":"88017ffbec8f1e1f2fdfaa83612f23fec6d40519cc2bf7ab1c7944f9334ca06a"} Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.144024 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.143840 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="proxy-httpd" containerID="cri-o://88017ffbec8f1e1f2fdfaa83612f23fec6d40519cc2bf7ab1c7944f9334ca06a" gracePeriod=30 Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.143621 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="ceilometer-central-agent" containerID="cri-o://36edc36220dc490ccd56619630b355363f14702467e8e728ba5b532fa5d79dc2" gracePeriod=30 Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.143859 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="ceilometer-notification-agent" containerID="cri-o://f66f06cabe0051182dc6470ed27b096c58fda0d8560264dc26d62abaa44f98f8" gracePeriod=30 Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.143851 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="sg-core" containerID="cri-o://f465f468da2e96ad167ecb5889ef0b393c487f154b84cd8a569ea587e0052914" gracePeriod=30 Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.571897 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.592065 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.273513232 podStartE2EDuration="5.592048312s" podCreationTimestamp="2025-10-02 02:01:24 +0000 UTC" firstStartedPulling="2025-10-02 02:01:24.959705034 +0000 UTC m=+1222.126449074" lastFinishedPulling="2025-10-02 02:01:28.278240094 +0000 UTC m=+1225.444984154" observedRunningTime="2025-10-02 02:01:29.17174415 +0000 UTC m=+1226.338488190" watchObservedRunningTime="2025-10-02 02:01:29.592048312 +0000 UTC m=+1226.758792352" Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.742396 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-762zr\" (UniqueName: \"kubernetes.io/projected/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-kube-api-access-762zr\") pod \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.742451 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-combined-ca-bundle\") pod \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.742625 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-httpd-config\") pod \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.742669 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-config\") pod \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.742723 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-ovndb-tls-certs\") pod \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\" (UID: \"bb7c56b0-e35f-4fb9-9746-71fe1be1b027\") " Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.749134 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "bb7c56b0-e35f-4fb9-9746-71fe1be1b027" (UID: "bb7c56b0-e35f-4fb9-9746-71fe1be1b027"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.749976 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-kube-api-access-762zr" (OuterVolumeSpecName: "kube-api-access-762zr") pod "bb7c56b0-e35f-4fb9-9746-71fe1be1b027" (UID: "bb7c56b0-e35f-4fb9-9746-71fe1be1b027"). InnerVolumeSpecName "kube-api-access-762zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.813194 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb7c56b0-e35f-4fb9-9746-71fe1be1b027" (UID: "bb7c56b0-e35f-4fb9-9746-71fe1be1b027"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.839305 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-config" (OuterVolumeSpecName: "config") pod "bb7c56b0-e35f-4fb9-9746-71fe1be1b027" (UID: "bb7c56b0-e35f-4fb9-9746-71fe1be1b027"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.842256 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "bb7c56b0-e35f-4fb9-9746-71fe1be1b027" (UID: "bb7c56b0-e35f-4fb9-9746-71fe1be1b027"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.845199 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.845228 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.845238 4775 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.845250 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-762zr\" (UniqueName: \"kubernetes.io/projected/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-kube-api-access-762zr\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:29 crc kubenswrapper[4775]: I1002 02:01:29.845259 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb7c56b0-e35f-4fb9-9746-71fe1be1b027-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.163458 4775 generic.go:334] "Generic (PLEG): container finished" podID="b05c13f8-c390-4c14-b405-312f24c169a7" containerID="88017ffbec8f1e1f2fdfaa83612f23fec6d40519cc2bf7ab1c7944f9334ca06a" exitCode=0 Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.163487 4775 generic.go:334] "Generic (PLEG): container finished" podID="b05c13f8-c390-4c14-b405-312f24c169a7" containerID="f465f468da2e96ad167ecb5889ef0b393c487f154b84cd8a569ea587e0052914" exitCode=2 Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.163494 4775 generic.go:334] "Generic (PLEG): container finished" podID="b05c13f8-c390-4c14-b405-312f24c169a7" containerID="f66f06cabe0051182dc6470ed27b096c58fda0d8560264dc26d62abaa44f98f8" exitCode=0 Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.163548 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b05c13f8-c390-4c14-b405-312f24c169a7","Type":"ContainerDied","Data":"88017ffbec8f1e1f2fdfaa83612f23fec6d40519cc2bf7ab1c7944f9334ca06a"} Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.163605 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b05c13f8-c390-4c14-b405-312f24c169a7","Type":"ContainerDied","Data":"f465f468da2e96ad167ecb5889ef0b393c487f154b84cd8a569ea587e0052914"} Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.163623 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b05c13f8-c390-4c14-b405-312f24c169a7","Type":"ContainerDied","Data":"f66f06cabe0051182dc6470ed27b096c58fda0d8560264dc26d62abaa44f98f8"} Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.169714 4775 generic.go:334] "Generic (PLEG): container finished" podID="bb7c56b0-e35f-4fb9-9746-71fe1be1b027" containerID="2b17fcefad2d7e332e250dd41af49214922dfec671fef00c94a8e45d9d0d89ed" exitCode=0 Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.169757 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-749868ff64-jpnxz" event={"ID":"bb7c56b0-e35f-4fb9-9746-71fe1be1b027","Type":"ContainerDied","Data":"2b17fcefad2d7e332e250dd41af49214922dfec671fef00c94a8e45d9d0d89ed"} Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.169785 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-749868ff64-jpnxz" Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.169815 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-749868ff64-jpnxz" event={"ID":"bb7c56b0-e35f-4fb9-9746-71fe1be1b027","Type":"ContainerDied","Data":"0d9edc310c8f5745221c21900728828e618c061e1c09f2f5eade243c870abd9e"} Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.169850 4775 scope.go:117] "RemoveContainer" containerID="a9ab79565384ff2e9f02e123a0e766415185a2be063ed2fb25fd5db22c85c64f" Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.230477 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-749868ff64-jpnxz"] Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.235202 4775 scope.go:117] "RemoveContainer" containerID="2b17fcefad2d7e332e250dd41af49214922dfec671fef00c94a8e45d9d0d89ed" Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.238107 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-749868ff64-jpnxz"] Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.244873 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.245224 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" containerName="glance-log" containerID="cri-o://6c7739e2aa0e5396b03e846410f94dee4c0629af1af171c37f25be6c5d7aee75" gracePeriod=30 Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.245362 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" containerName="glance-httpd" containerID="cri-o://1e4b5c78eaff0fb64d29ae3fd2247940e4dbd27d85e5d5983cb6837acb633966" gracePeriod=30 Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.258471 4775 scope.go:117] "RemoveContainer" containerID="a9ab79565384ff2e9f02e123a0e766415185a2be063ed2fb25fd5db22c85c64f" Oct 02 02:01:30 crc kubenswrapper[4775]: E1002 02:01:30.261305 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9ab79565384ff2e9f02e123a0e766415185a2be063ed2fb25fd5db22c85c64f\": container with ID starting with a9ab79565384ff2e9f02e123a0e766415185a2be063ed2fb25fd5db22c85c64f not found: ID does not exist" containerID="a9ab79565384ff2e9f02e123a0e766415185a2be063ed2fb25fd5db22c85c64f" Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.261404 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9ab79565384ff2e9f02e123a0e766415185a2be063ed2fb25fd5db22c85c64f"} err="failed to get container status \"a9ab79565384ff2e9f02e123a0e766415185a2be063ed2fb25fd5db22c85c64f\": rpc error: code = NotFound desc = could not find container \"a9ab79565384ff2e9f02e123a0e766415185a2be063ed2fb25fd5db22c85c64f\": container with ID starting with a9ab79565384ff2e9f02e123a0e766415185a2be063ed2fb25fd5db22c85c64f not found: ID does not exist" Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.261442 4775 scope.go:117] "RemoveContainer" containerID="2b17fcefad2d7e332e250dd41af49214922dfec671fef00c94a8e45d9d0d89ed" Oct 02 02:01:30 crc kubenswrapper[4775]: E1002 02:01:30.261700 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b17fcefad2d7e332e250dd41af49214922dfec671fef00c94a8e45d9d0d89ed\": container with ID starting with 2b17fcefad2d7e332e250dd41af49214922dfec671fef00c94a8e45d9d0d89ed not found: ID does not exist" containerID="2b17fcefad2d7e332e250dd41af49214922dfec671fef00c94a8e45d9d0d89ed" Oct 02 02:01:30 crc kubenswrapper[4775]: I1002 02:01:30.261719 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b17fcefad2d7e332e250dd41af49214922dfec671fef00c94a8e45d9d0d89ed"} err="failed to get container status \"2b17fcefad2d7e332e250dd41af49214922dfec671fef00c94a8e45d9d0d89ed\": rpc error: code = NotFound desc = could not find container \"2b17fcefad2d7e332e250dd41af49214922dfec671fef00c94a8e45d9d0d89ed\": container with ID starting with 2b17fcefad2d7e332e250dd41af49214922dfec671fef00c94a8e45d9d0d89ed not found: ID does not exist" Oct 02 02:01:31 crc kubenswrapper[4775]: I1002 02:01:31.180421 4775 generic.go:334] "Generic (PLEG): container finished" podID="b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" containerID="6c7739e2aa0e5396b03e846410f94dee4c0629af1af171c37f25be6c5d7aee75" exitCode=143 Oct 02 02:01:31 crc kubenswrapper[4775]: I1002 02:01:31.180462 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb","Type":"ContainerDied","Data":"6c7739e2aa0e5396b03e846410f94dee4c0629af1af171c37f25be6c5d7aee75"} Oct 02 02:01:31 crc kubenswrapper[4775]: I1002 02:01:31.776569 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb7c56b0-e35f-4fb9-9746-71fe1be1b027" path="/var/lib/kubelet/pods/bb7c56b0-e35f-4fb9-9746-71fe1be1b027/volumes" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.776205 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.776450 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" containerName="glance-log" containerID="cri-o://229dd58f2fa05172f374734e3ca5ed77fc063711c568f4fe9c41b9a071b85798" gracePeriod=30 Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.776875 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" containerName="glance-httpd" containerID="cri-o://4b96b0239299787f213eec6812a29703fc5113d42a71f339aa0bca3867dc4fa0" gracePeriod=30 Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.783226 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/glance-default-internal-api-0" podUID="d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.152:9292/healthcheck\": EOF" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.919285 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-f3f8-account-create-86qvz"] Oct 02 02:01:32 crc kubenswrapper[4775]: E1002 02:01:32.919692 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb7c56b0-e35f-4fb9-9746-71fe1be1b027" containerName="neutron-api" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.919712 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb7c56b0-e35f-4fb9-9746-71fe1be1b027" containerName="neutron-api" Oct 02 02:01:32 crc kubenswrapper[4775]: E1002 02:01:32.919737 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb7c56b0-e35f-4fb9-9746-71fe1be1b027" containerName="neutron-httpd" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.919746 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb7c56b0-e35f-4fb9-9746-71fe1be1b027" containerName="neutron-httpd" Oct 02 02:01:32 crc kubenswrapper[4775]: E1002 02:01:32.919762 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2d22306-6f52-462a-a980-210c208c595e" containerName="mariadb-database-create" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.919770 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2d22306-6f52-462a-a980-210c208c595e" containerName="mariadb-database-create" Oct 02 02:01:32 crc kubenswrapper[4775]: E1002 02:01:32.919794 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b01ed253-a630-4b6c-9282-929b34b69b6d" containerName="mariadb-database-create" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.919802 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b01ed253-a630-4b6c-9282-929b34b69b6d" containerName="mariadb-database-create" Oct 02 02:01:32 crc kubenswrapper[4775]: E1002 02:01:32.919829 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9" containerName="mariadb-database-create" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.919838 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9" containerName="mariadb-database-create" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.920074 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b01ed253-a630-4b6c-9282-929b34b69b6d" containerName="mariadb-database-create" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.920104 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2d22306-6f52-462a-a980-210c208c595e" containerName="mariadb-database-create" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.920113 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb7c56b0-e35f-4fb9-9746-71fe1be1b027" containerName="neutron-api" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.920131 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb7c56b0-e35f-4fb9-9746-71fe1be1b027" containerName="neutron-httpd" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.920158 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9" containerName="mariadb-database-create" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.920840 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f3f8-account-create-86qvz" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.925223 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 02 02:01:32 crc kubenswrapper[4775]: I1002 02:01:32.930170 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f3f8-account-create-86qvz"] Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.032536 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt6pw\" (UniqueName: \"kubernetes.io/projected/a120b1c5-156f-432c-96ca-f9a4dfc6dfa7-kube-api-access-zt6pw\") pod \"nova-api-f3f8-account-create-86qvz\" (UID: \"a120b1c5-156f-432c-96ca-f9a4dfc6dfa7\") " pod="openstack/nova-api-f3f8-account-create-86qvz" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.118456 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-0729-account-create-v8vz9"] Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.119609 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0729-account-create-v8vz9" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.121440 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.128869 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0729-account-create-v8vz9"] Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.133656 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt6pw\" (UniqueName: \"kubernetes.io/projected/a120b1c5-156f-432c-96ca-f9a4dfc6dfa7-kube-api-access-zt6pw\") pod \"nova-api-f3f8-account-create-86qvz\" (UID: \"a120b1c5-156f-432c-96ca-f9a4dfc6dfa7\") " pod="openstack/nova-api-f3f8-account-create-86qvz" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.156074 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt6pw\" (UniqueName: \"kubernetes.io/projected/a120b1c5-156f-432c-96ca-f9a4dfc6dfa7-kube-api-access-zt6pw\") pod \"nova-api-f3f8-account-create-86qvz\" (UID: \"a120b1c5-156f-432c-96ca-f9a4dfc6dfa7\") " pod="openstack/nova-api-f3f8-account-create-86qvz" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.198761 4775 generic.go:334] "Generic (PLEG): container finished" podID="d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" containerID="229dd58f2fa05172f374734e3ca5ed77fc063711c568f4fe9c41b9a071b85798" exitCode=143 Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.198806 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08","Type":"ContainerDied","Data":"229dd58f2fa05172f374734e3ca5ed77fc063711c568f4fe9c41b9a071b85798"} Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.235628 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4vhz\" (UniqueName: \"kubernetes.io/projected/e19ef24d-2e7c-47a9-943c-a49f650df3a7-kube-api-access-q4vhz\") pod \"nova-cell0-0729-account-create-v8vz9\" (UID: \"e19ef24d-2e7c-47a9-943c-a49f650df3a7\") " pod="openstack/nova-cell0-0729-account-create-v8vz9" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.237271 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f3f8-account-create-86qvz" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.320938 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-5350-account-create-lk56t"] Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.322134 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5350-account-create-lk56t" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.326880 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.337106 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4vhz\" (UniqueName: \"kubernetes.io/projected/e19ef24d-2e7c-47a9-943c-a49f650df3a7-kube-api-access-q4vhz\") pod \"nova-cell0-0729-account-create-v8vz9\" (UID: \"e19ef24d-2e7c-47a9-943c-a49f650df3a7\") " pod="openstack/nova-cell0-0729-account-create-v8vz9" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.349980 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5350-account-create-lk56t"] Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.357104 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4vhz\" (UniqueName: \"kubernetes.io/projected/e19ef24d-2e7c-47a9-943c-a49f650df3a7-kube-api-access-q4vhz\") pod \"nova-cell0-0729-account-create-v8vz9\" (UID: \"e19ef24d-2e7c-47a9-943c-a49f650df3a7\") " pod="openstack/nova-cell0-0729-account-create-v8vz9" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.436624 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0729-account-create-v8vz9" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.438805 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl78n\" (UniqueName: \"kubernetes.io/projected/833fd434-3bdc-4be7-be6f-f06915509129-kube-api-access-tl78n\") pod \"nova-cell1-5350-account-create-lk56t\" (UID: \"833fd434-3bdc-4be7-be6f-f06915509129\") " pod="openstack/nova-cell1-5350-account-create-lk56t" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.544626 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl78n\" (UniqueName: \"kubernetes.io/projected/833fd434-3bdc-4be7-be6f-f06915509129-kube-api-access-tl78n\") pod \"nova-cell1-5350-account-create-lk56t\" (UID: \"833fd434-3bdc-4be7-be6f-f06915509129\") " pod="openstack/nova-cell1-5350-account-create-lk56t" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.566049 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl78n\" (UniqueName: \"kubernetes.io/projected/833fd434-3bdc-4be7-be6f-f06915509129-kube-api-access-tl78n\") pod \"nova-cell1-5350-account-create-lk56t\" (UID: \"833fd434-3bdc-4be7-be6f-f06915509129\") " pod="openstack/nova-cell1-5350-account-create-lk56t" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.707353 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5350-account-create-lk56t" Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.861919 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f3f8-account-create-86qvz"] Oct 02 02:01:33 crc kubenswrapper[4775]: I1002 02:01:33.975721 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-0729-account-create-v8vz9"] Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.203179 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.220335 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f3f8-account-create-86qvz" event={"ID":"a120b1c5-156f-432c-96ca-f9a4dfc6dfa7","Type":"ContainerStarted","Data":"aa2f27d228e26a006ad01a4dceff0bf93b6b28e03f034ad9b684c6df38e9734a"} Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.240418 4775 generic.go:334] "Generic (PLEG): container finished" podID="b05c13f8-c390-4c14-b405-312f24c169a7" containerID="36edc36220dc490ccd56619630b355363f14702467e8e728ba5b532fa5d79dc2" exitCode=0 Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.240485 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b05c13f8-c390-4c14-b405-312f24c169a7","Type":"ContainerDied","Data":"36edc36220dc490ccd56619630b355363f14702467e8e728ba5b532fa5d79dc2"} Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.264291 4775 generic.go:334] "Generic (PLEG): container finished" podID="b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" containerID="1e4b5c78eaff0fb64d29ae3fd2247940e4dbd27d85e5d5983cb6837acb633966" exitCode=0 Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.264357 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb","Type":"ContainerDied","Data":"1e4b5c78eaff0fb64d29ae3fd2247940e4dbd27d85e5d5983cb6837acb633966"} Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.264383 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb","Type":"ContainerDied","Data":"586fe3ca1019c7dbd5e301cb818102473c665e29499d8629e9cb0d7dcd2f8cea"} Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.264398 4775 scope.go:117] "RemoveContainer" containerID="1e4b5c78eaff0fb64d29ae3fd2247940e4dbd27d85e5d5983cb6837acb633966" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.264516 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.272554 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0729-account-create-v8vz9" event={"ID":"e19ef24d-2e7c-47a9-943c-a49f650df3a7","Type":"ContainerStarted","Data":"6e3e3d0f497edd4f71d791cee09b0583a93344aeb9d2a618873a5c901104c27c"} Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.282532 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-config-data\") pod \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.282582 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-httpd-run\") pod \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.282598 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-combined-ca-bundle\") pod \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.282622 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-public-tls-certs\") pod \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.283138 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" (UID: "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.311554 4775 scope.go:117] "RemoveContainer" containerID="6c7739e2aa0e5396b03e846410f94dee4c0629af1af171c37f25be6c5d7aee75" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.333657 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-5350-account-create-lk56t"] Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.339377 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" (UID: "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.343857 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-config-data" (OuterVolumeSpecName: "config-data") pod "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" (UID: "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.384605 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.384693 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-logs\") pod \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.384731 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkq55\" (UniqueName: \"kubernetes.io/projected/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-kube-api-access-lkq55\") pod \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.384896 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-scripts\") pod \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\" (UID: \"b59797fc-1aa1-4106-ba3f-7239d6f4c7fb\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.385156 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-logs" (OuterVolumeSpecName: "logs") pod "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" (UID: "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.385304 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.385318 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.385328 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.385338 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.387329 4775 scope.go:117] "RemoveContainer" containerID="1e4b5c78eaff0fb64d29ae3fd2247940e4dbd27d85e5d5983cb6837acb633966" Oct 02 02:01:34 crc kubenswrapper[4775]: E1002 02:01:34.388594 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e4b5c78eaff0fb64d29ae3fd2247940e4dbd27d85e5d5983cb6837acb633966\": container with ID starting with 1e4b5c78eaff0fb64d29ae3fd2247940e4dbd27d85e5d5983cb6837acb633966 not found: ID does not exist" containerID="1e4b5c78eaff0fb64d29ae3fd2247940e4dbd27d85e5d5983cb6837acb633966" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.388635 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e4b5c78eaff0fb64d29ae3fd2247940e4dbd27d85e5d5983cb6837acb633966"} err="failed to get container status \"1e4b5c78eaff0fb64d29ae3fd2247940e4dbd27d85e5d5983cb6837acb633966\": rpc error: code = NotFound desc = could not find container \"1e4b5c78eaff0fb64d29ae3fd2247940e4dbd27d85e5d5983cb6837acb633966\": container with ID starting with 1e4b5c78eaff0fb64d29ae3fd2247940e4dbd27d85e5d5983cb6837acb633966 not found: ID does not exist" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.388662 4775 scope.go:117] "RemoveContainer" containerID="6c7739e2aa0e5396b03e846410f94dee4c0629af1af171c37f25be6c5d7aee75" Oct 02 02:01:34 crc kubenswrapper[4775]: E1002 02:01:34.389321 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c7739e2aa0e5396b03e846410f94dee4c0629af1af171c37f25be6c5d7aee75\": container with ID starting with 6c7739e2aa0e5396b03e846410f94dee4c0629af1af171c37f25be6c5d7aee75 not found: ID does not exist" containerID="6c7739e2aa0e5396b03e846410f94dee4c0629af1af171c37f25be6c5d7aee75" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.389352 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c7739e2aa0e5396b03e846410f94dee4c0629af1af171c37f25be6c5d7aee75"} err="failed to get container status \"6c7739e2aa0e5396b03e846410f94dee4c0629af1af171c37f25be6c5d7aee75\": rpc error: code = NotFound desc = could not find container \"6c7739e2aa0e5396b03e846410f94dee4c0629af1af171c37f25be6c5d7aee75\": container with ID starting with 6c7739e2aa0e5396b03e846410f94dee4c0629af1af171c37f25be6c5d7aee75 not found: ID does not exist" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.394819 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-kube-api-access-lkq55" (OuterVolumeSpecName: "kube-api-access-lkq55") pod "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" (UID: "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb"). InnerVolumeSpecName "kube-api-access-lkq55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.400041 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-scripts" (OuterVolumeSpecName: "scripts") pod "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" (UID: "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.400131 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" (UID: "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.410173 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" (UID: "b59797fc-1aa1-4106-ba3f-7239d6f4c7fb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.425306 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.486304 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-config-data\") pod \"b05c13f8-c390-4c14-b405-312f24c169a7\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.486337 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b05c13f8-c390-4c14-b405-312f24c169a7-log-httpd\") pod \"b05c13f8-c390-4c14-b405-312f24c169a7\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.486392 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xhqg\" (UniqueName: \"kubernetes.io/projected/b05c13f8-c390-4c14-b405-312f24c169a7-kube-api-access-2xhqg\") pod \"b05c13f8-c390-4c14-b405-312f24c169a7\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.486470 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b05c13f8-c390-4c14-b405-312f24c169a7-run-httpd\") pod \"b05c13f8-c390-4c14-b405-312f24c169a7\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.486489 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-sg-core-conf-yaml\") pod \"b05c13f8-c390-4c14-b405-312f24c169a7\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.486519 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-combined-ca-bundle\") pod \"b05c13f8-c390-4c14-b405-312f24c169a7\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.486534 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-scripts\") pod \"b05c13f8-c390-4c14-b405-312f24c169a7\" (UID: \"b05c13f8-c390-4c14-b405-312f24c169a7\") " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.486748 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkq55\" (UniqueName: \"kubernetes.io/projected/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-kube-api-access-lkq55\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.486758 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.486767 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.486784 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.493316 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b05c13f8-c390-4c14-b405-312f24c169a7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b05c13f8-c390-4c14-b405-312f24c169a7" (UID: "b05c13f8-c390-4c14-b405-312f24c169a7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.493581 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b05c13f8-c390-4c14-b405-312f24c169a7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b05c13f8-c390-4c14-b405-312f24c169a7" (UID: "b05c13f8-c390-4c14-b405-312f24c169a7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.497173 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b05c13f8-c390-4c14-b405-312f24c169a7-kube-api-access-2xhqg" (OuterVolumeSpecName: "kube-api-access-2xhqg") pod "b05c13f8-c390-4c14-b405-312f24c169a7" (UID: "b05c13f8-c390-4c14-b405-312f24c169a7"). InnerVolumeSpecName "kube-api-access-2xhqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.498495 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-scripts" (OuterVolumeSpecName: "scripts") pod "b05c13f8-c390-4c14-b405-312f24c169a7" (UID: "b05c13f8-c390-4c14-b405-312f24c169a7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.518522 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b05c13f8-c390-4c14-b405-312f24c169a7" (UID: "b05c13f8-c390-4c14-b405-312f24c169a7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.519552 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.583171 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b05c13f8-c390-4c14-b405-312f24c169a7" (UID: "b05c13f8-c390-4c14-b405-312f24c169a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.583583 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-config-data" (OuterVolumeSpecName: "config-data") pod "b05c13f8-c390-4c14-b405-312f24c169a7" (UID: "b05c13f8-c390-4c14-b405-312f24c169a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.588917 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.588947 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.588978 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.588988 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b05c13f8-c390-4c14-b405-312f24c169a7-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.588997 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b05c13f8-c390-4c14-b405-312f24c169a7-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.589007 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.589042 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xhqg\" (UniqueName: \"kubernetes.io/projected/b05c13f8-c390-4c14-b405-312f24c169a7-kube-api-access-2xhqg\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.589052 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b05c13f8-c390-4c14-b405-312f24c169a7-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.632194 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.640528 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.668106 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:01:34 crc kubenswrapper[4775]: E1002 02:01:34.668446 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="sg-core" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.668461 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="sg-core" Oct 02 02:01:34 crc kubenswrapper[4775]: E1002 02:01:34.668491 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="proxy-httpd" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.668499 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="proxy-httpd" Oct 02 02:01:34 crc kubenswrapper[4775]: E1002 02:01:34.668517 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="ceilometer-notification-agent" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.668524 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="ceilometer-notification-agent" Oct 02 02:01:34 crc kubenswrapper[4775]: E1002 02:01:34.668532 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" containerName="glance-httpd" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.668537 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" containerName="glance-httpd" Oct 02 02:01:34 crc kubenswrapper[4775]: E1002 02:01:34.668547 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="ceilometer-central-agent" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.668554 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="ceilometer-central-agent" Oct 02 02:01:34 crc kubenswrapper[4775]: E1002 02:01:34.668570 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" containerName="glance-log" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.668577 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" containerName="glance-log" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.668735 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="ceilometer-central-agent" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.668748 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" containerName="glance-log" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.668758 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="proxy-httpd" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.668767 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" containerName="glance-httpd" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.668774 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="ceilometer-notification-agent" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.668782 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" containerName="sg-core" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.670573 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.674610 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.674817 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.698589 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.792103 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfkzl\" (UniqueName: \"kubernetes.io/projected/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-kube-api-access-kfkzl\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.792159 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-config-data\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.792201 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-logs\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.792219 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.792239 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.792258 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.792373 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-scripts\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.792436 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.894486 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfkzl\" (UniqueName: \"kubernetes.io/projected/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-kube-api-access-kfkzl\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.894544 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-config-data\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.894591 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-logs\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.894615 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.894644 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.894671 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.894763 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-scripts\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.894792 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.895278 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.895373 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-logs\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.895557 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.899381 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.900758 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-scripts\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.901202 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-config-data\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.903478 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.918809 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfkzl\" (UniqueName: \"kubernetes.io/projected/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-kube-api-access-kfkzl\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:34 crc kubenswrapper[4775]: I1002 02:01:34.933678 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-0\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " pod="openstack/glance-default-external-api-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.005918 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.286314 4775 generic.go:334] "Generic (PLEG): container finished" podID="a120b1c5-156f-432c-96ca-f9a4dfc6dfa7" containerID="220d9b9728d9af47f9f7f6b6a307303f3baa2a782a03e007f8e2182217bf116d" exitCode=0 Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.286392 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f3f8-account-create-86qvz" event={"ID":"a120b1c5-156f-432c-96ca-f9a4dfc6dfa7","Type":"ContainerDied","Data":"220d9b9728d9af47f9f7f6b6a307303f3baa2a782a03e007f8e2182217bf116d"} Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.293436 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b05c13f8-c390-4c14-b405-312f24c169a7","Type":"ContainerDied","Data":"5cd0141732377a65cd43805f01ae31a4b93755faca90868fc4d07fa40781c09d"} Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.293559 4775 scope.go:117] "RemoveContainer" containerID="88017ffbec8f1e1f2fdfaa83612f23fec6d40519cc2bf7ab1c7944f9334ca06a" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.293736 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.309938 4775 generic.go:334] "Generic (PLEG): container finished" podID="833fd434-3bdc-4be7-be6f-f06915509129" containerID="424d08625ff4c08c37f7520793f0493961f7668b4af123a929558e4a08e9f894" exitCode=0 Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.310180 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5350-account-create-lk56t" event={"ID":"833fd434-3bdc-4be7-be6f-f06915509129","Type":"ContainerDied","Data":"424d08625ff4c08c37f7520793f0493961f7668b4af123a929558e4a08e9f894"} Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.310258 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5350-account-create-lk56t" event={"ID":"833fd434-3bdc-4be7-be6f-f06915509129","Type":"ContainerStarted","Data":"7437a851faf84e55c1c6a48acb6aa1ded3fde445e5563c74507480004c79db67"} Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.317018 4775 generic.go:334] "Generic (PLEG): container finished" podID="e19ef24d-2e7c-47a9-943c-a49f650df3a7" containerID="cdf8eb01511f36ccb375b3270b0dcf26234a99df6e6ec88e6e129895d780779d" exitCode=0 Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.317137 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0729-account-create-v8vz9" event={"ID":"e19ef24d-2e7c-47a9-943c-a49f650df3a7","Type":"ContainerDied","Data":"cdf8eb01511f36ccb375b3270b0dcf26234a99df6e6ec88e6e129895d780779d"} Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.327436 4775 scope.go:117] "RemoveContainer" containerID="f465f468da2e96ad167ecb5889ef0b393c487f154b84cd8a569ea587e0052914" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.364828 4775 scope.go:117] "RemoveContainer" containerID="f66f06cabe0051182dc6470ed27b096c58fda0d8560264dc26d62abaa44f98f8" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.389320 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.397174 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.399936 4775 scope.go:117] "RemoveContainer" containerID="36edc36220dc490ccd56619630b355363f14702467e8e728ba5b532fa5d79dc2" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.415366 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.418042 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.422410 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.423146 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.432658 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.536619 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-scripts\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.536738 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.537065 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1319e744-d454-4024-898e-ac338b1b8443-log-httpd\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.537303 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.538148 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-config-data\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.538218 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slbvn\" (UniqueName: \"kubernetes.io/projected/1319e744-d454-4024-898e-ac338b1b8443-kube-api-access-slbvn\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.538286 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1319e744-d454-4024-898e-ac338b1b8443-run-httpd\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.608297 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.640376 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-scripts\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.640424 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.640444 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1319e744-d454-4024-898e-ac338b1b8443-log-httpd\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.640492 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.640523 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-config-data\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.640546 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slbvn\" (UniqueName: \"kubernetes.io/projected/1319e744-d454-4024-898e-ac338b1b8443-kube-api-access-slbvn\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.640576 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1319e744-d454-4024-898e-ac338b1b8443-run-httpd\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.640967 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1319e744-d454-4024-898e-ac338b1b8443-run-httpd\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.641073 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1319e744-d454-4024-898e-ac338b1b8443-log-httpd\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.649129 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-config-data\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.649638 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.649866 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-scripts\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.649924 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.655044 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slbvn\" (UniqueName: \"kubernetes.io/projected/1319e744-d454-4024-898e-ac338b1b8443-kube-api-access-slbvn\") pod \"ceilometer-0\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.738118 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.775695 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b05c13f8-c390-4c14-b405-312f24c169a7" path="/var/lib/kubelet/pods/b05c13f8-c390-4c14-b405-312f24c169a7/volumes" Oct 02 02:01:35 crc kubenswrapper[4775]: I1002 02:01:35.776644 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b59797fc-1aa1-4106-ba3f-7239d6f4c7fb" path="/var/lib/kubelet/pods/b59797fc-1aa1-4106-ba3f-7239d6f4c7fb/volumes" Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.192986 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:01:36 crc kubenswrapper[4775]: W1002 02:01:36.222805 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1319e744_d454_4024_898e_ac338b1b8443.slice/crio-687d57ea852571f9784926dff665df1db5d839665a912bd929a9e716a4f50bbe WatchSource:0}: Error finding container 687d57ea852571f9784926dff665df1db5d839665a912bd929a9e716a4f50bbe: Status 404 returned error can't find the container with id 687d57ea852571f9784926dff665df1db5d839665a912bd929a9e716a4f50bbe Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.332095 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1319e744-d454-4024-898e-ac338b1b8443","Type":"ContainerStarted","Data":"687d57ea852571f9784926dff665df1db5d839665a912bd929a9e716a4f50bbe"} Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.333893 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51","Type":"ContainerStarted","Data":"12f66def0a878a3dc74caa0f75419c15f378c889154267260d12b5003e5ebee3"} Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.333924 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51","Type":"ContainerStarted","Data":"7a38314d1ae75675da87ea02ecbe9d1fac0745fb118de7950280d4535aa753e7"} Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.780745 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0729-account-create-v8vz9" Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.826925 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5350-account-create-lk56t" Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.832679 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f3f8-account-create-86qvz" Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.878746 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl78n\" (UniqueName: \"kubernetes.io/projected/833fd434-3bdc-4be7-be6f-f06915509129-kube-api-access-tl78n\") pod \"833fd434-3bdc-4be7-be6f-f06915509129\" (UID: \"833fd434-3bdc-4be7-be6f-f06915509129\") " Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.878854 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zt6pw\" (UniqueName: \"kubernetes.io/projected/a120b1c5-156f-432c-96ca-f9a4dfc6dfa7-kube-api-access-zt6pw\") pod \"a120b1c5-156f-432c-96ca-f9a4dfc6dfa7\" (UID: \"a120b1c5-156f-432c-96ca-f9a4dfc6dfa7\") " Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.878937 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4vhz\" (UniqueName: \"kubernetes.io/projected/e19ef24d-2e7c-47a9-943c-a49f650df3a7-kube-api-access-q4vhz\") pod \"e19ef24d-2e7c-47a9-943c-a49f650df3a7\" (UID: \"e19ef24d-2e7c-47a9-943c-a49f650df3a7\") " Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.883636 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a120b1c5-156f-432c-96ca-f9a4dfc6dfa7-kube-api-access-zt6pw" (OuterVolumeSpecName: "kube-api-access-zt6pw") pod "a120b1c5-156f-432c-96ca-f9a4dfc6dfa7" (UID: "a120b1c5-156f-432c-96ca-f9a4dfc6dfa7"). InnerVolumeSpecName "kube-api-access-zt6pw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.884055 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e19ef24d-2e7c-47a9-943c-a49f650df3a7-kube-api-access-q4vhz" (OuterVolumeSpecName: "kube-api-access-q4vhz") pod "e19ef24d-2e7c-47a9-943c-a49f650df3a7" (UID: "e19ef24d-2e7c-47a9-943c-a49f650df3a7"). InnerVolumeSpecName "kube-api-access-q4vhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.898666 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/833fd434-3bdc-4be7-be6f-f06915509129-kube-api-access-tl78n" (OuterVolumeSpecName: "kube-api-access-tl78n") pod "833fd434-3bdc-4be7-be6f-f06915509129" (UID: "833fd434-3bdc-4be7-be6f-f06915509129"). InnerVolumeSpecName "kube-api-access-tl78n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.982049 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4vhz\" (UniqueName: \"kubernetes.io/projected/e19ef24d-2e7c-47a9-943c-a49f650df3a7-kube-api-access-q4vhz\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.982085 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl78n\" (UniqueName: \"kubernetes.io/projected/833fd434-3bdc-4be7-be6f-f06915509129-kube-api-access-tl78n\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:36 crc kubenswrapper[4775]: I1002 02:01:36.982093 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zt6pw\" (UniqueName: \"kubernetes.io/projected/a120b1c5-156f-432c-96ca-f9a4dfc6dfa7-kube-api-access-zt6pw\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:37 crc kubenswrapper[4775]: I1002 02:01:37.390229 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-0729-account-create-v8vz9" Oct 02 02:01:37 crc kubenswrapper[4775]: I1002 02:01:37.394059 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-0729-account-create-v8vz9" event={"ID":"e19ef24d-2e7c-47a9-943c-a49f650df3a7","Type":"ContainerDied","Data":"6e3e3d0f497edd4f71d791cee09b0583a93344aeb9d2a618873a5c901104c27c"} Oct 02 02:01:37 crc kubenswrapper[4775]: I1002 02:01:37.394107 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e3e3d0f497edd4f71d791cee09b0583a93344aeb9d2a618873a5c901104c27c" Oct 02 02:01:37 crc kubenswrapper[4775]: I1002 02:01:37.402689 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51","Type":"ContainerStarted","Data":"63252743cedda5425e3430e58790235c337b97aaa0683e299068d0092eca5e7f"} Oct 02 02:01:37 crc kubenswrapper[4775]: I1002 02:01:37.415117 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f3f8-account-create-86qvz" event={"ID":"a120b1c5-156f-432c-96ca-f9a4dfc6dfa7","Type":"ContainerDied","Data":"aa2f27d228e26a006ad01a4dceff0bf93b6b28e03f034ad9b684c6df38e9734a"} Oct 02 02:01:37 crc kubenswrapper[4775]: I1002 02:01:37.415152 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa2f27d228e26a006ad01a4dceff0bf93b6b28e03f034ad9b684c6df38e9734a" Oct 02 02:01:37 crc kubenswrapper[4775]: I1002 02:01:37.415210 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f3f8-account-create-86qvz" Oct 02 02:01:37 crc kubenswrapper[4775]: I1002 02:01:37.432286 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1319e744-d454-4024-898e-ac338b1b8443","Type":"ContainerStarted","Data":"c315218740c2142dcb8618c469086e3dcefad74469b76f74a24edf1a27bfc666"} Oct 02 02:01:37 crc kubenswrapper[4775]: I1002 02:01:37.444318 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.444299479 podStartE2EDuration="3.444299479s" podCreationTimestamp="2025-10-02 02:01:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:01:37.440376609 +0000 UTC m=+1234.607120649" watchObservedRunningTime="2025-10-02 02:01:37.444299479 +0000 UTC m=+1234.611043519" Oct 02 02:01:37 crc kubenswrapper[4775]: I1002 02:01:37.445244 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-5350-account-create-lk56t" event={"ID":"833fd434-3bdc-4be7-be6f-f06915509129","Type":"ContainerDied","Data":"7437a851faf84e55c1c6a48acb6aa1ded3fde445e5563c74507480004c79db67"} Oct 02 02:01:37 crc kubenswrapper[4775]: I1002 02:01:37.445272 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7437a851faf84e55c1c6a48acb6aa1ded3fde445e5563c74507480004c79db67" Oct 02 02:01:37 crc kubenswrapper[4775]: I1002 02:01:37.445318 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-5350-account-create-lk56t" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.031745 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.204304 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.204351 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-combined-ca-bundle\") pod \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.204404 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-scripts\") pod \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.204439 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-httpd-run\") pod \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.204471 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-internal-tls-certs\") pod \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.204534 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hkdv\" (UniqueName: \"kubernetes.io/projected/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-kube-api-access-4hkdv\") pod \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.204550 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-config-data\") pod \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.204635 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-logs\") pod \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\" (UID: \"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08\") " Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.205293 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-logs" (OuterVolumeSpecName: "logs") pod "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" (UID: "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.206263 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" (UID: "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.208125 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-scripts" (OuterVolumeSpecName: "scripts") pod "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" (UID: "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.208322 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-kube-api-access-4hkdv" (OuterVolumeSpecName: "kube-api-access-4hkdv") pod "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" (UID: "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08"). InnerVolumeSpecName "kube-api-access-4hkdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.209111 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" (UID: "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.273828 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" (UID: "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.287780 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" (UID: "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.309312 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.309384 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.309397 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.309407 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.309416 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.309424 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.309461 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hkdv\" (UniqueName: \"kubernetes.io/projected/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-kube-api-access-4hkdv\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.312372 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-config-data" (OuterVolumeSpecName: "config-data") pod "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" (UID: "d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.365650 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.374075 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7jgdw"] Oct 02 02:01:38 crc kubenswrapper[4775]: E1002 02:01:38.374420 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a120b1c5-156f-432c-96ca-f9a4dfc6dfa7" containerName="mariadb-account-create" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.374432 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a120b1c5-156f-432c-96ca-f9a4dfc6dfa7" containerName="mariadb-account-create" Oct 02 02:01:38 crc kubenswrapper[4775]: E1002 02:01:38.374449 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" containerName="glance-httpd" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.374455 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" containerName="glance-httpd" Oct 02 02:01:38 crc kubenswrapper[4775]: E1002 02:01:38.374679 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e19ef24d-2e7c-47a9-943c-a49f650df3a7" containerName="mariadb-account-create" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.374686 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e19ef24d-2e7c-47a9-943c-a49f650df3a7" containerName="mariadb-account-create" Oct 02 02:01:38 crc kubenswrapper[4775]: E1002 02:01:38.374701 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" containerName="glance-log" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.374706 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" containerName="glance-log" Oct 02 02:01:38 crc kubenswrapper[4775]: E1002 02:01:38.374716 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="833fd434-3bdc-4be7-be6f-f06915509129" containerName="mariadb-account-create" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.374722 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="833fd434-3bdc-4be7-be6f-f06915509129" containerName="mariadb-account-create" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.374899 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e19ef24d-2e7c-47a9-943c-a49f650df3a7" containerName="mariadb-account-create" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.374919 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" containerName="glance-httpd" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.374930 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" containerName="glance-log" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.374941 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a120b1c5-156f-432c-96ca-f9a4dfc6dfa7" containerName="mariadb-account-create" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.375027 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="833fd434-3bdc-4be7-be6f-f06915509129" containerName="mariadb-account-create" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.375608 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.381551 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7jgdw"] Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.389239 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.389404 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.389846 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-lhw96" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.414001 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7jgdw\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.414058 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-scripts\") pod \"nova-cell0-conductor-db-sync-7jgdw\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.414085 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmll8\" (UniqueName: \"kubernetes.io/projected/1d054374-ba0f-41f4-8894-f1181d003bb3-kube-api-access-vmll8\") pod \"nova-cell0-conductor-db-sync-7jgdw\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.414152 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-config-data\") pod \"nova-cell0-conductor-db-sync-7jgdw\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.414223 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.414233 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.469205 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1319e744-d454-4024-898e-ac338b1b8443","Type":"ContainerStarted","Data":"0f915dee9ad3972ac1aaeee7f51be90d1210c168ef78f9afce2d8bf3c6fa6f2d"} Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.469245 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1319e744-d454-4024-898e-ac338b1b8443","Type":"ContainerStarted","Data":"c26f10b5e5a033ef447a32716656df3e4eb53bccf0a74b1bac910f359b58ff24"} Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.471029 4775 generic.go:334] "Generic (PLEG): container finished" podID="d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" containerID="4b96b0239299787f213eec6812a29703fc5113d42a71f339aa0bca3867dc4fa0" exitCode=0 Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.471843 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.472013 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08","Type":"ContainerDied","Data":"4b96b0239299787f213eec6812a29703fc5113d42a71f339aa0bca3867dc4fa0"} Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.472077 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08","Type":"ContainerDied","Data":"bec1ab733d5e9f88875433a033f9f18e650c465e9b2bf5fafa524188bbcf1c8d"} Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.472097 4775 scope.go:117] "RemoveContainer" containerID="4b96b0239299787f213eec6812a29703fc5113d42a71f339aa0bca3867dc4fa0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.515844 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-config-data\") pod \"nova-cell0-conductor-db-sync-7jgdw\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.515977 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7jgdw\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.516005 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-scripts\") pod \"nova-cell0-conductor-db-sync-7jgdw\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.516032 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmll8\" (UniqueName: \"kubernetes.io/projected/1d054374-ba0f-41f4-8894-f1181d003bb3-kube-api-access-vmll8\") pod \"nova-cell0-conductor-db-sync-7jgdw\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.519774 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-config-data\") pod \"nova-cell0-conductor-db-sync-7jgdw\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.521336 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-scripts\") pod \"nova-cell0-conductor-db-sync-7jgdw\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.528600 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-7jgdw\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.528997 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.537230 4775 scope.go:117] "RemoveContainer" containerID="229dd58f2fa05172f374734e3ca5ed77fc063711c568f4fe9c41b9a071b85798" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.537600 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmll8\" (UniqueName: \"kubernetes.io/projected/1d054374-ba0f-41f4-8894-f1181d003bb3-kube-api-access-vmll8\") pod \"nova-cell0-conductor-db-sync-7jgdw\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.542621 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.552002 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.553508 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.557037 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.557240 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.570945 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.579426 4775 scope.go:117] "RemoveContainer" containerID="4b96b0239299787f213eec6812a29703fc5113d42a71f339aa0bca3867dc4fa0" Oct 02 02:01:38 crc kubenswrapper[4775]: E1002 02:01:38.583357 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b96b0239299787f213eec6812a29703fc5113d42a71f339aa0bca3867dc4fa0\": container with ID starting with 4b96b0239299787f213eec6812a29703fc5113d42a71f339aa0bca3867dc4fa0 not found: ID does not exist" containerID="4b96b0239299787f213eec6812a29703fc5113d42a71f339aa0bca3867dc4fa0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.583396 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b96b0239299787f213eec6812a29703fc5113d42a71f339aa0bca3867dc4fa0"} err="failed to get container status \"4b96b0239299787f213eec6812a29703fc5113d42a71f339aa0bca3867dc4fa0\": rpc error: code = NotFound desc = could not find container \"4b96b0239299787f213eec6812a29703fc5113d42a71f339aa0bca3867dc4fa0\": container with ID starting with 4b96b0239299787f213eec6812a29703fc5113d42a71f339aa0bca3867dc4fa0 not found: ID does not exist" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.583421 4775 scope.go:117] "RemoveContainer" containerID="229dd58f2fa05172f374734e3ca5ed77fc063711c568f4fe9c41b9a071b85798" Oct 02 02:01:38 crc kubenswrapper[4775]: E1002 02:01:38.583913 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"229dd58f2fa05172f374734e3ca5ed77fc063711c568f4fe9c41b9a071b85798\": container with ID starting with 229dd58f2fa05172f374734e3ca5ed77fc063711c568f4fe9c41b9a071b85798 not found: ID does not exist" containerID="229dd58f2fa05172f374734e3ca5ed77fc063711c568f4fe9c41b9a071b85798" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.583963 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"229dd58f2fa05172f374734e3ca5ed77fc063711c568f4fe9c41b9a071b85798"} err="failed to get container status \"229dd58f2fa05172f374734e3ca5ed77fc063711c568f4fe9c41b9a071b85798\": rpc error: code = NotFound desc = could not find container \"229dd58f2fa05172f374734e3ca5ed77fc063711c568f4fe9c41b9a071b85798\": container with ID starting with 229dd58f2fa05172f374734e3ca5ed77fc063711c568f4fe9c41b9a071b85798 not found: ID does not exist" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.702468 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.719924 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.720100 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.720174 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.720243 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f286b47-509d-479d-bad5-0bbec930558a-logs\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.720281 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f286b47-509d-479d-bad5-0bbec930558a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.720332 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sv9w\" (UniqueName: \"kubernetes.io/projected/8f286b47-509d-479d-bad5-0bbec930558a-kube-api-access-5sv9w\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.720392 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.720440 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.845715 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.846016 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.846058 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.846083 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.846112 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f286b47-509d-479d-bad5-0bbec930558a-logs\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.846131 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f286b47-509d-479d-bad5-0bbec930558a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.847429 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sv9w\" (UniqueName: \"kubernetes.io/projected/8f286b47-509d-479d-bad5-0bbec930558a-kube-api-access-5sv9w\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.847498 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.848630 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.849498 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f286b47-509d-479d-bad5-0bbec930558a-logs\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.849839 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f286b47-509d-479d-bad5-0bbec930558a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.854909 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.855512 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.856292 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.857405 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.872279 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sv9w\" (UniqueName: \"kubernetes.io/projected/8f286b47-509d-479d-bad5-0bbec930558a-kube-api-access-5sv9w\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:38 crc kubenswrapper[4775]: I1002 02:01:38.885356 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " pod="openstack/glance-default-internal-api-0" Oct 02 02:01:39 crc kubenswrapper[4775]: I1002 02:01:39.171775 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7jgdw"] Oct 02 02:01:39 crc kubenswrapper[4775]: I1002 02:01:39.184478 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:01:39 crc kubenswrapper[4775]: I1002 02:01:39.482278 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7jgdw" event={"ID":"1d054374-ba0f-41f4-8894-f1181d003bb3","Type":"ContainerStarted","Data":"2de894361751beaf0ccb04d4e86754a11fa62fb10310da28c841bf7a12801d47"} Oct 02 02:01:39 crc kubenswrapper[4775]: I1002 02:01:39.709131 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:01:39 crc kubenswrapper[4775]: W1002 02:01:39.711210 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f286b47_509d_479d_bad5_0bbec930558a.slice/crio-a8f89d176e72a63eb5c4007babeb5a609d0afffddf3f01497b66de43cffb3e2c WatchSource:0}: Error finding container a8f89d176e72a63eb5c4007babeb5a609d0afffddf3f01497b66de43cffb3e2c: Status 404 returned error can't find the container with id a8f89d176e72a63eb5c4007babeb5a609d0afffddf3f01497b66de43cffb3e2c Oct 02 02:01:39 crc kubenswrapper[4775]: I1002 02:01:39.777931 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08" path="/var/lib/kubelet/pods/d6b84d82-4d4d-4f75-b6a4-3dc31ce6dc08/volumes" Oct 02 02:01:40 crc kubenswrapper[4775]: I1002 02:01:40.502785 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1319e744-d454-4024-898e-ac338b1b8443","Type":"ContainerStarted","Data":"e3133fad4451f93b08768783e43e7c4303dcbf17baae0f5714dfb7c95d316e71"} Oct 02 02:01:40 crc kubenswrapper[4775]: I1002 02:01:40.503229 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:01:40 crc kubenswrapper[4775]: I1002 02:01:40.513976 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f286b47-509d-479d-bad5-0bbec930558a","Type":"ContainerStarted","Data":"35fba6a1f824b8a5a30f3f0d8b52922737c0bdb3b40b6987fcc54b88ccfc67ab"} Oct 02 02:01:40 crc kubenswrapper[4775]: I1002 02:01:40.514026 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f286b47-509d-479d-bad5-0bbec930558a","Type":"ContainerStarted","Data":"a8f89d176e72a63eb5c4007babeb5a609d0afffddf3f01497b66de43cffb3e2c"} Oct 02 02:01:40 crc kubenswrapper[4775]: I1002 02:01:40.538209 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.320301785 podStartE2EDuration="5.538181117s" podCreationTimestamp="2025-10-02 02:01:35 +0000 UTC" firstStartedPulling="2025-10-02 02:01:36.235387179 +0000 UTC m=+1233.402131229" lastFinishedPulling="2025-10-02 02:01:39.453266501 +0000 UTC m=+1236.620010561" observedRunningTime="2025-10-02 02:01:40.535847148 +0000 UTC m=+1237.702591208" watchObservedRunningTime="2025-10-02 02:01:40.538181117 +0000 UTC m=+1237.704925167" Oct 02 02:01:41 crc kubenswrapper[4775]: I1002 02:01:41.525230 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f286b47-509d-479d-bad5-0bbec930558a","Type":"ContainerStarted","Data":"db8059deef67f936d22e07a37161aed4eda44bdc81845a2918822f9946fd9c00"} Oct 02 02:01:41 crc kubenswrapper[4775]: I1002 02:01:41.564294 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.564273788 podStartE2EDuration="3.564273788s" podCreationTimestamp="2025-10-02 02:01:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:01:41.556026369 +0000 UTC m=+1238.722770429" watchObservedRunningTime="2025-10-02 02:01:41.564273788 +0000 UTC m=+1238.731017838" Oct 02 02:01:45 crc kubenswrapper[4775]: I1002 02:01:45.007756 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 02:01:45 crc kubenswrapper[4775]: I1002 02:01:45.008031 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 02:01:45 crc kubenswrapper[4775]: I1002 02:01:45.037056 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 02:01:45 crc kubenswrapper[4775]: I1002 02:01:45.067451 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 02:01:45 crc kubenswrapper[4775]: I1002 02:01:45.573029 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 02:01:45 crc kubenswrapper[4775]: I1002 02:01:45.573467 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 02:01:46 crc kubenswrapper[4775]: I1002 02:01:46.590832 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7jgdw" event={"ID":"1d054374-ba0f-41f4-8894-f1181d003bb3","Type":"ContainerStarted","Data":"e16c239ca1126dd791cb5053ad8e44fb7f55ca4a02d7a26f73e67a82f58c593d"} Oct 02 02:01:46 crc kubenswrapper[4775]: I1002 02:01:46.637627 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-7jgdw" podStartSLOduration=1.7218744529999999 podStartE2EDuration="8.637601408s" podCreationTimestamp="2025-10-02 02:01:38 +0000 UTC" firstStartedPulling="2025-10-02 02:01:39.181372885 +0000 UTC m=+1236.348116925" lastFinishedPulling="2025-10-02 02:01:46.09709981 +0000 UTC m=+1243.263843880" observedRunningTime="2025-10-02 02:01:46.636194063 +0000 UTC m=+1243.802938103" watchObservedRunningTime="2025-10-02 02:01:46.637601408 +0000 UTC m=+1243.804345488" Oct 02 02:01:47 crc kubenswrapper[4775]: I1002 02:01:47.483912 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 02:01:47 crc kubenswrapper[4775]: I1002 02:01:47.488876 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 02:01:49 crc kubenswrapper[4775]: I1002 02:01:49.184905 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 02:01:49 crc kubenswrapper[4775]: I1002 02:01:49.185187 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 02:01:49 crc kubenswrapper[4775]: I1002 02:01:49.237118 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 02:01:49 crc kubenswrapper[4775]: I1002 02:01:49.250970 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 02:01:49 crc kubenswrapper[4775]: I1002 02:01:49.652048 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 02:01:49 crc kubenswrapper[4775]: I1002 02:01:49.652101 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 02:01:51 crc kubenswrapper[4775]: I1002 02:01:51.502669 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 02:01:51 crc kubenswrapper[4775]: I1002 02:01:51.507762 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 02:01:57 crc kubenswrapper[4775]: I1002 02:01:57.753230 4775 generic.go:334] "Generic (PLEG): container finished" podID="1d054374-ba0f-41f4-8894-f1181d003bb3" containerID="e16c239ca1126dd791cb5053ad8e44fb7f55ca4a02d7a26f73e67a82f58c593d" exitCode=0 Oct 02 02:01:57 crc kubenswrapper[4775]: I1002 02:01:57.753326 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7jgdw" event={"ID":"1d054374-ba0f-41f4-8894-f1181d003bb3","Type":"ContainerDied","Data":"e16c239ca1126dd791cb5053ad8e44fb7f55ca4a02d7a26f73e67a82f58c593d"} Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.162941 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.296796 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmll8\" (UniqueName: \"kubernetes.io/projected/1d054374-ba0f-41f4-8894-f1181d003bb3-kube-api-access-vmll8\") pod \"1d054374-ba0f-41f4-8894-f1181d003bb3\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.296859 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-combined-ca-bundle\") pod \"1d054374-ba0f-41f4-8894-f1181d003bb3\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.297076 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-scripts\") pod \"1d054374-ba0f-41f4-8894-f1181d003bb3\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.297180 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-config-data\") pod \"1d054374-ba0f-41f4-8894-f1181d003bb3\" (UID: \"1d054374-ba0f-41f4-8894-f1181d003bb3\") " Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.304521 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d054374-ba0f-41f4-8894-f1181d003bb3-kube-api-access-vmll8" (OuterVolumeSpecName: "kube-api-access-vmll8") pod "1d054374-ba0f-41f4-8894-f1181d003bb3" (UID: "1d054374-ba0f-41f4-8894-f1181d003bb3"). InnerVolumeSpecName "kube-api-access-vmll8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.304838 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-scripts" (OuterVolumeSpecName: "scripts") pod "1d054374-ba0f-41f4-8894-f1181d003bb3" (UID: "1d054374-ba0f-41f4-8894-f1181d003bb3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.324234 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d054374-ba0f-41f4-8894-f1181d003bb3" (UID: "1d054374-ba0f-41f4-8894-f1181d003bb3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.348179 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-config-data" (OuterVolumeSpecName: "config-data") pod "1d054374-ba0f-41f4-8894-f1181d003bb3" (UID: "1d054374-ba0f-41f4-8894-f1181d003bb3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.399811 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.399864 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmll8\" (UniqueName: \"kubernetes.io/projected/1d054374-ba0f-41f4-8894-f1181d003bb3-kube-api-access-vmll8\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.399887 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.399905 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d054374-ba0f-41f4-8894-f1181d003bb3-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.783185 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-7jgdw" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.799789 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-7jgdw" event={"ID":"1d054374-ba0f-41f4-8894-f1181d003bb3","Type":"ContainerDied","Data":"2de894361751beaf0ccb04d4e86754a11fa62fb10310da28c841bf7a12801d47"} Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.799852 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2de894361751beaf0ccb04d4e86754a11fa62fb10310da28c841bf7a12801d47" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.969270 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 02:01:59 crc kubenswrapper[4775]: E1002 02:01:59.969781 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d054374-ba0f-41f4-8894-f1181d003bb3" containerName="nova-cell0-conductor-db-sync" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.969805 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d054374-ba0f-41f4-8894-f1181d003bb3" containerName="nova-cell0-conductor-db-sync" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.970073 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d054374-ba0f-41f4-8894-f1181d003bb3" containerName="nova-cell0-conductor-db-sync" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.970846 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.973992 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-lhw96" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.976365 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 02:01:59 crc kubenswrapper[4775]: I1002 02:01:59.989027 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.016701 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74e0a1d-125d-43f7-9ba3-cea70453fc01-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.016931 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74e0a1d-125d-43f7-9ba3-cea70453fc01-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.017093 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvwrj\" (UniqueName: \"kubernetes.io/projected/c74e0a1d-125d-43f7-9ba3-cea70453fc01-kube-api-access-lvwrj\") pod \"nova-cell0-conductor-0\" (UID: \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.118377 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74e0a1d-125d-43f7-9ba3-cea70453fc01-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.118451 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvwrj\" (UniqueName: \"kubernetes.io/projected/c74e0a1d-125d-43f7-9ba3-cea70453fc01-kube-api-access-lvwrj\") pod \"nova-cell0-conductor-0\" (UID: \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.118517 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74e0a1d-125d-43f7-9ba3-cea70453fc01-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.122676 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74e0a1d-125d-43f7-9ba3-cea70453fc01-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.124022 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74e0a1d-125d-43f7-9ba3-cea70453fc01-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.135578 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvwrj\" (UniqueName: \"kubernetes.io/projected/c74e0a1d-125d-43f7-9ba3-cea70453fc01-kube-api-access-lvwrj\") pod \"nova-cell0-conductor-0\" (UID: \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\") " pod="openstack/nova-cell0-conductor-0" Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.304449 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.562367 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 02:02:00 crc kubenswrapper[4775]: W1002 02:02:00.564850 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc74e0a1d_125d_43f7_9ba3_cea70453fc01.slice/crio-436ed97e2972853f43721bb36b94dc770df3bfe0aed42fb78e37eb4ddd990d75 WatchSource:0}: Error finding container 436ed97e2972853f43721bb36b94dc770df3bfe0aed42fb78e37eb4ddd990d75: Status 404 returned error can't find the container with id 436ed97e2972853f43721bb36b94dc770df3bfe0aed42fb78e37eb4ddd990d75 Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.795048 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c74e0a1d-125d-43f7-9ba3-cea70453fc01","Type":"ContainerStarted","Data":"2cf05392cd7bff8e546c2fbad0965375d71482567f75af41133d05802bad7f02"} Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.795144 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c74e0a1d-125d-43f7-9ba3-cea70453fc01","Type":"ContainerStarted","Data":"436ed97e2972853f43721bb36b94dc770df3bfe0aed42fb78e37eb4ddd990d75"} Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.795237 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 02 02:02:00 crc kubenswrapper[4775]: I1002 02:02:00.823220 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.823200905 podStartE2EDuration="1.823200905s" podCreationTimestamp="2025-10-02 02:01:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:00.816653139 +0000 UTC m=+1257.983397219" watchObservedRunningTime="2025-10-02 02:02:00.823200905 +0000 UTC m=+1257.989944955" Oct 02 02:02:05 crc kubenswrapper[4775]: I1002 02:02:05.342757 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 02 02:02:05 crc kubenswrapper[4775]: I1002 02:02:05.744081 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 02:02:05 crc kubenswrapper[4775]: I1002 02:02:05.898249 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-fxjf5"] Oct 02 02:02:05 crc kubenswrapper[4775]: I1002 02:02:05.899634 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:05 crc kubenswrapper[4775]: I1002 02:02:05.903852 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 02 02:02:05 crc kubenswrapper[4775]: I1002 02:02:05.904872 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 02 02:02:05 crc kubenswrapper[4775]: I1002 02:02:05.916301 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fxjf5"] Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.052794 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fxjf5\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.052859 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-scripts\") pod \"nova-cell0-cell-mapping-fxjf5\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.052876 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-config-data\") pod \"nova-cell0-cell-mapping-fxjf5\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.052946 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zm2k\" (UniqueName: \"kubernetes.io/projected/777deed6-1940-4ae7-b4bb-e9f71fa4429e-kube-api-access-4zm2k\") pod \"nova-cell0-cell-mapping-fxjf5\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.072712 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.079582 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.085884 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.090962 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.155062 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-scripts\") pod \"nova-cell0-cell-mapping-fxjf5\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.155120 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-config-data\") pod \"nova-cell0-cell-mapping-fxjf5\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.155226 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zm2k\" (UniqueName: \"kubernetes.io/projected/777deed6-1940-4ae7-b4bb-e9f71fa4429e-kube-api-access-4zm2k\") pod \"nova-cell0-cell-mapping-fxjf5\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.155315 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fxjf5\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.162264 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-scripts\") pod \"nova-cell0-cell-mapping-fxjf5\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.163419 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-config-data\") pod \"nova-cell0-cell-mapping-fxjf5\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.167905 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.168717 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-fxjf5\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.169483 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.173685 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.177923 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.203556 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zm2k\" (UniqueName: \"kubernetes.io/projected/777deed6-1940-4ae7-b4bb-e9f71fa4429e-kube-api-access-4zm2k\") pod \"nova-cell0-cell-mapping-fxjf5\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.230680 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.231949 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.239197 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.240716 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.249430 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.256561 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-config-data\") pod \"nova-scheduler-0\" (UID: \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.256989 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htjr6\" (UniqueName: \"kubernetes.io/projected/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-kube-api-access-htjr6\") pod \"nova-scheduler-0\" (UID: \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.257129 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.298952 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.300496 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.313943 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.336834 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.358022 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf8nf\" (UniqueName: \"kubernetes.io/projected/20b54b85-0b61-4332-869c-1f2f8c07b4fb-kube-api-access-gf8nf\") pod \"nova-api-0\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " pod="openstack/nova-api-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.358072 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20b54b85-0b61-4332-869c-1f2f8c07b4fb-logs\") pod \"nova-api-0\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " pod="openstack/nova-api-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.358102 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.358137 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-695st\" (UniqueName: \"kubernetes.io/projected/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-kube-api-access-695st\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.358179 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.358208 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20b54b85-0b61-4332-869c-1f2f8c07b4fb-config-data\") pod \"nova-api-0\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " pod="openstack/nova-api-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.358237 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-config-data\") pod \"nova-scheduler-0\" (UID: \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.358264 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htjr6\" (UniqueName: \"kubernetes.io/projected/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-kube-api-access-htjr6\") pod \"nova-scheduler-0\" (UID: \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.358284 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20b54b85-0b61-4332-869c-1f2f8c07b4fb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " pod="openstack/nova-api-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.358316 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.363446 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.365994 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-config-data\") pod \"nova-scheduler-0\" (UID: \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.408056 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htjr6\" (UniqueName: \"kubernetes.io/projected/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-kube-api-access-htjr6\") pod \"nova-scheduler-0\" (UID: \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.438816 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-6qqkn"] Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.446490 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.469446 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-695st\" (UniqueName: \"kubernetes.io/projected/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-kube-api-access-695st\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.469684 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " pod="openstack/nova-metadata-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.469854 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.469993 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-config-data\") pod \"nova-metadata-0\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " pod="openstack/nova-metadata-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.470146 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20b54b85-0b61-4332-869c-1f2f8c07b4fb-config-data\") pod \"nova-api-0\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " pod="openstack/nova-api-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.470288 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxmks\" (UniqueName: \"kubernetes.io/projected/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-kube-api-access-lxmks\") pod \"nova-metadata-0\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " pod="openstack/nova-metadata-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.470451 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20b54b85-0b61-4332-869c-1f2f8c07b4fb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " pod="openstack/nova-api-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.470573 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-logs\") pod \"nova-metadata-0\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " pod="openstack/nova-metadata-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.470688 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf8nf\" (UniqueName: \"kubernetes.io/projected/20b54b85-0b61-4332-869c-1f2f8c07b4fb-kube-api-access-gf8nf\") pod \"nova-api-0\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " pod="openstack/nova-api-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.474481 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20b54b85-0b61-4332-869c-1f2f8c07b4fb-logs\") pod \"nova-api-0\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " pod="openstack/nova-api-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.474709 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.475351 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20b54b85-0b61-4332-869c-1f2f8c07b4fb-logs\") pod \"nova-api-0\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " pod="openstack/nova-api-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.478878 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20b54b85-0b61-4332-869c-1f2f8c07b4fb-config-data\") pod \"nova-api-0\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " pod="openstack/nova-api-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.479539 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.481633 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20b54b85-0b61-4332-869c-1f2f8c07b4fb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " pod="openstack/nova-api-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.481774 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.506684 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-6qqkn"] Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.511305 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-695st\" (UniqueName: \"kubernetes.io/projected/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-kube-api-access-695st\") pod \"nova-cell1-novncproxy-0\" (UID: \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.511359 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf8nf\" (UniqueName: \"kubernetes.io/projected/20b54b85-0b61-4332-869c-1f2f8c07b4fb-kube-api-access-gf8nf\") pod \"nova-api-0\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " pod="openstack/nova-api-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.562910 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.578500 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6r4k\" (UniqueName: \"kubernetes.io/projected/ebf491cd-4ab2-42a8-a53d-94193bfceb88-kube-api-access-f6r4k\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.578558 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-logs\") pod \"nova-metadata-0\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " pod="openstack/nova-metadata-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.578583 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-config\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.578603 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.578641 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.578684 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.578700 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " pod="openstack/nova-metadata-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.578724 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.578759 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-config-data\") pod \"nova-metadata-0\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " pod="openstack/nova-metadata-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.578785 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxmks\" (UniqueName: \"kubernetes.io/projected/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-kube-api-access-lxmks\") pod \"nova-metadata-0\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " pod="openstack/nova-metadata-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.579407 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-logs\") pod \"nova-metadata-0\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " pod="openstack/nova-metadata-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.583135 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-config-data\") pod \"nova-metadata-0\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " pod="openstack/nova-metadata-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.583362 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " pod="openstack/nova-metadata-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.596132 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxmks\" (UniqueName: \"kubernetes.io/projected/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-kube-api-access-lxmks\") pod \"nova-metadata-0\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " pod="openstack/nova-metadata-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.654641 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.679912 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.679974 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.680052 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6r4k\" (UniqueName: \"kubernetes.io/projected/ebf491cd-4ab2-42a8-a53d-94193bfceb88-kube-api-access-f6r4k\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.680083 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-config\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.680099 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.680136 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.680879 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.681506 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.681904 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-config\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.682054 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.682742 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.697383 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6r4k\" (UniqueName: \"kubernetes.io/projected/ebf491cd-4ab2-42a8-a53d-94193bfceb88-kube-api-access-f6r4k\") pod \"dnsmasq-dns-845d6d6f59-6qqkn\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.705924 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.734235 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.767457 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:06 crc kubenswrapper[4775]: I1002 02:02:06.921014 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-fxjf5"] Oct 02 02:02:06 crc kubenswrapper[4775]: W1002 02:02:06.927844 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod777deed6_1940_4ae7_b4bb_e9f71fa4429e.slice/crio-9cc81756615486c11400175a916740cc550e43fa60ff6361ba8b02fbf6e90b8f WatchSource:0}: Error finding container 9cc81756615486c11400175a916740cc550e43fa60ff6361ba8b02fbf6e90b8f: Status 404 returned error can't find the container with id 9cc81756615486c11400175a916740cc550e43fa60ff6361ba8b02fbf6e90b8f Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.001214 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7jsjh"] Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.002448 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.019194 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.019436 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.049070 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7jsjh"] Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.194919 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-config-data\") pod \"nova-cell1-conductor-db-sync-7jsjh\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.195026 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66g7q\" (UniqueName: \"kubernetes.io/projected/8ee94784-1d45-48fa-b7ad-81f12013c44c-kube-api-access-66g7q\") pod \"nova-cell1-conductor-db-sync-7jsjh\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.195049 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-scripts\") pod \"nova-cell1-conductor-db-sync-7jsjh\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.195079 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7jsjh\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.196685 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.302233 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-config-data\") pod \"nova-cell1-conductor-db-sync-7jsjh\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.302390 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66g7q\" (UniqueName: \"kubernetes.io/projected/8ee94784-1d45-48fa-b7ad-81f12013c44c-kube-api-access-66g7q\") pod \"nova-cell1-conductor-db-sync-7jsjh\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.302428 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-scripts\") pod \"nova-cell1-conductor-db-sync-7jsjh\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.312131 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7jsjh\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.322684 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-scripts\") pod \"nova-cell1-conductor-db-sync-7jsjh\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.326575 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-7jsjh\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.331771 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-config-data\") pod \"nova-cell1-conductor-db-sync-7jsjh\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.337529 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66g7q\" (UniqueName: \"kubernetes.io/projected/8ee94784-1d45-48fa-b7ad-81f12013c44c-kube-api-access-66g7q\") pod \"nova-cell1-conductor-db-sync-7jsjh\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.363229 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.371783 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.380327 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.381678 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.622449 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-6qqkn"] Oct 02 02:02:07 crc kubenswrapper[4775]: W1002 02:02:07.638005 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebf491cd_4ab2_42a8_a53d_94193bfceb88.slice/crio-4d0984404da4fcd2276b10a3b9ddba0f0e4b1e69eb142099539e0bad2cb966e6 WatchSource:0}: Error finding container 4d0984404da4fcd2276b10a3b9ddba0f0e4b1e69eb142099539e0bad2cb966e6: Status 404 returned error can't find the container with id 4d0984404da4fcd2276b10a3b9ddba0f0e4b1e69eb142099539e0bad2cb966e6 Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.864591 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7jsjh"] Oct 02 02:02:07 crc kubenswrapper[4775]: W1002 02:02:07.871442 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ee94784_1d45_48fa_b7ad_81f12013c44c.slice/crio-a4ce69e6d4e95ac5dc563807b38923e3f20371c303752e81cc7a46b7b73aa427 WatchSource:0}: Error finding container a4ce69e6d4e95ac5dc563807b38923e3f20371c303752e81cc7a46b7b73aa427: Status 404 returned error can't find the container with id a4ce69e6d4e95ac5dc563807b38923e3f20371c303752e81cc7a46b7b73aa427 Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.896071 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fxjf5" event={"ID":"777deed6-1940-4ae7-b4bb-e9f71fa4429e","Type":"ContainerStarted","Data":"6a762875446f3a628ec2e350627ea55effedef200ef0ae26856c73ea9d59025d"} Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.896120 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fxjf5" event={"ID":"777deed6-1940-4ae7-b4bb-e9f71fa4429e","Type":"ContainerStarted","Data":"9cc81756615486c11400175a916740cc550e43fa60ff6361ba8b02fbf6e90b8f"} Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.899403 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20b54b85-0b61-4332-869c-1f2f8c07b4fb","Type":"ContainerStarted","Data":"70a17366be9d6b080684e25d522b5c3ec4fc11b171629239a9475d3a19ba627c"} Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.900746 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a","Type":"ContainerStarted","Data":"da4ab5d083bd8f445a3ec0d0bc84aaaebe5e998b90fdbe9008988e92bf637884"} Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.901818 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c19ba10d-7dfd-4022-b29e-59e7f7d577d9","Type":"ContainerStarted","Data":"5ac7cd69e7470d62c860047c0f468418a2432cacf48f81204a83fbc300d75d3d"} Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.906865 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3","Type":"ContainerStarted","Data":"1102614fc06f60a9344f26ba887143892fafdd9ed2c392fb196d172ff7753009"} Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.921019 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-fxjf5" podStartSLOduration=2.921004641 podStartE2EDuration="2.921004641s" podCreationTimestamp="2025-10-02 02:02:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:07.911655913 +0000 UTC m=+1265.078399953" watchObservedRunningTime="2025-10-02 02:02:07.921004641 +0000 UTC m=+1265.087748681" Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.922545 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" event={"ID":"ebf491cd-4ab2-42a8-a53d-94193bfceb88","Type":"ContainerStarted","Data":"4d0984404da4fcd2276b10a3b9ddba0f0e4b1e69eb142099539e0bad2cb966e6"} Oct 02 02:02:07 crc kubenswrapper[4775]: I1002 02:02:07.932114 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7jsjh" event={"ID":"8ee94784-1d45-48fa-b7ad-81f12013c44c","Type":"ContainerStarted","Data":"a4ce69e6d4e95ac5dc563807b38923e3f20371c303752e81cc7a46b7b73aa427"} Oct 02 02:02:08 crc kubenswrapper[4775]: I1002 02:02:08.948150 4775 generic.go:334] "Generic (PLEG): container finished" podID="ebf491cd-4ab2-42a8-a53d-94193bfceb88" containerID="a3f2f5fb1878415b74883963e59abf3f91c197081a8aec84bd142cc8c640ad9d" exitCode=0 Oct 02 02:02:08 crc kubenswrapper[4775]: I1002 02:02:08.948707 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" event={"ID":"ebf491cd-4ab2-42a8-a53d-94193bfceb88","Type":"ContainerDied","Data":"a3f2f5fb1878415b74883963e59abf3f91c197081a8aec84bd142cc8c640ad9d"} Oct 02 02:02:08 crc kubenswrapper[4775]: I1002 02:02:08.960058 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7jsjh" event={"ID":"8ee94784-1d45-48fa-b7ad-81f12013c44c","Type":"ContainerStarted","Data":"95e4dfafbb440c4413dfe2954b28be460f01b7d55d95e1321662f71a709463a5"} Oct 02 02:02:09 crc kubenswrapper[4775]: I1002 02:02:09.656410 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-7jsjh" podStartSLOduration=3.656394463 podStartE2EDuration="3.656394463s" podCreationTimestamp="2025-10-02 02:02:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:08.997361749 +0000 UTC m=+1266.164105789" watchObservedRunningTime="2025-10-02 02:02:09.656394463 +0000 UTC m=+1266.823138503" Oct 02 02:02:09 crc kubenswrapper[4775]: I1002 02:02:09.662423 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:09 crc kubenswrapper[4775]: I1002 02:02:09.676769 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:02:11 crc kubenswrapper[4775]: I1002 02:02:11.988823 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a","Type":"ContainerStarted","Data":"6551e9c5f62eb1393127c59a3b81e28867173c615ce4b3cc689c3b24c62e2044"} Oct 02 02:02:11 crc kubenswrapper[4775]: I1002 02:02:11.988918 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="8e6e2e8f-33ab-4eea-9c51-471533aa9f4a" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://6551e9c5f62eb1393127c59a3b81e28867173c615ce4b3cc689c3b24c62e2044" gracePeriod=30 Oct 02 02:02:11 crc kubenswrapper[4775]: I1002 02:02:11.991286 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c19ba10d-7dfd-4022-b29e-59e7f7d577d9","Type":"ContainerStarted","Data":"ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c"} Oct 02 02:02:11 crc kubenswrapper[4775]: I1002 02:02:11.991329 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c19ba10d-7dfd-4022-b29e-59e7f7d577d9","Type":"ContainerStarted","Data":"90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f"} Oct 02 02:02:11 crc kubenswrapper[4775]: I1002 02:02:11.991378 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c19ba10d-7dfd-4022-b29e-59e7f7d577d9" containerName="nova-metadata-log" containerID="cri-o://90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f" gracePeriod=30 Oct 02 02:02:11 crc kubenswrapper[4775]: I1002 02:02:11.991418 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="c19ba10d-7dfd-4022-b29e-59e7f7d577d9" containerName="nova-metadata-metadata" containerID="cri-o://ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c" gracePeriod=30 Oct 02 02:02:11 crc kubenswrapper[4775]: I1002 02:02:11.994590 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3","Type":"ContainerStarted","Data":"c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b"} Oct 02 02:02:11 crc kubenswrapper[4775]: I1002 02:02:11.997829 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" event={"ID":"ebf491cd-4ab2-42a8-a53d-94193bfceb88","Type":"ContainerStarted","Data":"00b230cb5807d0b8515a0dd77cc58a129ee4a0f812e8343c5cd23ba4ec66cfc9"} Oct 02 02:02:11 crc kubenswrapper[4775]: I1002 02:02:11.998401 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.000674 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20b54b85-0b61-4332-869c-1f2f8c07b4fb","Type":"ContainerStarted","Data":"9c58be88cd4130340b3ca2a4f3bc647e10ec86b0ed1878d66aeab0f478633046"} Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.000815 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20b54b85-0b61-4332-869c-1f2f8c07b4fb","Type":"ContainerStarted","Data":"58c0cbe1782d0bf708643480ecd6e72fcb640e1aab07bded85208680637d966d"} Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.011265 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.891896218 podStartE2EDuration="6.011244974s" podCreationTimestamp="2025-10-02 02:02:06 +0000 UTC" firstStartedPulling="2025-10-02 02:02:07.408503604 +0000 UTC m=+1264.575247644" lastFinishedPulling="2025-10-02 02:02:10.52785236 +0000 UTC m=+1267.694596400" observedRunningTime="2025-10-02 02:02:12.00718243 +0000 UTC m=+1269.173926470" watchObservedRunningTime="2025-10-02 02:02:12.011244974 +0000 UTC m=+1269.177989024" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.027167 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.746269263 podStartE2EDuration="6.027143788s" podCreationTimestamp="2025-10-02 02:02:06 +0000 UTC" firstStartedPulling="2025-10-02 02:02:07.2471444 +0000 UTC m=+1264.413888440" lastFinishedPulling="2025-10-02 02:02:10.528018925 +0000 UTC m=+1267.694762965" observedRunningTime="2025-10-02 02:02:12.026541053 +0000 UTC m=+1269.193285093" watchObservedRunningTime="2025-10-02 02:02:12.027143788 +0000 UTC m=+1269.193887838" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.044643 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.924452216 podStartE2EDuration="6.044624973s" podCreationTimestamp="2025-10-02 02:02:06 +0000 UTC" firstStartedPulling="2025-10-02 02:02:07.408759081 +0000 UTC m=+1264.575503121" lastFinishedPulling="2025-10-02 02:02:10.528931838 +0000 UTC m=+1267.695675878" observedRunningTime="2025-10-02 02:02:12.042008846 +0000 UTC m=+1269.208752916" watchObservedRunningTime="2025-10-02 02:02:12.044624973 +0000 UTC m=+1269.211369033" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.065938 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.9100615 podStartE2EDuration="6.065924665s" podCreationTimestamp="2025-10-02 02:02:06 +0000 UTC" firstStartedPulling="2025-10-02 02:02:07.379221449 +0000 UTC m=+1264.545965489" lastFinishedPulling="2025-10-02 02:02:10.535084584 +0000 UTC m=+1267.701828654" observedRunningTime="2025-10-02 02:02:12.06219748 +0000 UTC m=+1269.228941520" watchObservedRunningTime="2025-10-02 02:02:12.065924665 +0000 UTC m=+1269.232668705" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.098561 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" podStartSLOduration=6.098543924 podStartE2EDuration="6.098543924s" podCreationTimestamp="2025-10-02 02:02:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:12.095037705 +0000 UTC m=+1269.261781735" watchObservedRunningTime="2025-10-02 02:02:12.098543924 +0000 UTC m=+1269.265287954" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.145638 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.145829 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="6f55acc8-9fa8-496b-a882-d86ea3f28730" containerName="kube-state-metrics" containerID="cri-o://991d076b8c7fdaa82481667a7c1adc27601fb8bff13b965c51aee091cf3511ff" gracePeriod=30 Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.610052 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.633527 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-combined-ca-bundle\") pod \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.633619 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-config-data\") pod \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.633644 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxmks\" (UniqueName: \"kubernetes.io/projected/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-kube-api-access-lxmks\") pod \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.633762 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-logs\") pod \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\" (UID: \"c19ba10d-7dfd-4022-b29e-59e7f7d577d9\") " Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.634409 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-logs" (OuterVolumeSpecName: "logs") pod "c19ba10d-7dfd-4022-b29e-59e7f7d577d9" (UID: "c19ba10d-7dfd-4022-b29e-59e7f7d577d9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.639385 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-kube-api-access-lxmks" (OuterVolumeSpecName: "kube-api-access-lxmks") pod "c19ba10d-7dfd-4022-b29e-59e7f7d577d9" (UID: "c19ba10d-7dfd-4022-b29e-59e7f7d577d9"). InnerVolumeSpecName "kube-api-access-lxmks". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.664167 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-config-data" (OuterVolumeSpecName: "config-data") pod "c19ba10d-7dfd-4022-b29e-59e7f7d577d9" (UID: "c19ba10d-7dfd-4022-b29e-59e7f7d577d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.677114 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c19ba10d-7dfd-4022-b29e-59e7f7d577d9" (UID: "c19ba10d-7dfd-4022-b29e-59e7f7d577d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.736014 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.736047 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.736076 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxmks\" (UniqueName: \"kubernetes.io/projected/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-kube-api-access-lxmks\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.736089 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c19ba10d-7dfd-4022-b29e-59e7f7d577d9-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.765640 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.839743 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw9m9\" (UniqueName: \"kubernetes.io/projected/6f55acc8-9fa8-496b-a882-d86ea3f28730-kube-api-access-fw9m9\") pod \"6f55acc8-9fa8-496b-a882-d86ea3f28730\" (UID: \"6f55acc8-9fa8-496b-a882-d86ea3f28730\") " Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.844914 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f55acc8-9fa8-496b-a882-d86ea3f28730-kube-api-access-fw9m9" (OuterVolumeSpecName: "kube-api-access-fw9m9") pod "6f55acc8-9fa8-496b-a882-d86ea3f28730" (UID: "6f55acc8-9fa8-496b-a882-d86ea3f28730"). InnerVolumeSpecName "kube-api-access-fw9m9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:12 crc kubenswrapper[4775]: I1002 02:02:12.942682 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw9m9\" (UniqueName: \"kubernetes.io/projected/6f55acc8-9fa8-496b-a882-d86ea3f28730-kube-api-access-fw9m9\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.010794 4775 generic.go:334] "Generic (PLEG): container finished" podID="6f55acc8-9fa8-496b-a882-d86ea3f28730" containerID="991d076b8c7fdaa82481667a7c1adc27601fb8bff13b965c51aee091cf3511ff" exitCode=2 Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.010867 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6f55acc8-9fa8-496b-a882-d86ea3f28730","Type":"ContainerDied","Data":"991d076b8c7fdaa82481667a7c1adc27601fb8bff13b965c51aee091cf3511ff"} Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.010893 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"6f55acc8-9fa8-496b-a882-d86ea3f28730","Type":"ContainerDied","Data":"a7952639931629a679e5654b1879e61531b64cacd1cb10b624464e76d444153e"} Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.010909 4775 scope.go:117] "RemoveContainer" containerID="991d076b8c7fdaa82481667a7c1adc27601fb8bff13b965c51aee091cf3511ff" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.011019 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.021557 4775 generic.go:334] "Generic (PLEG): container finished" podID="c19ba10d-7dfd-4022-b29e-59e7f7d577d9" containerID="ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c" exitCode=0 Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.021588 4775 generic.go:334] "Generic (PLEG): container finished" podID="c19ba10d-7dfd-4022-b29e-59e7f7d577d9" containerID="90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f" exitCode=143 Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.022487 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.023946 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c19ba10d-7dfd-4022-b29e-59e7f7d577d9","Type":"ContainerDied","Data":"ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c"} Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.024044 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c19ba10d-7dfd-4022-b29e-59e7f7d577d9","Type":"ContainerDied","Data":"90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f"} Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.024065 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c19ba10d-7dfd-4022-b29e-59e7f7d577d9","Type":"ContainerDied","Data":"5ac7cd69e7470d62c860047c0f468418a2432cacf48f81204a83fbc300d75d3d"} Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.049535 4775 scope.go:117] "RemoveContainer" containerID="991d076b8c7fdaa82481667a7c1adc27601fb8bff13b965c51aee091cf3511ff" Oct 02 02:02:13 crc kubenswrapper[4775]: E1002 02:02:13.050175 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"991d076b8c7fdaa82481667a7c1adc27601fb8bff13b965c51aee091cf3511ff\": container with ID starting with 991d076b8c7fdaa82481667a7c1adc27601fb8bff13b965c51aee091cf3511ff not found: ID does not exist" containerID="991d076b8c7fdaa82481667a7c1adc27601fb8bff13b965c51aee091cf3511ff" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.050287 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"991d076b8c7fdaa82481667a7c1adc27601fb8bff13b965c51aee091cf3511ff"} err="failed to get container status \"991d076b8c7fdaa82481667a7c1adc27601fb8bff13b965c51aee091cf3511ff\": rpc error: code = NotFound desc = could not find container \"991d076b8c7fdaa82481667a7c1adc27601fb8bff13b965c51aee091cf3511ff\": container with ID starting with 991d076b8c7fdaa82481667a7c1adc27601fb8bff13b965c51aee091cf3511ff not found: ID does not exist" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.050404 4775 scope.go:117] "RemoveContainer" containerID="ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.072218 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.089563 4775 scope.go:117] "RemoveContainer" containerID="90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.104821 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.119847 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:02:13 crc kubenswrapper[4775]: E1002 02:02:13.120254 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f55acc8-9fa8-496b-a882-d86ea3f28730" containerName="kube-state-metrics" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.120266 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f55acc8-9fa8-496b-a882-d86ea3f28730" containerName="kube-state-metrics" Oct 02 02:02:13 crc kubenswrapper[4775]: E1002 02:02:13.120286 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19ba10d-7dfd-4022-b29e-59e7f7d577d9" containerName="nova-metadata-log" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.120293 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19ba10d-7dfd-4022-b29e-59e7f7d577d9" containerName="nova-metadata-log" Oct 02 02:02:13 crc kubenswrapper[4775]: E1002 02:02:13.120316 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19ba10d-7dfd-4022-b29e-59e7f7d577d9" containerName="nova-metadata-metadata" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.120322 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19ba10d-7dfd-4022-b29e-59e7f7d577d9" containerName="nova-metadata-metadata" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.120506 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c19ba10d-7dfd-4022-b29e-59e7f7d577d9" containerName="nova-metadata-log" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.120519 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f55acc8-9fa8-496b-a882-d86ea3f28730" containerName="kube-state-metrics" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.120542 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c19ba10d-7dfd-4022-b29e-59e7f7d577d9" containerName="nova-metadata-metadata" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.121189 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.125942 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.126141 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.130073 4775 scope.go:117] "RemoveContainer" containerID="ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c" Oct 02 02:02:13 crc kubenswrapper[4775]: E1002 02:02:13.131198 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c\": container with ID starting with ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c not found: ID does not exist" containerID="ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.131234 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c"} err="failed to get container status \"ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c\": rpc error: code = NotFound desc = could not find container \"ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c\": container with ID starting with ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c not found: ID does not exist" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.131261 4775 scope.go:117] "RemoveContainer" containerID="90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f" Oct 02 02:02:13 crc kubenswrapper[4775]: E1002 02:02:13.131671 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f\": container with ID starting with 90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f not found: ID does not exist" containerID="90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.131688 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f"} err="failed to get container status \"90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f\": rpc error: code = NotFound desc = could not find container \"90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f\": container with ID starting with 90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f not found: ID does not exist" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.131700 4775 scope.go:117] "RemoveContainer" containerID="ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.132055 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c"} err="failed to get container status \"ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c\": rpc error: code = NotFound desc = could not find container \"ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c\": container with ID starting with ff6ac027004e9265cb3fd97039da1fafe81d0dbf23117bfa4a6d88a350483a3c not found: ID does not exist" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.132093 4775 scope.go:117] "RemoveContainer" containerID="90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.132327 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f"} err="failed to get container status \"90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f\": rpc error: code = NotFound desc = could not find container \"90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f\": container with ID starting with 90497801e16eb31684d57fc15155003d6788c5c3e8fa548dcfaf089f51af376f not found: ID does not exist" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.133906 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.146194 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.149833 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.149943 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr8rv\" (UniqueName: \"kubernetes.io/projected/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-api-access-pr8rv\") pod \"kube-state-metrics-0\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.150110 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.150155 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.155201 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.167288 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.169375 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.170674 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.170731 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.178244 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.252982 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.253400 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.253513 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-logs\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.253663 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7wwl\" (UniqueName: \"kubernetes.io/projected/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-kube-api-access-k7wwl\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.253791 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.253877 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.253981 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-config-data\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.254102 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.254274 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr8rv\" (UniqueName: \"kubernetes.io/projected/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-api-access-pr8rv\") pod \"kube-state-metrics-0\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.259274 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.272538 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.275641 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.276784 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr8rv\" (UniqueName: \"kubernetes.io/projected/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-api-access-pr8rv\") pod \"kube-state-metrics-0\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.356355 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-logs\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.356425 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7wwl\" (UniqueName: \"kubernetes.io/projected/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-kube-api-access-k7wwl\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.356473 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.356497 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-config-data\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.356535 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.356906 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-logs\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.360589 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.360843 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.360986 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-config-data\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.374004 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7wwl\" (UniqueName: \"kubernetes.io/projected/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-kube-api-access-k7wwl\") pod \"nova-metadata-0\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.440551 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.482934 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.778626 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f55acc8-9fa8-496b-a882-d86ea3f28730" path="/var/lib/kubelet/pods/6f55acc8-9fa8-496b-a882-d86ea3f28730/volumes" Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.779486 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c19ba10d-7dfd-4022-b29e-59e7f7d577d9" path="/var/lib/kubelet/pods/c19ba10d-7dfd-4022-b29e-59e7f7d577d9/volumes" Oct 02 02:02:13 crc kubenswrapper[4775]: W1002 02:02:13.935377 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24ef8940_eb43_459c_9cfe_854df57cfe74.slice/crio-4588a06ce5aa5eae7ebbeddab0850d8aaef02484e59ea3376959ccc6a6facdbe WatchSource:0}: Error finding container 4588a06ce5aa5eae7ebbeddab0850d8aaef02484e59ea3376959ccc6a6facdbe: Status 404 returned error can't find the container with id 4588a06ce5aa5eae7ebbeddab0850d8aaef02484e59ea3376959ccc6a6facdbe Oct 02 02:02:13 crc kubenswrapper[4775]: I1002 02:02:13.937855 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:02:14 crc kubenswrapper[4775]: I1002 02:02:14.036092 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"24ef8940-eb43-459c-9cfe-854df57cfe74","Type":"ContainerStarted","Data":"4588a06ce5aa5eae7ebbeddab0850d8aaef02484e59ea3376959ccc6a6facdbe"} Oct 02 02:02:14 crc kubenswrapper[4775]: I1002 02:02:14.054371 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:14 crc kubenswrapper[4775]: W1002 02:02:14.055860 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0fa1011f_07f0_4a21_9d46_00ba9b51c89d.slice/crio-96dc9ec0c6196eeaf2a6da8dda6877d49aee8cb78d3d5083ac42600044dd348c WatchSource:0}: Error finding container 96dc9ec0c6196eeaf2a6da8dda6877d49aee8cb78d3d5083ac42600044dd348c: Status 404 returned error can't find the container with id 96dc9ec0c6196eeaf2a6da8dda6877d49aee8cb78d3d5083ac42600044dd348c Oct 02 02:02:14 crc kubenswrapper[4775]: I1002 02:02:14.145619 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:02:14 crc kubenswrapper[4775]: I1002 02:02:14.146280 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="ceilometer-central-agent" containerID="cri-o://c315218740c2142dcb8618c469086e3dcefad74469b76f74a24edf1a27bfc666" gracePeriod=30 Oct 02 02:02:14 crc kubenswrapper[4775]: I1002 02:02:14.146329 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="proxy-httpd" containerID="cri-o://e3133fad4451f93b08768783e43e7c4303dcbf17baae0f5714dfb7c95d316e71" gracePeriod=30 Oct 02 02:02:14 crc kubenswrapper[4775]: I1002 02:02:14.146373 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="ceilometer-notification-agent" containerID="cri-o://c26f10b5e5a033ef447a32716656df3e4eb53bccf0a74b1bac910f359b58ff24" gracePeriod=30 Oct 02 02:02:14 crc kubenswrapper[4775]: I1002 02:02:14.146388 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="sg-core" containerID="cri-o://0f915dee9ad3972ac1aaeee7f51be90d1210c168ef78f9afce2d8bf3c6fa6f2d" gracePeriod=30 Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.047557 4775 generic.go:334] "Generic (PLEG): container finished" podID="777deed6-1940-4ae7-b4bb-e9f71fa4429e" containerID="6a762875446f3a628ec2e350627ea55effedef200ef0ae26856c73ea9d59025d" exitCode=0 Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.047656 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fxjf5" event={"ID":"777deed6-1940-4ae7-b4bb-e9f71fa4429e","Type":"ContainerDied","Data":"6a762875446f3a628ec2e350627ea55effedef200ef0ae26856c73ea9d59025d"} Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.050664 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fa1011f-07f0-4a21-9d46-00ba9b51c89d","Type":"ContainerStarted","Data":"fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702"} Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.050704 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fa1011f-07f0-4a21-9d46-00ba9b51c89d","Type":"ContainerStarted","Data":"21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f"} Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.050719 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fa1011f-07f0-4a21-9d46-00ba9b51c89d","Type":"ContainerStarted","Data":"96dc9ec0c6196eeaf2a6da8dda6877d49aee8cb78d3d5083ac42600044dd348c"} Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.055785 4775 generic.go:334] "Generic (PLEG): container finished" podID="1319e744-d454-4024-898e-ac338b1b8443" containerID="e3133fad4451f93b08768783e43e7c4303dcbf17baae0f5714dfb7c95d316e71" exitCode=0 Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.055818 4775 generic.go:334] "Generic (PLEG): container finished" podID="1319e744-d454-4024-898e-ac338b1b8443" containerID="0f915dee9ad3972ac1aaeee7f51be90d1210c168ef78f9afce2d8bf3c6fa6f2d" exitCode=2 Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.055826 4775 generic.go:334] "Generic (PLEG): container finished" podID="1319e744-d454-4024-898e-ac338b1b8443" containerID="c315218740c2142dcb8618c469086e3dcefad74469b76f74a24edf1a27bfc666" exitCode=0 Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.055874 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1319e744-d454-4024-898e-ac338b1b8443","Type":"ContainerDied","Data":"e3133fad4451f93b08768783e43e7c4303dcbf17baae0f5714dfb7c95d316e71"} Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.055928 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1319e744-d454-4024-898e-ac338b1b8443","Type":"ContainerDied","Data":"0f915dee9ad3972ac1aaeee7f51be90d1210c168ef78f9afce2d8bf3c6fa6f2d"} Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.055939 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1319e744-d454-4024-898e-ac338b1b8443","Type":"ContainerDied","Data":"c315218740c2142dcb8618c469086e3dcefad74469b76f74a24edf1a27bfc666"} Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.058558 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"24ef8940-eb43-459c-9cfe-854df57cfe74","Type":"ContainerStarted","Data":"7faaf4aaa749333ce8970b0b3bfce0b23a77558417d2d73450a187551b1a850f"} Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.058804 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.093275 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.09325612 podStartE2EDuration="2.09325612s" podCreationTimestamp="2025-10-02 02:02:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:15.084099368 +0000 UTC m=+1272.250843428" watchObservedRunningTime="2025-10-02 02:02:15.09325612 +0000 UTC m=+1272.260000160" Oct 02 02:02:15 crc kubenswrapper[4775]: I1002 02:02:15.113908 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.764818695 podStartE2EDuration="2.113890305s" podCreationTimestamp="2025-10-02 02:02:13 +0000 UTC" firstStartedPulling="2025-10-02 02:02:13.938343162 +0000 UTC m=+1271.105087202" lastFinishedPulling="2025-10-02 02:02:14.287414772 +0000 UTC m=+1271.454158812" observedRunningTime="2025-10-02 02:02:15.110097349 +0000 UTC m=+1272.276841399" watchObservedRunningTime="2025-10-02 02:02:15.113890305 +0000 UTC m=+1272.280634345" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.068445 4775 generic.go:334] "Generic (PLEG): container finished" podID="8ee94784-1d45-48fa-b7ad-81f12013c44c" containerID="95e4dfafbb440c4413dfe2954b28be460f01b7d55d95e1321662f71a709463a5" exitCode=0 Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.068811 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7jsjh" event={"ID":"8ee94784-1d45-48fa-b7ad-81f12013c44c","Type":"ContainerDied","Data":"95e4dfafbb440c4413dfe2954b28be460f01b7d55d95e1321662f71a709463a5"} Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.442218 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.533584 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-combined-ca-bundle\") pod \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.533659 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-config-data\") pod \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.533828 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-scripts\") pod \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.533976 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zm2k\" (UniqueName: \"kubernetes.io/projected/777deed6-1940-4ae7-b4bb-e9f71fa4429e-kube-api-access-4zm2k\") pod \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\" (UID: \"777deed6-1940-4ae7-b4bb-e9f71fa4429e\") " Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.539100 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-scripts" (OuterVolumeSpecName: "scripts") pod "777deed6-1940-4ae7-b4bb-e9f71fa4429e" (UID: "777deed6-1940-4ae7-b4bb-e9f71fa4429e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.539609 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/777deed6-1940-4ae7-b4bb-e9f71fa4429e-kube-api-access-4zm2k" (OuterVolumeSpecName: "kube-api-access-4zm2k") pod "777deed6-1940-4ae7-b4bb-e9f71fa4429e" (UID: "777deed6-1940-4ae7-b4bb-e9f71fa4429e"). InnerVolumeSpecName "kube-api-access-4zm2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.563491 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.563559 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.575551 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "777deed6-1940-4ae7-b4bb-e9f71fa4429e" (UID: "777deed6-1940-4ae7-b4bb-e9f71fa4429e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.578175 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-config-data" (OuterVolumeSpecName: "config-data") pod "777deed6-1940-4ae7-b4bb-e9f71fa4429e" (UID: "777deed6-1940-4ae7-b4bb-e9f71fa4429e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.636787 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.636818 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zm2k\" (UniqueName: \"kubernetes.io/projected/777deed6-1940-4ae7-b4bb-e9f71fa4429e-kube-api-access-4zm2k\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.636828 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.636839 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/777deed6-1940-4ae7-b4bb-e9f71fa4429e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.656008 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.706987 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.707197 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.743858 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.769202 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.880784 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-rf268"] Oct 02 02:02:16 crc kubenswrapper[4775]: I1002 02:02:16.881226 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-rf268" podUID="9aa78020-05ca-40e3-abd3-d7a960b8b2ad" containerName="dnsmasq-dns" containerID="cri-o://c7e1a112ab7147ecb551a7b9f6aef7efdc4bb4197ccd00766ebfc02583672984" gracePeriod=10 Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.080815 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-fxjf5" event={"ID":"777deed6-1940-4ae7-b4bb-e9f71fa4429e","Type":"ContainerDied","Data":"9cc81756615486c11400175a916740cc550e43fa60ff6361ba8b02fbf6e90b8f"} Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.080875 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cc81756615486c11400175a916740cc550e43fa60ff6361ba8b02fbf6e90b8f" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.080845 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-fxjf5" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.087182 4775 generic.go:334] "Generic (PLEG): container finished" podID="9aa78020-05ca-40e3-abd3-d7a960b8b2ad" containerID="c7e1a112ab7147ecb551a7b9f6aef7efdc4bb4197ccd00766ebfc02583672984" exitCode=0 Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.087602 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-rf268" event={"ID":"9aa78020-05ca-40e3-abd3-d7a960b8b2ad","Type":"ContainerDied","Data":"c7e1a112ab7147ecb551a7b9f6aef7efdc4bb4197ccd00766ebfc02583672984"} Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.129161 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.248946 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.249174 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="20b54b85-0b61-4332-869c-1f2f8c07b4fb" containerName="nova-api-log" containerID="cri-o://58c0cbe1782d0bf708643480ecd6e72fcb640e1aab07bded85208680637d966d" gracePeriod=30 Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.249250 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="20b54b85-0b61-4332-869c-1f2f8c07b4fb" containerName="nova-api-api" containerID="cri-o://9c58be88cd4130340b3ca2a4f3bc647e10ec86b0ed1878d66aeab0f478633046" gracePeriod=30 Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.261175 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="20b54b85-0b61-4332-869c-1f2f8c07b4fb" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": EOF" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.261274 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="20b54b85-0b61-4332-869c-1f2f8c07b4fb" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": EOF" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.267829 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.268054 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0fa1011f-07f0-4a21-9d46-00ba9b51c89d" containerName="nova-metadata-log" containerID="cri-o://21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f" gracePeriod=30 Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.268234 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0fa1011f-07f0-4a21-9d46-00ba9b51c89d" containerName="nova-metadata-metadata" containerID="cri-o://fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702" gracePeriod=30 Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.322872 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.361805 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hzwf\" (UniqueName: \"kubernetes.io/projected/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-kube-api-access-6hzwf\") pod \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.361984 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-dns-svc\") pod \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.362102 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-ovsdbserver-sb\") pod \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.362138 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-ovsdbserver-nb\") pod \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.362163 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-dns-swift-storage-0\") pod \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.362276 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-config\") pod \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\" (UID: \"9aa78020-05ca-40e3-abd3-d7a960b8b2ad\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.373143 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-kube-api-access-6hzwf" (OuterVolumeSpecName: "kube-api-access-6hzwf") pod "9aa78020-05ca-40e3-abd3-d7a960b8b2ad" (UID: "9aa78020-05ca-40e3-abd3-d7a960b8b2ad"). InnerVolumeSpecName "kube-api-access-6hzwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.456466 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9aa78020-05ca-40e3-abd3-d7a960b8b2ad" (UID: "9aa78020-05ca-40e3-abd3-d7a960b8b2ad"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.464071 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hzwf\" (UniqueName: \"kubernetes.io/projected/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-kube-api-access-6hzwf\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.464099 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.464268 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9aa78020-05ca-40e3-abd3-d7a960b8b2ad" (UID: "9aa78020-05ca-40e3-abd3-d7a960b8b2ad"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.473808 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-config" (OuterVolumeSpecName: "config") pod "9aa78020-05ca-40e3-abd3-d7a960b8b2ad" (UID: "9aa78020-05ca-40e3-abd3-d7a960b8b2ad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.490629 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9aa78020-05ca-40e3-abd3-d7a960b8b2ad" (UID: "9aa78020-05ca-40e3-abd3-d7a960b8b2ad"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.499863 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9aa78020-05ca-40e3-abd3-d7a960b8b2ad" (UID: "9aa78020-05ca-40e3-abd3-d7a960b8b2ad"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.566294 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.566485 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.566545 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.566625 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aa78020-05ca-40e3-abd3-d7a960b8b2ad-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.644071 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.703587 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.762909 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.777349 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66g7q\" (UniqueName: \"kubernetes.io/projected/8ee94784-1d45-48fa-b7ad-81f12013c44c-kube-api-access-66g7q\") pod \"8ee94784-1d45-48fa-b7ad-81f12013c44c\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.777447 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-config-data\") pod \"8ee94784-1d45-48fa-b7ad-81f12013c44c\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.777652 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-combined-ca-bundle\") pod \"8ee94784-1d45-48fa-b7ad-81f12013c44c\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.777703 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-scripts\") pod \"8ee94784-1d45-48fa-b7ad-81f12013c44c\" (UID: \"8ee94784-1d45-48fa-b7ad-81f12013c44c\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.784500 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-scripts" (OuterVolumeSpecName: "scripts") pod "8ee94784-1d45-48fa-b7ad-81f12013c44c" (UID: "8ee94784-1d45-48fa-b7ad-81f12013c44c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.784677 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ee94784-1d45-48fa-b7ad-81f12013c44c-kube-api-access-66g7q" (OuterVolumeSpecName: "kube-api-access-66g7q") pod "8ee94784-1d45-48fa-b7ad-81f12013c44c" (UID: "8ee94784-1d45-48fa-b7ad-81f12013c44c"). InnerVolumeSpecName "kube-api-access-66g7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.791621 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66g7q\" (UniqueName: \"kubernetes.io/projected/8ee94784-1d45-48fa-b7ad-81f12013c44c-kube-api-access-66g7q\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.791643 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.826477 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-config-data" (OuterVolumeSpecName: "config-data") pod "8ee94784-1d45-48fa-b7ad-81f12013c44c" (UID: "8ee94784-1d45-48fa-b7ad-81f12013c44c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.826841 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ee94784-1d45-48fa-b7ad-81f12013c44c" (UID: "8ee94784-1d45-48fa-b7ad-81f12013c44c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.892643 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-nova-metadata-tls-certs\") pod \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.892719 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7wwl\" (UniqueName: \"kubernetes.io/projected/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-kube-api-access-k7wwl\") pod \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.892835 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-combined-ca-bundle\") pod \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.892904 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-config-data\") pod \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.892978 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-logs\") pod \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\" (UID: \"0fa1011f-07f0-4a21-9d46-00ba9b51c89d\") " Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.893325 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.893343 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ee94784-1d45-48fa-b7ad-81f12013c44c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.893608 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-logs" (OuterVolumeSpecName: "logs") pod "0fa1011f-07f0-4a21-9d46-00ba9b51c89d" (UID: "0fa1011f-07f0-4a21-9d46-00ba9b51c89d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.896108 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-kube-api-access-k7wwl" (OuterVolumeSpecName: "kube-api-access-k7wwl") pod "0fa1011f-07f0-4a21-9d46-00ba9b51c89d" (UID: "0fa1011f-07f0-4a21-9d46-00ba9b51c89d"). InnerVolumeSpecName "kube-api-access-k7wwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.916729 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0fa1011f-07f0-4a21-9d46-00ba9b51c89d" (UID: "0fa1011f-07f0-4a21-9d46-00ba9b51c89d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.918477 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-config-data" (OuterVolumeSpecName: "config-data") pod "0fa1011f-07f0-4a21-9d46-00ba9b51c89d" (UID: "0fa1011f-07f0-4a21-9d46-00ba9b51c89d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.934736 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "0fa1011f-07f0-4a21-9d46-00ba9b51c89d" (UID: "0fa1011f-07f0-4a21-9d46-00ba9b51c89d"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.995350 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.995380 4775 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.995391 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7wwl\" (UniqueName: \"kubernetes.io/projected/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-kube-api-access-k7wwl\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.995401 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:17 crc kubenswrapper[4775]: I1002 02:02:17.995409 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0fa1011f-07f0-4a21-9d46-00ba9b51c89d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.099509 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-7jsjh" event={"ID":"8ee94784-1d45-48fa-b7ad-81f12013c44c","Type":"ContainerDied","Data":"a4ce69e6d4e95ac5dc563807b38923e3f20371c303752e81cc7a46b7b73aa427"} Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.099548 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4ce69e6d4e95ac5dc563807b38923e3f20371c303752e81cc7a46b7b73aa427" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.099623 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-7jsjh" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.103022 4775 generic.go:334] "Generic (PLEG): container finished" podID="0fa1011f-07f0-4a21-9d46-00ba9b51c89d" containerID="fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702" exitCode=0 Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.103077 4775 generic.go:334] "Generic (PLEG): container finished" podID="0fa1011f-07f0-4a21-9d46-00ba9b51c89d" containerID="21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f" exitCode=143 Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.103155 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fa1011f-07f0-4a21-9d46-00ba9b51c89d","Type":"ContainerDied","Data":"fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702"} Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.103206 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fa1011f-07f0-4a21-9d46-00ba9b51c89d","Type":"ContainerDied","Data":"21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f"} Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.103226 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0fa1011f-07f0-4a21-9d46-00ba9b51c89d","Type":"ContainerDied","Data":"96dc9ec0c6196eeaf2a6da8dda6877d49aee8cb78d3d5083ac42600044dd348c"} Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.103252 4775 scope.go:117] "RemoveContainer" containerID="fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.103432 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.114093 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-rf268" event={"ID":"9aa78020-05ca-40e3-abd3-d7a960b8b2ad","Type":"ContainerDied","Data":"4757c352696a93e19189bf85ac66b6319d45160727f5310ed6fff72f7634afe8"} Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.114269 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-rf268" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.122329 4775 generic.go:334] "Generic (PLEG): container finished" podID="20b54b85-0b61-4332-869c-1f2f8c07b4fb" containerID="58c0cbe1782d0bf708643480ecd6e72fcb640e1aab07bded85208680637d966d" exitCode=143 Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.122526 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20b54b85-0b61-4332-869c-1f2f8c07b4fb","Type":"ContainerDied","Data":"58c0cbe1782d0bf708643480ecd6e72fcb640e1aab07bded85208680637d966d"} Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.175527 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-rf268"] Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.186880 4775 scope.go:117] "RemoveContainer" containerID="21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.190659 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-rf268"] Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.203282 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 02:02:18 crc kubenswrapper[4775]: E1002 02:02:18.203714 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa78020-05ca-40e3-abd3-d7a960b8b2ad" containerName="init" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.203733 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa78020-05ca-40e3-abd3-d7a960b8b2ad" containerName="init" Oct 02 02:02:18 crc kubenswrapper[4775]: E1002 02:02:18.203743 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fa1011f-07f0-4a21-9d46-00ba9b51c89d" containerName="nova-metadata-log" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.203752 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fa1011f-07f0-4a21-9d46-00ba9b51c89d" containerName="nova-metadata-log" Oct 02 02:02:18 crc kubenswrapper[4775]: E1002 02:02:18.203785 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="777deed6-1940-4ae7-b4bb-e9f71fa4429e" containerName="nova-manage" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.203791 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="777deed6-1940-4ae7-b4bb-e9f71fa4429e" containerName="nova-manage" Oct 02 02:02:18 crc kubenswrapper[4775]: E1002 02:02:18.203807 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa78020-05ca-40e3-abd3-d7a960b8b2ad" containerName="dnsmasq-dns" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.203814 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa78020-05ca-40e3-abd3-d7a960b8b2ad" containerName="dnsmasq-dns" Oct 02 02:02:18 crc kubenswrapper[4775]: E1002 02:02:18.203823 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fa1011f-07f0-4a21-9d46-00ba9b51c89d" containerName="nova-metadata-metadata" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.203829 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fa1011f-07f0-4a21-9d46-00ba9b51c89d" containerName="nova-metadata-metadata" Oct 02 02:02:18 crc kubenswrapper[4775]: E1002 02:02:18.203846 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ee94784-1d45-48fa-b7ad-81f12013c44c" containerName="nova-cell1-conductor-db-sync" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.203851 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ee94784-1d45-48fa-b7ad-81f12013c44c" containerName="nova-cell1-conductor-db-sync" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.204051 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fa1011f-07f0-4a21-9d46-00ba9b51c89d" containerName="nova-metadata-log" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.204067 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aa78020-05ca-40e3-abd3-d7a960b8b2ad" containerName="dnsmasq-dns" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.204081 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="777deed6-1940-4ae7-b4bb-e9f71fa4429e" containerName="nova-manage" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.204098 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ee94784-1d45-48fa-b7ad-81f12013c44c" containerName="nova-cell1-conductor-db-sync" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.204114 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fa1011f-07f0-4a21-9d46-00ba9b51c89d" containerName="nova-metadata-metadata" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.204376 4775 scope.go:117] "RemoveContainer" containerID="fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.204714 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 02:02:18 crc kubenswrapper[4775]: E1002 02:02:18.204884 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702\": container with ID starting with fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702 not found: ID does not exist" containerID="fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.204917 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702"} err="failed to get container status \"fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702\": rpc error: code = NotFound desc = could not find container \"fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702\": container with ID starting with fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702 not found: ID does not exist" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.204937 4775 scope.go:117] "RemoveContainer" containerID="21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f" Oct 02 02:02:18 crc kubenswrapper[4775]: E1002 02:02:18.205250 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f\": container with ID starting with 21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f not found: ID does not exist" containerID="21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.205305 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f"} err="failed to get container status \"21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f\": rpc error: code = NotFound desc = could not find container \"21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f\": container with ID starting with 21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f not found: ID does not exist" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.205339 4775 scope.go:117] "RemoveContainer" containerID="fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.205633 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702"} err="failed to get container status \"fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702\": rpc error: code = NotFound desc = could not find container \"fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702\": container with ID starting with fd3e734eb0740402b418b2956a6bf54b9d606f36a04eee6ad367185469692702 not found: ID does not exist" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.205656 4775 scope.go:117] "RemoveContainer" containerID="21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.206026 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f"} err="failed to get container status \"21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f\": rpc error: code = NotFound desc = could not find container \"21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f\": container with ID starting with 21ba2fbb996b51ff535637200a2ee7d24b0d32c95b3b71a5a2b3742c12795b6f not found: ID does not exist" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.206055 4775 scope.go:117] "RemoveContainer" containerID="c7e1a112ab7147ecb551a7b9f6aef7efdc4bb4197ccd00766ebfc02583672984" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.237420 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.241063 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.259239 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.274455 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.287098 4775 scope.go:117] "RemoveContainer" containerID="9e75c63ad61e752ebf2677a4126de52817646ff1a38139c6459d1e49b6e73721" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.300220 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46453bb2-6af0-4dd4-83ba-b14bf3f96310-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.300328 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46453bb2-6af0-4dd4-83ba-b14bf3f96310-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.300360 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rglgf\" (UniqueName: \"kubernetes.io/projected/46453bb2-6af0-4dd4-83ba-b14bf3f96310-kube-api-access-rglgf\") pod \"nova-cell1-conductor-0\" (UID: \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.321659 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.323074 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.325797 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.325810 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.328388 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.401867 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-config-data\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.401972 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46453bb2-6af0-4dd4-83ba-b14bf3f96310-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.402029 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.402071 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.402104 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f59877f-b40e-4585-a6a0-b22a8cf885eb-logs\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.402120 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7rtk\" (UniqueName: \"kubernetes.io/projected/6f59877f-b40e-4585-a6a0-b22a8cf885eb-kube-api-access-l7rtk\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.402149 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46453bb2-6af0-4dd4-83ba-b14bf3f96310-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.402181 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rglgf\" (UniqueName: \"kubernetes.io/projected/46453bb2-6af0-4dd4-83ba-b14bf3f96310-kube-api-access-rglgf\") pod \"nova-cell1-conductor-0\" (UID: \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.406332 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46453bb2-6af0-4dd4-83ba-b14bf3f96310-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.406586 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46453bb2-6af0-4dd4-83ba-b14bf3f96310-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.420612 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rglgf\" (UniqueName: \"kubernetes.io/projected/46453bb2-6af0-4dd4-83ba-b14bf3f96310-kube-api-access-rglgf\") pod \"nova-cell1-conductor-0\" (UID: \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\") " pod="openstack/nova-cell1-conductor-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.503458 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.503521 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.503556 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f59877f-b40e-4585-a6a0-b22a8cf885eb-logs\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.503574 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7rtk\" (UniqueName: \"kubernetes.io/projected/6f59877f-b40e-4585-a6a0-b22a8cf885eb-kube-api-access-l7rtk\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.503632 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-config-data\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.503973 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f59877f-b40e-4585-a6a0-b22a8cf885eb-logs\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.508306 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-config-data\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.509052 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.509475 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.521990 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7rtk\" (UniqueName: \"kubernetes.io/projected/6f59877f-b40e-4585-a6a0-b22a8cf885eb-kube-api-access-l7rtk\") pod \"nova-metadata-0\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.563612 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.643649 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:02:18 crc kubenswrapper[4775]: I1002 02:02:18.999547 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 02:02:19 crc kubenswrapper[4775]: W1002 02:02:19.004153 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46453bb2_6af0_4dd4_83ba_b14bf3f96310.slice/crio-e79277171e00d6730ea668e1d3a8b2e2495caf310572d88755b3b04fcb9a6364 WatchSource:0}: Error finding container e79277171e00d6730ea668e1d3a8b2e2495caf310572d88755b3b04fcb9a6364: Status 404 returned error can't find the container with id e79277171e00d6730ea668e1d3a8b2e2495caf310572d88755b3b04fcb9a6364 Oct 02 02:02:19 crc kubenswrapper[4775]: I1002 02:02:19.134076 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:02:19 crc kubenswrapper[4775]: I1002 02:02:19.147310 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"46453bb2-6af0-4dd4-83ba-b14bf3f96310","Type":"ContainerStarted","Data":"e79277171e00d6730ea668e1d3a8b2e2495caf310572d88755b3b04fcb9a6364"} Oct 02 02:02:19 crc kubenswrapper[4775]: I1002 02:02:19.150103 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="78473f6a-8832-40d9-8d3d-0ea8ee59c1f3" containerName="nova-scheduler-scheduler" containerID="cri-o://c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b" gracePeriod=30 Oct 02 02:02:19 crc kubenswrapper[4775]: I1002 02:02:19.776890 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fa1011f-07f0-4a21-9d46-00ba9b51c89d" path="/var/lib/kubelet/pods/0fa1011f-07f0-4a21-9d46-00ba9b51c89d/volumes" Oct 02 02:02:19 crc kubenswrapper[4775]: I1002 02:02:19.778057 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aa78020-05ca-40e3-abd3-d7a960b8b2ad" path="/var/lib/kubelet/pods/9aa78020-05ca-40e3-abd3-d7a960b8b2ad/volumes" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.112627 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.166282 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"46453bb2-6af0-4dd4-83ba-b14bf3f96310","Type":"ContainerStarted","Data":"b66fc6b4d3be69dcb9c2d949e6dfe8d6ab6ddb59f7b8ad17c8380c0b6034ccf6"} Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.168305 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.172649 4775 generic.go:334] "Generic (PLEG): container finished" podID="1319e744-d454-4024-898e-ac338b1b8443" containerID="c26f10b5e5a033ef447a32716656df3e4eb53bccf0a74b1bac910f359b58ff24" exitCode=0 Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.172715 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1319e744-d454-4024-898e-ac338b1b8443","Type":"ContainerDied","Data":"c26f10b5e5a033ef447a32716656df3e4eb53bccf0a74b1bac910f359b58ff24"} Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.172736 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1319e744-d454-4024-898e-ac338b1b8443","Type":"ContainerDied","Data":"687d57ea852571f9784926dff665df1db5d839665a912bd929a9e716a4f50bbe"} Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.172754 4775 scope.go:117] "RemoveContainer" containerID="e3133fad4451f93b08768783e43e7c4303dcbf17baae0f5714dfb7c95d316e71" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.172890 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.180404 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.180389672 podStartE2EDuration="2.180389672s" podCreationTimestamp="2025-10-02 02:02:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:20.17955427 +0000 UTC m=+1277.346298300" watchObservedRunningTime="2025-10-02 02:02:20.180389672 +0000 UTC m=+1277.347133712" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.186407 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f59877f-b40e-4585-a6a0-b22a8cf885eb","Type":"ContainerStarted","Data":"f68aaa46749cbbb1fff57d01d3426721b116a24cc8c1b3ee14b7fd2f36824d8f"} Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.186452 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f59877f-b40e-4585-a6a0-b22a8cf885eb","Type":"ContainerStarted","Data":"5de23478e0ccd06671708d54ffe109e061c817a4c1d2f255498395c9c5eda273"} Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.186467 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f59877f-b40e-4585-a6a0-b22a8cf885eb","Type":"ContainerStarted","Data":"28ab43a939e893a3d6986354846470f9bd135dfd2190d23da9e2504608e69600"} Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.222296 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.222270557 podStartE2EDuration="2.222270557s" podCreationTimestamp="2025-10-02 02:02:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:20.207208894 +0000 UTC m=+1277.373952944" watchObservedRunningTime="2025-10-02 02:02:20.222270557 +0000 UTC m=+1277.389014597" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.232086 4775 scope.go:117] "RemoveContainer" containerID="0f915dee9ad3972ac1aaeee7f51be90d1210c168ef78f9afce2d8bf3c6fa6f2d" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.243640 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-scripts\") pod \"1319e744-d454-4024-898e-ac338b1b8443\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.243688 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-config-data\") pod \"1319e744-d454-4024-898e-ac338b1b8443\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.243730 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1319e744-d454-4024-898e-ac338b1b8443-run-httpd\") pod \"1319e744-d454-4024-898e-ac338b1b8443\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.243753 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1319e744-d454-4024-898e-ac338b1b8443-log-httpd\") pod \"1319e744-d454-4024-898e-ac338b1b8443\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.243853 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slbvn\" (UniqueName: \"kubernetes.io/projected/1319e744-d454-4024-898e-ac338b1b8443-kube-api-access-slbvn\") pod \"1319e744-d454-4024-898e-ac338b1b8443\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.243871 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-sg-core-conf-yaml\") pod \"1319e744-d454-4024-898e-ac338b1b8443\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.243919 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-combined-ca-bundle\") pod \"1319e744-d454-4024-898e-ac338b1b8443\" (UID: \"1319e744-d454-4024-898e-ac338b1b8443\") " Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.244716 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1319e744-d454-4024-898e-ac338b1b8443-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1319e744-d454-4024-898e-ac338b1b8443" (UID: "1319e744-d454-4024-898e-ac338b1b8443"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.245273 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1319e744-d454-4024-898e-ac338b1b8443-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1319e744-d454-4024-898e-ac338b1b8443" (UID: "1319e744-d454-4024-898e-ac338b1b8443"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.252815 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-scripts" (OuterVolumeSpecName: "scripts") pod "1319e744-d454-4024-898e-ac338b1b8443" (UID: "1319e744-d454-4024-898e-ac338b1b8443"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.253683 4775 scope.go:117] "RemoveContainer" containerID="c26f10b5e5a033ef447a32716656df3e4eb53bccf0a74b1bac910f359b58ff24" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.257314 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1319e744-d454-4024-898e-ac338b1b8443-kube-api-access-slbvn" (OuterVolumeSpecName: "kube-api-access-slbvn") pod "1319e744-d454-4024-898e-ac338b1b8443" (UID: "1319e744-d454-4024-898e-ac338b1b8443"). InnerVolumeSpecName "kube-api-access-slbvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.274561 4775 scope.go:117] "RemoveContainer" containerID="c315218740c2142dcb8618c469086e3dcefad74469b76f74a24edf1a27bfc666" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.277104 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1319e744-d454-4024-898e-ac338b1b8443" (UID: "1319e744-d454-4024-898e-ac338b1b8443"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.299231 4775 scope.go:117] "RemoveContainer" containerID="e3133fad4451f93b08768783e43e7c4303dcbf17baae0f5714dfb7c95d316e71" Oct 02 02:02:20 crc kubenswrapper[4775]: E1002 02:02:20.299705 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3133fad4451f93b08768783e43e7c4303dcbf17baae0f5714dfb7c95d316e71\": container with ID starting with e3133fad4451f93b08768783e43e7c4303dcbf17baae0f5714dfb7c95d316e71 not found: ID does not exist" containerID="e3133fad4451f93b08768783e43e7c4303dcbf17baae0f5714dfb7c95d316e71" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.299735 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3133fad4451f93b08768783e43e7c4303dcbf17baae0f5714dfb7c95d316e71"} err="failed to get container status \"e3133fad4451f93b08768783e43e7c4303dcbf17baae0f5714dfb7c95d316e71\": rpc error: code = NotFound desc = could not find container \"e3133fad4451f93b08768783e43e7c4303dcbf17baae0f5714dfb7c95d316e71\": container with ID starting with e3133fad4451f93b08768783e43e7c4303dcbf17baae0f5714dfb7c95d316e71 not found: ID does not exist" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.299758 4775 scope.go:117] "RemoveContainer" containerID="0f915dee9ad3972ac1aaeee7f51be90d1210c168ef78f9afce2d8bf3c6fa6f2d" Oct 02 02:02:20 crc kubenswrapper[4775]: E1002 02:02:20.300278 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f915dee9ad3972ac1aaeee7f51be90d1210c168ef78f9afce2d8bf3c6fa6f2d\": container with ID starting with 0f915dee9ad3972ac1aaeee7f51be90d1210c168ef78f9afce2d8bf3c6fa6f2d not found: ID does not exist" containerID="0f915dee9ad3972ac1aaeee7f51be90d1210c168ef78f9afce2d8bf3c6fa6f2d" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.300336 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f915dee9ad3972ac1aaeee7f51be90d1210c168ef78f9afce2d8bf3c6fa6f2d"} err="failed to get container status \"0f915dee9ad3972ac1aaeee7f51be90d1210c168ef78f9afce2d8bf3c6fa6f2d\": rpc error: code = NotFound desc = could not find container \"0f915dee9ad3972ac1aaeee7f51be90d1210c168ef78f9afce2d8bf3c6fa6f2d\": container with ID starting with 0f915dee9ad3972ac1aaeee7f51be90d1210c168ef78f9afce2d8bf3c6fa6f2d not found: ID does not exist" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.300373 4775 scope.go:117] "RemoveContainer" containerID="c26f10b5e5a033ef447a32716656df3e4eb53bccf0a74b1bac910f359b58ff24" Oct 02 02:02:20 crc kubenswrapper[4775]: E1002 02:02:20.300847 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c26f10b5e5a033ef447a32716656df3e4eb53bccf0a74b1bac910f359b58ff24\": container with ID starting with c26f10b5e5a033ef447a32716656df3e4eb53bccf0a74b1bac910f359b58ff24 not found: ID does not exist" containerID="c26f10b5e5a033ef447a32716656df3e4eb53bccf0a74b1bac910f359b58ff24" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.300890 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c26f10b5e5a033ef447a32716656df3e4eb53bccf0a74b1bac910f359b58ff24"} err="failed to get container status \"c26f10b5e5a033ef447a32716656df3e4eb53bccf0a74b1bac910f359b58ff24\": rpc error: code = NotFound desc = could not find container \"c26f10b5e5a033ef447a32716656df3e4eb53bccf0a74b1bac910f359b58ff24\": container with ID starting with c26f10b5e5a033ef447a32716656df3e4eb53bccf0a74b1bac910f359b58ff24 not found: ID does not exist" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.300916 4775 scope.go:117] "RemoveContainer" containerID="c315218740c2142dcb8618c469086e3dcefad74469b76f74a24edf1a27bfc666" Oct 02 02:02:20 crc kubenswrapper[4775]: E1002 02:02:20.301330 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c315218740c2142dcb8618c469086e3dcefad74469b76f74a24edf1a27bfc666\": container with ID starting with c315218740c2142dcb8618c469086e3dcefad74469b76f74a24edf1a27bfc666 not found: ID does not exist" containerID="c315218740c2142dcb8618c469086e3dcefad74469b76f74a24edf1a27bfc666" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.301356 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c315218740c2142dcb8618c469086e3dcefad74469b76f74a24edf1a27bfc666"} err="failed to get container status \"c315218740c2142dcb8618c469086e3dcefad74469b76f74a24edf1a27bfc666\": rpc error: code = NotFound desc = could not find container \"c315218740c2142dcb8618c469086e3dcefad74469b76f74a24edf1a27bfc666\": container with ID starting with c315218740c2142dcb8618c469086e3dcefad74469b76f74a24edf1a27bfc666 not found: ID does not exist" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.314602 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1319e744-d454-4024-898e-ac338b1b8443" (UID: "1319e744-d454-4024-898e-ac338b1b8443"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.341679 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-config-data" (OuterVolumeSpecName: "config-data") pod "1319e744-d454-4024-898e-ac338b1b8443" (UID: "1319e744-d454-4024-898e-ac338b1b8443"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.345731 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slbvn\" (UniqueName: \"kubernetes.io/projected/1319e744-d454-4024-898e-ac338b1b8443-kube-api-access-slbvn\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.345762 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.345771 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.345781 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.345790 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1319e744-d454-4024-898e-ac338b1b8443-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.345799 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1319e744-d454-4024-898e-ac338b1b8443-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.345808 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1319e744-d454-4024-898e-ac338b1b8443-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.505835 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.512949 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.535126 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:02:20 crc kubenswrapper[4775]: E1002 02:02:20.535573 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="sg-core" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.535585 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="sg-core" Oct 02 02:02:20 crc kubenswrapper[4775]: E1002 02:02:20.535604 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="ceilometer-notification-agent" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.535610 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="ceilometer-notification-agent" Oct 02 02:02:20 crc kubenswrapper[4775]: E1002 02:02:20.535644 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="proxy-httpd" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.535650 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="proxy-httpd" Oct 02 02:02:20 crc kubenswrapper[4775]: E1002 02:02:20.535667 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="ceilometer-central-agent" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.535674 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="ceilometer-central-agent" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.535842 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="ceilometer-notification-agent" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.535856 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="proxy-httpd" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.535868 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="ceilometer-central-agent" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.535878 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1319e744-d454-4024-898e-ac338b1b8443" containerName="sg-core" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.537632 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.539149 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.539662 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.540918 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.541144 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.661311 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-config-data\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.661379 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-scripts\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.661511 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47865629-abc2-4da9-94d7-a0d389f9570c-run-httpd\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.661683 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.661706 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.661749 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.661818 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pj6d\" (UniqueName: \"kubernetes.io/projected/47865629-abc2-4da9-94d7-a0d389f9570c-kube-api-access-7pj6d\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.661932 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47865629-abc2-4da9-94d7-a0d389f9570c-log-httpd\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.763876 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-config-data\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.763933 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-scripts\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.763977 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47865629-abc2-4da9-94d7-a0d389f9570c-run-httpd\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.764027 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.764067 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.764099 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.764137 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pj6d\" (UniqueName: \"kubernetes.io/projected/47865629-abc2-4da9-94d7-a0d389f9570c-kube-api-access-7pj6d\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.764185 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47865629-abc2-4da9-94d7-a0d389f9570c-log-httpd\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.764698 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47865629-abc2-4da9-94d7-a0d389f9570c-log-httpd\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.764892 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47865629-abc2-4da9-94d7-a0d389f9570c-run-httpd\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.770391 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-config-data\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.773211 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.774400 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.780169 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-scripts\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.783679 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.795869 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pj6d\" (UniqueName: \"kubernetes.io/projected/47865629-abc2-4da9-94d7-a0d389f9570c-kube-api-access-7pj6d\") pod \"ceilometer-0\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " pod="openstack/ceilometer-0" Oct 02 02:02:20 crc kubenswrapper[4775]: I1002 02:02:20.884189 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:02:21 crc kubenswrapper[4775]: I1002 02:02:21.358208 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:02:21 crc kubenswrapper[4775]: W1002 02:02:21.377195 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47865629_abc2_4da9_94d7_a0d389f9570c.slice/crio-d536196b445f1a21ca895b5a9063aaf4845b6b511f28bcb9a60e162278c267f0 WatchSource:0}: Error finding container d536196b445f1a21ca895b5a9063aaf4845b6b511f28bcb9a60e162278c267f0: Status 404 returned error can't find the container with id d536196b445f1a21ca895b5a9063aaf4845b6b511f28bcb9a60e162278c267f0 Oct 02 02:02:21 crc kubenswrapper[4775]: E1002 02:02:21.709360 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 02:02:21 crc kubenswrapper[4775]: E1002 02:02:21.711991 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 02:02:21 crc kubenswrapper[4775]: E1002 02:02:21.713497 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 02:02:21 crc kubenswrapper[4775]: E1002 02:02:21.713532 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="78473f6a-8832-40d9-8d3d-0ea8ee59c1f3" containerName="nova-scheduler-scheduler" Oct 02 02:02:21 crc kubenswrapper[4775]: I1002 02:02:21.789566 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1319e744-d454-4024-898e-ac338b1b8443" path="/var/lib/kubelet/pods/1319e744-d454-4024-898e-ac338b1b8443/volumes" Oct 02 02:02:22 crc kubenswrapper[4775]: I1002 02:02:22.207409 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47865629-abc2-4da9-94d7-a0d389f9570c","Type":"ContainerStarted","Data":"6e2f3b70a5c27a2136fce6a4efd75732be583ec15cbf867ee9f595b46f1953ea"} Oct 02 02:02:22 crc kubenswrapper[4775]: I1002 02:02:22.207455 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47865629-abc2-4da9-94d7-a0d389f9570c","Type":"ContainerStarted","Data":"d536196b445f1a21ca895b5a9063aaf4845b6b511f28bcb9a60e162278c267f0"} Oct 02 02:02:22 crc kubenswrapper[4775]: I1002 02:02:22.973699 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.144568 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20b54b85-0b61-4332-869c-1f2f8c07b4fb-combined-ca-bundle\") pod \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.144782 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf8nf\" (UniqueName: \"kubernetes.io/projected/20b54b85-0b61-4332-869c-1f2f8c07b4fb-kube-api-access-gf8nf\") pod \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.144806 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20b54b85-0b61-4332-869c-1f2f8c07b4fb-config-data\") pod \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.144827 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20b54b85-0b61-4332-869c-1f2f8c07b4fb-logs\") pod \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\" (UID: \"20b54b85-0b61-4332-869c-1f2f8c07b4fb\") " Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.145649 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20b54b85-0b61-4332-869c-1f2f8c07b4fb-logs" (OuterVolumeSpecName: "logs") pod "20b54b85-0b61-4332-869c-1f2f8c07b4fb" (UID: "20b54b85-0b61-4332-869c-1f2f8c07b4fb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.145703 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.150423 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b54b85-0b61-4332-869c-1f2f8c07b4fb-kube-api-access-gf8nf" (OuterVolumeSpecName: "kube-api-access-gf8nf") pod "20b54b85-0b61-4332-869c-1f2f8c07b4fb" (UID: "20b54b85-0b61-4332-869c-1f2f8c07b4fb"). InnerVolumeSpecName "kube-api-access-gf8nf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.176098 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b54b85-0b61-4332-869c-1f2f8c07b4fb-config-data" (OuterVolumeSpecName: "config-data") pod "20b54b85-0b61-4332-869c-1f2f8c07b4fb" (UID: "20b54b85-0b61-4332-869c-1f2f8c07b4fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.182564 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b54b85-0b61-4332-869c-1f2f8c07b4fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "20b54b85-0b61-4332-869c-1f2f8c07b4fb" (UID: "20b54b85-0b61-4332-869c-1f2f8c07b4fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.218148 4775 generic.go:334] "Generic (PLEG): container finished" podID="20b54b85-0b61-4332-869c-1f2f8c07b4fb" containerID="9c58be88cd4130340b3ca2a4f3bc647e10ec86b0ed1878d66aeab0f478633046" exitCode=0 Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.218255 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.219821 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20b54b85-0b61-4332-869c-1f2f8c07b4fb","Type":"ContainerDied","Data":"9c58be88cd4130340b3ca2a4f3bc647e10ec86b0ed1878d66aeab0f478633046"} Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.219874 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20b54b85-0b61-4332-869c-1f2f8c07b4fb","Type":"ContainerDied","Data":"70a17366be9d6b080684e25d522b5c3ec4fc11b171629239a9475d3a19ba627c"} Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.219896 4775 scope.go:117] "RemoveContainer" containerID="9c58be88cd4130340b3ca2a4f3bc647e10ec86b0ed1878d66aeab0f478633046" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.222186 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47865629-abc2-4da9-94d7-a0d389f9570c","Type":"ContainerStarted","Data":"e15f9bb8da4554af6844bb4153fc54e557917adfaafc48f55338ba8ccfa96018"} Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.223415 4775 generic.go:334] "Generic (PLEG): container finished" podID="78473f6a-8832-40d9-8d3d-0ea8ee59c1f3" containerID="c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b" exitCode=0 Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.223441 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3","Type":"ContainerDied","Data":"c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b"} Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.223455 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3","Type":"ContainerDied","Data":"1102614fc06f60a9344f26ba887143892fafdd9ed2c392fb196d172ff7753009"} Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.223498 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.246180 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htjr6\" (UniqueName: \"kubernetes.io/projected/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-kube-api-access-htjr6\") pod \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\" (UID: \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\") " Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.246268 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-combined-ca-bundle\") pod \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\" (UID: \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\") " Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.246417 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-config-data\") pod \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\" (UID: \"78473f6a-8832-40d9-8d3d-0ea8ee59c1f3\") " Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.246733 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20b54b85-0b61-4332-869c-1f2f8c07b4fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.246749 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf8nf\" (UniqueName: \"kubernetes.io/projected/20b54b85-0b61-4332-869c-1f2f8c07b4fb-kube-api-access-gf8nf\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.246759 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20b54b85-0b61-4332-869c-1f2f8c07b4fb-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.246768 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20b54b85-0b61-4332-869c-1f2f8c07b4fb-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.251277 4775 scope.go:117] "RemoveContainer" containerID="58c0cbe1782d0bf708643480ecd6e72fcb640e1aab07bded85208680637d966d" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.262594 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.272401 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-kube-api-access-htjr6" (OuterVolumeSpecName: "kube-api-access-htjr6") pod "78473f6a-8832-40d9-8d3d-0ea8ee59c1f3" (UID: "78473f6a-8832-40d9-8d3d-0ea8ee59c1f3"). InnerVolumeSpecName "kube-api-access-htjr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.278416 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.280653 4775 scope.go:117] "RemoveContainer" containerID="9c58be88cd4130340b3ca2a4f3bc647e10ec86b0ed1878d66aeab0f478633046" Oct 02 02:02:23 crc kubenswrapper[4775]: E1002 02:02:23.281817 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c58be88cd4130340b3ca2a4f3bc647e10ec86b0ed1878d66aeab0f478633046\": container with ID starting with 9c58be88cd4130340b3ca2a4f3bc647e10ec86b0ed1878d66aeab0f478633046 not found: ID does not exist" containerID="9c58be88cd4130340b3ca2a4f3bc647e10ec86b0ed1878d66aeab0f478633046" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.281851 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c58be88cd4130340b3ca2a4f3bc647e10ec86b0ed1878d66aeab0f478633046"} err="failed to get container status \"9c58be88cd4130340b3ca2a4f3bc647e10ec86b0ed1878d66aeab0f478633046\": rpc error: code = NotFound desc = could not find container \"9c58be88cd4130340b3ca2a4f3bc647e10ec86b0ed1878d66aeab0f478633046\": container with ID starting with 9c58be88cd4130340b3ca2a4f3bc647e10ec86b0ed1878d66aeab0f478633046 not found: ID does not exist" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.281872 4775 scope.go:117] "RemoveContainer" containerID="58c0cbe1782d0bf708643480ecd6e72fcb640e1aab07bded85208680637d966d" Oct 02 02:02:23 crc kubenswrapper[4775]: E1002 02:02:23.285867 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58c0cbe1782d0bf708643480ecd6e72fcb640e1aab07bded85208680637d966d\": container with ID starting with 58c0cbe1782d0bf708643480ecd6e72fcb640e1aab07bded85208680637d966d not found: ID does not exist" containerID="58c0cbe1782d0bf708643480ecd6e72fcb640e1aab07bded85208680637d966d" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.285910 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58c0cbe1782d0bf708643480ecd6e72fcb640e1aab07bded85208680637d966d"} err="failed to get container status \"58c0cbe1782d0bf708643480ecd6e72fcb640e1aab07bded85208680637d966d\": rpc error: code = NotFound desc = could not find container \"58c0cbe1782d0bf708643480ecd6e72fcb640e1aab07bded85208680637d966d\": container with ID starting with 58c0cbe1782d0bf708643480ecd6e72fcb640e1aab07bded85208680637d966d not found: ID does not exist" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.285935 4775 scope.go:117] "RemoveContainer" containerID="c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.290177 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.290481 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "78473f6a-8832-40d9-8d3d-0ea8ee59c1f3" (UID: "78473f6a-8832-40d9-8d3d-0ea8ee59c1f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:23 crc kubenswrapper[4775]: E1002 02:02:23.290683 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78473f6a-8832-40d9-8d3d-0ea8ee59c1f3" containerName="nova-scheduler-scheduler" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.290702 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="78473f6a-8832-40d9-8d3d-0ea8ee59c1f3" containerName="nova-scheduler-scheduler" Oct 02 02:02:23 crc kubenswrapper[4775]: E1002 02:02:23.290735 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20b54b85-0b61-4332-869c-1f2f8c07b4fb" containerName="nova-api-log" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.290742 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="20b54b85-0b61-4332-869c-1f2f8c07b4fb" containerName="nova-api-log" Oct 02 02:02:23 crc kubenswrapper[4775]: E1002 02:02:23.290756 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20b54b85-0b61-4332-869c-1f2f8c07b4fb" containerName="nova-api-api" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.290762 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="20b54b85-0b61-4332-869c-1f2f8c07b4fb" containerName="nova-api-api" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.290931 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="20b54b85-0b61-4332-869c-1f2f8c07b4fb" containerName="nova-api-log" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.290966 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="78473f6a-8832-40d9-8d3d-0ea8ee59c1f3" containerName="nova-scheduler-scheduler" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.290977 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="20b54b85-0b61-4332-869c-1f2f8c07b4fb" containerName="nova-api-api" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.292041 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.295651 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.298320 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.305120 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-config-data" (OuterVolumeSpecName: "config-data") pod "78473f6a-8832-40d9-8d3d-0ea8ee59c1f3" (UID: "78473f6a-8832-40d9-8d3d-0ea8ee59c1f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.321267 4775 scope.go:117] "RemoveContainer" containerID="c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b" Oct 02 02:02:23 crc kubenswrapper[4775]: E1002 02:02:23.321684 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b\": container with ID starting with c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b not found: ID does not exist" containerID="c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.321715 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b"} err="failed to get container status \"c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b\": rpc error: code = NotFound desc = could not find container \"c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b\": container with ID starting with c040fa3ff40ab70ee12738ab0f31a94a4eab0f8fc0dd3d72101769d0f49e8f7b not found: ID does not exist" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.347888 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htjr6\" (UniqueName: \"kubernetes.io/projected/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-kube-api-access-htjr6\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.347920 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.347930 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.449257 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqcpv\" (UniqueName: \"kubernetes.io/projected/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-kube-api-access-bqcpv\") pod \"nova-api-0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.449314 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-config-data\") pod \"nova-api-0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.449333 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.449385 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-logs\") pod \"nova-api-0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.449782 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.551121 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqcpv\" (UniqueName: \"kubernetes.io/projected/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-kube-api-access-bqcpv\") pod \"nova-api-0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.551181 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.551196 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-config-data\") pod \"nova-api-0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.551258 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-logs\") pod \"nova-api-0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.551852 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-logs\") pod \"nova-api-0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.557807 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.561648 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-config-data\") pod \"nova-api-0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.562784 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.565185 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.585867 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqcpv\" (UniqueName: \"kubernetes.io/projected/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-kube-api-access-bqcpv\") pod \"nova-api-0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.590346 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.592526 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.599303 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.614433 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.617630 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.644176 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.644297 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.753994 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6zlx\" (UniqueName: \"kubernetes.io/projected/7203b75a-e8db-46d7-a80d-ae72b3e6b411-kube-api-access-b6zlx\") pod \"nova-scheduler-0\" (UID: \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.754231 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7203b75a-e8db-46d7-a80d-ae72b3e6b411-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.754270 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7203b75a-e8db-46d7-a80d-ae72b3e6b411-config-data\") pod \"nova-scheduler-0\" (UID: \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.782172 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b54b85-0b61-4332-869c-1f2f8c07b4fb" path="/var/lib/kubelet/pods/20b54b85-0b61-4332-869c-1f2f8c07b4fb/volumes" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.782750 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78473f6a-8832-40d9-8d3d-0ea8ee59c1f3" path="/var/lib/kubelet/pods/78473f6a-8832-40d9-8d3d-0ea8ee59c1f3/volumes" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.855465 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6zlx\" (UniqueName: \"kubernetes.io/projected/7203b75a-e8db-46d7-a80d-ae72b3e6b411-kube-api-access-b6zlx\") pod \"nova-scheduler-0\" (UID: \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.856327 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7203b75a-e8db-46d7-a80d-ae72b3e6b411-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.856413 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7203b75a-e8db-46d7-a80d-ae72b3e6b411-config-data\") pod \"nova-scheduler-0\" (UID: \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.862596 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7203b75a-e8db-46d7-a80d-ae72b3e6b411-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.873218 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7203b75a-e8db-46d7-a80d-ae72b3e6b411-config-data\") pod \"nova-scheduler-0\" (UID: \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:23 crc kubenswrapper[4775]: I1002 02:02:23.880481 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6zlx\" (UniqueName: \"kubernetes.io/projected/7203b75a-e8db-46d7-a80d-ae72b3e6b411-kube-api-access-b6zlx\") pod \"nova-scheduler-0\" (UID: \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\") " pod="openstack/nova-scheduler-0" Oct 02 02:02:24 crc kubenswrapper[4775]: I1002 02:02:24.041020 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:02:24 crc kubenswrapper[4775]: I1002 02:02:24.099100 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:24 crc kubenswrapper[4775]: W1002 02:02:24.111932 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podefa9cb7f_f30f_49d9_96f9_f00f9f23dbe0.slice/crio-93200b862398c218f8bdc30ce974edfad64a57f6508e4407de9ed6283e79ec16 WatchSource:0}: Error finding container 93200b862398c218f8bdc30ce974edfad64a57f6508e4407de9ed6283e79ec16: Status 404 returned error can't find the container with id 93200b862398c218f8bdc30ce974edfad64a57f6508e4407de9ed6283e79ec16 Oct 02 02:02:24 crc kubenswrapper[4775]: I1002 02:02:24.238590 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0","Type":"ContainerStarted","Data":"93200b862398c218f8bdc30ce974edfad64a57f6508e4407de9ed6283e79ec16"} Oct 02 02:02:24 crc kubenswrapper[4775]: I1002 02:02:24.246363 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47865629-abc2-4da9-94d7-a0d389f9570c","Type":"ContainerStarted","Data":"01b3fe72cdb48add7470fbc8ed1e2751751aea48e865723d88cc83c45a4f11d4"} Oct 02 02:02:24 crc kubenswrapper[4775]: I1002 02:02:24.505153 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:02:24 crc kubenswrapper[4775]: W1002 02:02:24.515547 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7203b75a_e8db_46d7_a80d_ae72b3e6b411.slice/crio-d5f747ad678cec38c9271d36ae4713e9188419662aeb3a1c5d570ca9f39c59d8 WatchSource:0}: Error finding container d5f747ad678cec38c9271d36ae4713e9188419662aeb3a1c5d570ca9f39c59d8: Status 404 returned error can't find the container with id d5f747ad678cec38c9271d36ae4713e9188419662aeb3a1c5d570ca9f39c59d8 Oct 02 02:02:25 crc kubenswrapper[4775]: I1002 02:02:25.262108 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7203b75a-e8db-46d7-a80d-ae72b3e6b411","Type":"ContainerStarted","Data":"727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a"} Oct 02 02:02:25 crc kubenswrapper[4775]: I1002 02:02:25.262742 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7203b75a-e8db-46d7-a80d-ae72b3e6b411","Type":"ContainerStarted","Data":"d5f747ad678cec38c9271d36ae4713e9188419662aeb3a1c5d570ca9f39c59d8"} Oct 02 02:02:25 crc kubenswrapper[4775]: I1002 02:02:25.263890 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0","Type":"ContainerStarted","Data":"1027f009f85bc9a134913d0a2ef631ba37aa5937bca451d499576333b28a7ab7"} Oct 02 02:02:25 crc kubenswrapper[4775]: I1002 02:02:25.263936 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0","Type":"ContainerStarted","Data":"f21e9214eb3f724ffd1fb23bd48cb675fe7032fdd08ebb34daa2d4f382ca3e63"} Oct 02 02:02:25 crc kubenswrapper[4775]: I1002 02:02:25.302415 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.302393509 podStartE2EDuration="2.302393509s" podCreationTimestamp="2025-10-02 02:02:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:25.299814954 +0000 UTC m=+1282.466559054" watchObservedRunningTime="2025-10-02 02:02:25.302393509 +0000 UTC m=+1282.469137549" Oct 02 02:02:25 crc kubenswrapper[4775]: I1002 02:02:25.304174 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.304164284 podStartE2EDuration="2.304164284s" podCreationTimestamp="2025-10-02 02:02:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:25.284248208 +0000 UTC m=+1282.450992278" watchObservedRunningTime="2025-10-02 02:02:25.304164284 +0000 UTC m=+1282.470908324" Oct 02 02:02:26 crc kubenswrapper[4775]: I1002 02:02:26.278618 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47865629-abc2-4da9-94d7-a0d389f9570c","Type":"ContainerStarted","Data":"f3bfa122906ccd823316624f30a5cab42c24e1ed44f513d06f0b75836ce957df"} Oct 02 02:02:26 crc kubenswrapper[4775]: I1002 02:02:26.279029 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:02:26 crc kubenswrapper[4775]: I1002 02:02:26.303868 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.401825369 podStartE2EDuration="6.303849384s" podCreationTimestamp="2025-10-02 02:02:20 +0000 UTC" firstStartedPulling="2025-10-02 02:02:21.379890273 +0000 UTC m=+1278.546634313" lastFinishedPulling="2025-10-02 02:02:25.281914288 +0000 UTC m=+1282.448658328" observedRunningTime="2025-10-02 02:02:26.296188199 +0000 UTC m=+1283.462932249" watchObservedRunningTime="2025-10-02 02:02:26.303849384 +0000 UTC m=+1283.470593434" Oct 02 02:02:28 crc kubenswrapper[4775]: I1002 02:02:28.616282 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 02 02:02:28 crc kubenswrapper[4775]: I1002 02:02:28.645068 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 02:02:28 crc kubenswrapper[4775]: I1002 02:02:28.645105 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 02:02:29 crc kubenswrapper[4775]: I1002 02:02:29.041464 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 02:02:29 crc kubenswrapper[4775]: I1002 02:02:29.665192 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 02:02:29 crc kubenswrapper[4775]: I1002 02:02:29.665225 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 02:02:33 crc kubenswrapper[4775]: I1002 02:02:33.621152 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 02:02:33 crc kubenswrapper[4775]: I1002 02:02:33.621646 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 02:02:34 crc kubenswrapper[4775]: I1002 02:02:34.041396 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 02:02:34 crc kubenswrapper[4775]: I1002 02:02:34.091613 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 02:02:34 crc kubenswrapper[4775]: I1002 02:02:34.447911 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 02:02:34 crc kubenswrapper[4775]: I1002 02:02:34.705107 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 02:02:34 crc kubenswrapper[4775]: I1002 02:02:34.705836 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 02:02:37 crc kubenswrapper[4775]: I1002 02:02:37.233382 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:02:37 crc kubenswrapper[4775]: I1002 02:02:37.234140 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:02:38 crc kubenswrapper[4775]: I1002 02:02:38.648617 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 02:02:38 crc kubenswrapper[4775]: I1002 02:02:38.652033 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 02:02:38 crc kubenswrapper[4775]: I1002 02:02:38.652831 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 02:02:39 crc kubenswrapper[4775]: I1002 02:02:39.474006 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 02:02:42 crc kubenswrapper[4775]: I1002 02:02:42.540398 4775 generic.go:334] "Generic (PLEG): container finished" podID="8e6e2e8f-33ab-4eea-9c51-471533aa9f4a" containerID="6551e9c5f62eb1393127c59a3b81e28867173c615ce4b3cc689c3b24c62e2044" exitCode=137 Oct 02 02:02:42 crc kubenswrapper[4775]: I1002 02:02:42.541021 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a","Type":"ContainerDied","Data":"6551e9c5f62eb1393127c59a3b81e28867173c615ce4b3cc689c3b24c62e2044"} Oct 02 02:02:42 crc kubenswrapper[4775]: I1002 02:02:42.541050 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a","Type":"ContainerDied","Data":"da4ab5d083bd8f445a3ec0d0bc84aaaebe5e998b90fdbe9008988e92bf637884"} Oct 02 02:02:42 crc kubenswrapper[4775]: I1002 02:02:42.541061 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da4ab5d083bd8f445a3ec0d0bc84aaaebe5e998b90fdbe9008988e92bf637884" Oct 02 02:02:42 crc kubenswrapper[4775]: I1002 02:02:42.585374 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:42 crc kubenswrapper[4775]: I1002 02:02:42.656690 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-695st\" (UniqueName: \"kubernetes.io/projected/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-kube-api-access-695st\") pod \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\" (UID: \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\") " Oct 02 02:02:42 crc kubenswrapper[4775]: I1002 02:02:42.656875 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-combined-ca-bundle\") pod \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\" (UID: \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\") " Oct 02 02:02:42 crc kubenswrapper[4775]: I1002 02:02:42.656975 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-config-data\") pod \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\" (UID: \"8e6e2e8f-33ab-4eea-9c51-471533aa9f4a\") " Oct 02 02:02:42 crc kubenswrapper[4775]: I1002 02:02:42.664466 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-kube-api-access-695st" (OuterVolumeSpecName: "kube-api-access-695st") pod "8e6e2e8f-33ab-4eea-9c51-471533aa9f4a" (UID: "8e6e2e8f-33ab-4eea-9c51-471533aa9f4a"). InnerVolumeSpecName "kube-api-access-695st". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:42 crc kubenswrapper[4775]: I1002 02:02:42.707555 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-config-data" (OuterVolumeSpecName: "config-data") pod "8e6e2e8f-33ab-4eea-9c51-471533aa9f4a" (UID: "8e6e2e8f-33ab-4eea-9c51-471533aa9f4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:42 crc kubenswrapper[4775]: I1002 02:02:42.707606 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e6e2e8f-33ab-4eea-9c51-471533aa9f4a" (UID: "8e6e2e8f-33ab-4eea-9c51-471533aa9f4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:42 crc kubenswrapper[4775]: I1002 02:02:42.758938 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-695st\" (UniqueName: \"kubernetes.io/projected/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-kube-api-access-695st\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:42 crc kubenswrapper[4775]: I1002 02:02:42.759005 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:42 crc kubenswrapper[4775]: I1002 02:02:42.759020 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.553712 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.608065 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.623366 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.634067 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.636435 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.655436 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.665540 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:02:43 crc kubenswrapper[4775]: E1002 02:02:43.666220 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e6e2e8f-33ab-4eea-9c51-471533aa9f4a" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.666249 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e6e2e8f-33ab-4eea-9c51-471533aa9f4a" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.666588 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e6e2e8f-33ab-4eea-9c51-471533aa9f4a" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.667515 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.669803 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.674018 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.674291 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.674342 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.678199 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.780912 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qfwn\" (UniqueName: \"kubernetes.io/projected/a5c60336-6e25-4c68-98e6-52801f34578b-kube-api-access-2qfwn\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.781012 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.781049 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.781444 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.781564 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.781595 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e6e2e8f-33ab-4eea-9c51-471533aa9f4a" path="/var/lib/kubelet/pods/8e6e2e8f-33ab-4eea-9c51-471533aa9f4a/volumes" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.883648 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qfwn\" (UniqueName: \"kubernetes.io/projected/a5c60336-6e25-4c68-98e6-52801f34578b-kube-api-access-2qfwn\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.883721 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.883789 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.883815 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.883843 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.891482 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.892145 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.892794 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.898790 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:43 crc kubenswrapper[4775]: I1002 02:02:43.905208 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qfwn\" (UniqueName: \"kubernetes.io/projected/a5c60336-6e25-4c68-98e6-52801f34578b-kube-api-access-2qfwn\") pod \"nova-cell1-novncproxy-0\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.000555 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.366746 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.564963 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a5c60336-6e25-4c68-98e6-52801f34578b","Type":"ContainerStarted","Data":"dbe64d055f38f78c1200ca47b2599fd920af93a48d37676c014b7689ac274217"} Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.564998 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.568443 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.746808 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-8jbsr"] Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.751094 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.768701 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-8jbsr"] Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.807009 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.807098 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.807118 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.807141 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt4xd\" (UniqueName: \"kubernetes.io/projected/1364df9e-151f-40d0-9a5a-606a54021ef7-kube-api-access-rt4xd\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.807156 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-config\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.807173 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.908688 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.908802 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.908828 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.908859 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt4xd\" (UniqueName: \"kubernetes.io/projected/1364df9e-151f-40d0-9a5a-606a54021ef7-kube-api-access-rt4xd\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.908880 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-config\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.908905 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.909889 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.910026 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.910054 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.910422 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.910586 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-config\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:44 crc kubenswrapper[4775]: I1002 02:02:44.929379 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt4xd\" (UniqueName: \"kubernetes.io/projected/1364df9e-151f-40d0-9a5a-606a54021ef7-kube-api-access-rt4xd\") pod \"dnsmasq-dns-59cf4bdb65-8jbsr\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:45 crc kubenswrapper[4775]: I1002 02:02:45.076274 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:45 crc kubenswrapper[4775]: I1002 02:02:45.535901 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-8jbsr"] Oct 02 02:02:45 crc kubenswrapper[4775]: I1002 02:02:45.593204 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a5c60336-6e25-4c68-98e6-52801f34578b","Type":"ContainerStarted","Data":"38c453bdb403040517a2812e8643acfa55f8124baa338e04c4028a69bd4cbbd1"} Oct 02 02:02:45 crc kubenswrapper[4775]: I1002 02:02:45.596036 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" event={"ID":"1364df9e-151f-40d0-9a5a-606a54021ef7","Type":"ContainerStarted","Data":"5da6a552854ebc54e69c6124436167ea7d5496a4afc985a7bb27725f03a1dc6e"} Oct 02 02:02:45 crc kubenswrapper[4775]: I1002 02:02:45.612683 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.612666238 podStartE2EDuration="2.612666238s" podCreationTimestamp="2025-10-02 02:02:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:45.606127982 +0000 UTC m=+1302.772872022" watchObservedRunningTime="2025-10-02 02:02:45.612666238 +0000 UTC m=+1302.779410278" Oct 02 02:02:46 crc kubenswrapper[4775]: I1002 02:02:46.606338 4775 generic.go:334] "Generic (PLEG): container finished" podID="1364df9e-151f-40d0-9a5a-606a54021ef7" containerID="5ed104134ed32605e39b2c63eeab848fad590d9ebf90dc94c858b92bc55ff88b" exitCode=0 Oct 02 02:02:46 crc kubenswrapper[4775]: I1002 02:02:46.606440 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" event={"ID":"1364df9e-151f-40d0-9a5a-606a54021ef7","Type":"ContainerDied","Data":"5ed104134ed32605e39b2c63eeab848fad590d9ebf90dc94c858b92bc55ff88b"} Oct 02 02:02:46 crc kubenswrapper[4775]: I1002 02:02:46.716517 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:02:46 crc kubenswrapper[4775]: I1002 02:02:46.716812 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="ceilometer-central-agent" containerID="cri-o://6e2f3b70a5c27a2136fce6a4efd75732be583ec15cbf867ee9f595b46f1953ea" gracePeriod=30 Oct 02 02:02:46 crc kubenswrapper[4775]: I1002 02:02:46.716905 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="sg-core" containerID="cri-o://01b3fe72cdb48add7470fbc8ed1e2751751aea48e865723d88cc83c45a4f11d4" gracePeriod=30 Oct 02 02:02:46 crc kubenswrapper[4775]: I1002 02:02:46.716921 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="ceilometer-notification-agent" containerID="cri-o://e15f9bb8da4554af6844bb4153fc54e557917adfaafc48f55338ba8ccfa96018" gracePeriod=30 Oct 02 02:02:46 crc kubenswrapper[4775]: I1002 02:02:46.717049 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="proxy-httpd" containerID="cri-o://f3bfa122906ccd823316624f30a5cab42c24e1ed44f513d06f0b75836ce957df" gracePeriod=30 Oct 02 02:02:46 crc kubenswrapper[4775]: I1002 02:02:46.737130 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.194:3000/\": EOF" Oct 02 02:02:47 crc kubenswrapper[4775]: I1002 02:02:47.239687 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:47 crc kubenswrapper[4775]: I1002 02:02:47.620821 4775 generic.go:334] "Generic (PLEG): container finished" podID="47865629-abc2-4da9-94d7-a0d389f9570c" containerID="f3bfa122906ccd823316624f30a5cab42c24e1ed44f513d06f0b75836ce957df" exitCode=0 Oct 02 02:02:47 crc kubenswrapper[4775]: I1002 02:02:47.620854 4775 generic.go:334] "Generic (PLEG): container finished" podID="47865629-abc2-4da9-94d7-a0d389f9570c" containerID="01b3fe72cdb48add7470fbc8ed1e2751751aea48e865723d88cc83c45a4f11d4" exitCode=2 Oct 02 02:02:47 crc kubenswrapper[4775]: I1002 02:02:47.620861 4775 generic.go:334] "Generic (PLEG): container finished" podID="47865629-abc2-4da9-94d7-a0d389f9570c" containerID="6e2f3b70a5c27a2136fce6a4efd75732be583ec15cbf867ee9f595b46f1953ea" exitCode=0 Oct 02 02:02:47 crc kubenswrapper[4775]: I1002 02:02:47.620901 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47865629-abc2-4da9-94d7-a0d389f9570c","Type":"ContainerDied","Data":"f3bfa122906ccd823316624f30a5cab42c24e1ed44f513d06f0b75836ce957df"} Oct 02 02:02:47 crc kubenswrapper[4775]: I1002 02:02:47.621034 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47865629-abc2-4da9-94d7-a0d389f9570c","Type":"ContainerDied","Data":"01b3fe72cdb48add7470fbc8ed1e2751751aea48e865723d88cc83c45a4f11d4"} Oct 02 02:02:47 crc kubenswrapper[4775]: I1002 02:02:47.621093 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47865629-abc2-4da9-94d7-a0d389f9570c","Type":"ContainerDied","Data":"6e2f3b70a5c27a2136fce6a4efd75732be583ec15cbf867ee9f595b46f1953ea"} Oct 02 02:02:47 crc kubenswrapper[4775]: I1002 02:02:47.623123 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" event={"ID":"1364df9e-151f-40d0-9a5a-606a54021ef7","Type":"ContainerStarted","Data":"9bdbd014e9940be2ba8f1e074e43a05086b36eec14f5338e99629708c2efdca9"} Oct 02 02:02:47 crc kubenswrapper[4775]: I1002 02:02:47.623209 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" containerName="nova-api-log" containerID="cri-o://f21e9214eb3f724ffd1fb23bd48cb675fe7032fdd08ebb34daa2d4f382ca3e63" gracePeriod=30 Oct 02 02:02:47 crc kubenswrapper[4775]: I1002 02:02:47.623339 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" containerName="nova-api-api" containerID="cri-o://1027f009f85bc9a134913d0a2ef631ba37aa5937bca451d499576333b28a7ab7" gracePeriod=30 Oct 02 02:02:47 crc kubenswrapper[4775]: I1002 02:02:47.623561 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:47 crc kubenswrapper[4775]: I1002 02:02:47.646096 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" podStartSLOduration=3.646070102 podStartE2EDuration="3.646070102s" podCreationTimestamp="2025-10-02 02:02:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:47.644178044 +0000 UTC m=+1304.810922114" watchObservedRunningTime="2025-10-02 02:02:47.646070102 +0000 UTC m=+1304.812814182" Oct 02 02:02:48 crc kubenswrapper[4775]: I1002 02:02:48.634643 4775 generic.go:334] "Generic (PLEG): container finished" podID="efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" containerID="f21e9214eb3f724ffd1fb23bd48cb675fe7032fdd08ebb34daa2d4f382ca3e63" exitCode=143 Oct 02 02:02:48 crc kubenswrapper[4775]: I1002 02:02:48.634811 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0","Type":"ContainerDied","Data":"f21e9214eb3f724ffd1fb23bd48cb675fe7032fdd08ebb34daa2d4f382ca3e63"} Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.000874 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.184968 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.332440 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-sg-core-conf-yaml\") pod \"47865629-abc2-4da9-94d7-a0d389f9570c\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.332553 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-combined-ca-bundle\") pod \"47865629-abc2-4da9-94d7-a0d389f9570c\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.332590 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47865629-abc2-4da9-94d7-a0d389f9570c-log-httpd\") pod \"47865629-abc2-4da9-94d7-a0d389f9570c\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.332693 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pj6d\" (UniqueName: \"kubernetes.io/projected/47865629-abc2-4da9-94d7-a0d389f9570c-kube-api-access-7pj6d\") pod \"47865629-abc2-4da9-94d7-a0d389f9570c\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.332773 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-scripts\") pod \"47865629-abc2-4da9-94d7-a0d389f9570c\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.332808 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-ceilometer-tls-certs\") pod \"47865629-abc2-4da9-94d7-a0d389f9570c\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.332849 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-config-data\") pod \"47865629-abc2-4da9-94d7-a0d389f9570c\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.332903 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47865629-abc2-4da9-94d7-a0d389f9570c-run-httpd\") pod \"47865629-abc2-4da9-94d7-a0d389f9570c\" (UID: \"47865629-abc2-4da9-94d7-a0d389f9570c\") " Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.333385 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47865629-abc2-4da9-94d7-a0d389f9570c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "47865629-abc2-4da9-94d7-a0d389f9570c" (UID: "47865629-abc2-4da9-94d7-a0d389f9570c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.333743 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47865629-abc2-4da9-94d7-a0d389f9570c-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.333840 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47865629-abc2-4da9-94d7-a0d389f9570c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "47865629-abc2-4da9-94d7-a0d389f9570c" (UID: "47865629-abc2-4da9-94d7-a0d389f9570c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.339777 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47865629-abc2-4da9-94d7-a0d389f9570c-kube-api-access-7pj6d" (OuterVolumeSpecName: "kube-api-access-7pj6d") pod "47865629-abc2-4da9-94d7-a0d389f9570c" (UID: "47865629-abc2-4da9-94d7-a0d389f9570c"). InnerVolumeSpecName "kube-api-access-7pj6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.345103 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-scripts" (OuterVolumeSpecName: "scripts") pod "47865629-abc2-4da9-94d7-a0d389f9570c" (UID: "47865629-abc2-4da9-94d7-a0d389f9570c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.369712 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "47865629-abc2-4da9-94d7-a0d389f9570c" (UID: "47865629-abc2-4da9-94d7-a0d389f9570c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.398549 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "47865629-abc2-4da9-94d7-a0d389f9570c" (UID: "47865629-abc2-4da9-94d7-a0d389f9570c"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.435228 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pj6d\" (UniqueName: \"kubernetes.io/projected/47865629-abc2-4da9-94d7-a0d389f9570c-kube-api-access-7pj6d\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.435545 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.435703 4775 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.435819 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/47865629-abc2-4da9-94d7-a0d389f9570c-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.435930 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.439928 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-config-data" (OuterVolumeSpecName: "config-data") pod "47865629-abc2-4da9-94d7-a0d389f9570c" (UID: "47865629-abc2-4da9-94d7-a0d389f9570c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.469926 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "47865629-abc2-4da9-94d7-a0d389f9570c" (UID: "47865629-abc2-4da9-94d7-a0d389f9570c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.537832 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.537879 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47865629-abc2-4da9-94d7-a0d389f9570c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.646915 4775 generic.go:334] "Generic (PLEG): container finished" podID="47865629-abc2-4da9-94d7-a0d389f9570c" containerID="e15f9bb8da4554af6844bb4153fc54e557917adfaafc48f55338ba8ccfa96018" exitCode=0 Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.647014 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47865629-abc2-4da9-94d7-a0d389f9570c","Type":"ContainerDied","Data":"e15f9bb8da4554af6844bb4153fc54e557917adfaafc48f55338ba8ccfa96018"} Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.647065 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.647100 4775 scope.go:117] "RemoveContainer" containerID="f3bfa122906ccd823316624f30a5cab42c24e1ed44f513d06f0b75836ce957df" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.647082 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"47865629-abc2-4da9-94d7-a0d389f9570c","Type":"ContainerDied","Data":"d536196b445f1a21ca895b5a9063aaf4845b6b511f28bcb9a60e162278c267f0"} Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.672260 4775 scope.go:117] "RemoveContainer" containerID="01b3fe72cdb48add7470fbc8ed1e2751751aea48e865723d88cc83c45a4f11d4" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.700754 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.702461 4775 scope.go:117] "RemoveContainer" containerID="e15f9bb8da4554af6844bb4153fc54e557917adfaafc48f55338ba8ccfa96018" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.715059 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.726292 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:02:49 crc kubenswrapper[4775]: E1002 02:02:49.735631 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="ceilometer-notification-agent" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.735670 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="ceilometer-notification-agent" Oct 02 02:02:49 crc kubenswrapper[4775]: E1002 02:02:49.735716 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="proxy-httpd" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.735726 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="proxy-httpd" Oct 02 02:02:49 crc kubenswrapper[4775]: E1002 02:02:49.735751 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="ceilometer-central-agent" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.735763 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="ceilometer-central-agent" Oct 02 02:02:49 crc kubenswrapper[4775]: E1002 02:02:49.735790 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="sg-core" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.735800 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="sg-core" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.736215 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="ceilometer-notification-agent" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.736248 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="ceilometer-central-agent" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.736264 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="proxy-httpd" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.736284 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" containerName="sg-core" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.739928 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.741852 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.744815 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.745066 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.753365 4775 scope.go:117] "RemoveContainer" containerID="6e2f3b70a5c27a2136fce6a4efd75732be583ec15cbf867ee9f595b46f1953ea" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.755697 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.787137 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47865629-abc2-4da9-94d7-a0d389f9570c" path="/var/lib/kubelet/pods/47865629-abc2-4da9-94d7-a0d389f9570c/volumes" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.791361 4775 scope.go:117] "RemoveContainer" containerID="f3bfa122906ccd823316624f30a5cab42c24e1ed44f513d06f0b75836ce957df" Oct 02 02:02:49 crc kubenswrapper[4775]: E1002 02:02:49.791808 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3bfa122906ccd823316624f30a5cab42c24e1ed44f513d06f0b75836ce957df\": container with ID starting with f3bfa122906ccd823316624f30a5cab42c24e1ed44f513d06f0b75836ce957df not found: ID does not exist" containerID="f3bfa122906ccd823316624f30a5cab42c24e1ed44f513d06f0b75836ce957df" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.791841 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3bfa122906ccd823316624f30a5cab42c24e1ed44f513d06f0b75836ce957df"} err="failed to get container status \"f3bfa122906ccd823316624f30a5cab42c24e1ed44f513d06f0b75836ce957df\": rpc error: code = NotFound desc = could not find container \"f3bfa122906ccd823316624f30a5cab42c24e1ed44f513d06f0b75836ce957df\": container with ID starting with f3bfa122906ccd823316624f30a5cab42c24e1ed44f513d06f0b75836ce957df not found: ID does not exist" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.791864 4775 scope.go:117] "RemoveContainer" containerID="01b3fe72cdb48add7470fbc8ed1e2751751aea48e865723d88cc83c45a4f11d4" Oct 02 02:02:49 crc kubenswrapper[4775]: E1002 02:02:49.792360 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01b3fe72cdb48add7470fbc8ed1e2751751aea48e865723d88cc83c45a4f11d4\": container with ID starting with 01b3fe72cdb48add7470fbc8ed1e2751751aea48e865723d88cc83c45a4f11d4 not found: ID does not exist" containerID="01b3fe72cdb48add7470fbc8ed1e2751751aea48e865723d88cc83c45a4f11d4" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.792405 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01b3fe72cdb48add7470fbc8ed1e2751751aea48e865723d88cc83c45a4f11d4"} err="failed to get container status \"01b3fe72cdb48add7470fbc8ed1e2751751aea48e865723d88cc83c45a4f11d4\": rpc error: code = NotFound desc = could not find container \"01b3fe72cdb48add7470fbc8ed1e2751751aea48e865723d88cc83c45a4f11d4\": container with ID starting with 01b3fe72cdb48add7470fbc8ed1e2751751aea48e865723d88cc83c45a4f11d4 not found: ID does not exist" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.792452 4775 scope.go:117] "RemoveContainer" containerID="e15f9bb8da4554af6844bb4153fc54e557917adfaafc48f55338ba8ccfa96018" Oct 02 02:02:49 crc kubenswrapper[4775]: E1002 02:02:49.792794 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e15f9bb8da4554af6844bb4153fc54e557917adfaafc48f55338ba8ccfa96018\": container with ID starting with e15f9bb8da4554af6844bb4153fc54e557917adfaafc48f55338ba8ccfa96018 not found: ID does not exist" containerID="e15f9bb8da4554af6844bb4153fc54e557917adfaafc48f55338ba8ccfa96018" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.792823 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e15f9bb8da4554af6844bb4153fc54e557917adfaafc48f55338ba8ccfa96018"} err="failed to get container status \"e15f9bb8da4554af6844bb4153fc54e557917adfaafc48f55338ba8ccfa96018\": rpc error: code = NotFound desc = could not find container \"e15f9bb8da4554af6844bb4153fc54e557917adfaafc48f55338ba8ccfa96018\": container with ID starting with e15f9bb8da4554af6844bb4153fc54e557917adfaafc48f55338ba8ccfa96018 not found: ID does not exist" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.792837 4775 scope.go:117] "RemoveContainer" containerID="6e2f3b70a5c27a2136fce6a4efd75732be583ec15cbf867ee9f595b46f1953ea" Oct 02 02:02:49 crc kubenswrapper[4775]: E1002 02:02:49.793809 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e2f3b70a5c27a2136fce6a4efd75732be583ec15cbf867ee9f595b46f1953ea\": container with ID starting with 6e2f3b70a5c27a2136fce6a4efd75732be583ec15cbf867ee9f595b46f1953ea not found: ID does not exist" containerID="6e2f3b70a5c27a2136fce6a4efd75732be583ec15cbf867ee9f595b46f1953ea" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.793850 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e2f3b70a5c27a2136fce6a4efd75732be583ec15cbf867ee9f595b46f1953ea"} err="failed to get container status \"6e2f3b70a5c27a2136fce6a4efd75732be583ec15cbf867ee9f595b46f1953ea\": rpc error: code = NotFound desc = could not find container \"6e2f3b70a5c27a2136fce6a4efd75732be583ec15cbf867ee9f595b46f1953ea\": container with ID starting with 6e2f3b70a5c27a2136fce6a4efd75732be583ec15cbf867ee9f595b46f1953ea not found: ID does not exist" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.844520 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk2vw\" (UniqueName: \"kubernetes.io/projected/d2d45342-c4de-48db-8f6b-23f26103497b-kube-api-access-rk2vw\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.844920 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-config-data\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.845083 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.845181 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2d45342-c4de-48db-8f6b-23f26103497b-run-httpd\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.845242 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.845342 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.845488 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-scripts\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.845634 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2d45342-c4de-48db-8f6b-23f26103497b-log-httpd\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.947575 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.947646 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2d45342-c4de-48db-8f6b-23f26103497b-run-httpd\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.947680 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.947733 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.947784 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-scripts\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.947883 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2d45342-c4de-48db-8f6b-23f26103497b-log-httpd\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.948036 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk2vw\" (UniqueName: \"kubernetes.io/projected/d2d45342-c4de-48db-8f6b-23f26103497b-kube-api-access-rk2vw\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.948168 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-config-data\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.948894 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2d45342-c4de-48db-8f6b-23f26103497b-run-httpd\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.949683 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2d45342-c4de-48db-8f6b-23f26103497b-log-httpd\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.954337 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.955096 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-scripts\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.955586 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.956306 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-config-data\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.956499 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:49 crc kubenswrapper[4775]: I1002 02:02:49.964295 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk2vw\" (UniqueName: \"kubernetes.io/projected/d2d45342-c4de-48db-8f6b-23f26103497b-kube-api-access-rk2vw\") pod \"ceilometer-0\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " pod="openstack/ceilometer-0" Oct 02 02:02:50 crc kubenswrapper[4775]: I1002 02:02:50.081419 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:02:50 crc kubenswrapper[4775]: I1002 02:02:50.615918 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:02:50 crc kubenswrapper[4775]: I1002 02:02:50.662148 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2d45342-c4de-48db-8f6b-23f26103497b","Type":"ContainerStarted","Data":"251614ec9eafc9d0dc22c0b2e4a21f7e76e5fea8e660ef31288764935585dffd"} Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.276267 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.382524 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqcpv\" (UniqueName: \"kubernetes.io/projected/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-kube-api-access-bqcpv\") pod \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.382733 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-combined-ca-bundle\") pod \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.382802 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-config-data\") pod \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.382872 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-logs\") pod \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\" (UID: \"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0\") " Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.383419 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-logs" (OuterVolumeSpecName: "logs") pod "efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" (UID: "efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.387178 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-kube-api-access-bqcpv" (OuterVolumeSpecName: "kube-api-access-bqcpv") pod "efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" (UID: "efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0"). InnerVolumeSpecName "kube-api-access-bqcpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.408530 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" (UID: "efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.414096 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-config-data" (OuterVolumeSpecName: "config-data") pod "efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" (UID: "efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.485320 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.485358 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.485371 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.485382 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqcpv\" (UniqueName: \"kubernetes.io/projected/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0-kube-api-access-bqcpv\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.675940 4775 generic.go:334] "Generic (PLEG): container finished" podID="efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" containerID="1027f009f85bc9a134913d0a2ef631ba37aa5937bca451d499576333b28a7ab7" exitCode=0 Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.676046 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0","Type":"ContainerDied","Data":"1027f009f85bc9a134913d0a2ef631ba37aa5937bca451d499576333b28a7ab7"} Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.676087 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.676402 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0","Type":"ContainerDied","Data":"93200b862398c218f8bdc30ce974edfad64a57f6508e4407de9ed6283e79ec16"} Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.676438 4775 scope.go:117] "RemoveContainer" containerID="1027f009f85bc9a134913d0a2ef631ba37aa5937bca451d499576333b28a7ab7" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.679659 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2d45342-c4de-48db-8f6b-23f26103497b","Type":"ContainerStarted","Data":"8df9de2a0e96e291b49efa835c2476332824829ee66a90f08c5fa720d9e412f8"} Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.713524 4775 scope.go:117] "RemoveContainer" containerID="f21e9214eb3f724ffd1fb23bd48cb675fe7032fdd08ebb34daa2d4f382ca3e63" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.714567 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.731087 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.741299 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:51 crc kubenswrapper[4775]: E1002 02:02:51.741705 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" containerName="nova-api-api" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.741724 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" containerName="nova-api-api" Oct 02 02:02:51 crc kubenswrapper[4775]: E1002 02:02:51.741752 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" containerName="nova-api-log" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.741759 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" containerName="nova-api-log" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.741968 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" containerName="nova-api-api" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.741983 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" containerName="nova-api-log" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.742946 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.745657 4775 scope.go:117] "RemoveContainer" containerID="1027f009f85bc9a134913d0a2ef631ba37aa5937bca451d499576333b28a7ab7" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.745883 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 02:02:51 crc kubenswrapper[4775]: E1002 02:02:51.746091 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1027f009f85bc9a134913d0a2ef631ba37aa5937bca451d499576333b28a7ab7\": container with ID starting with 1027f009f85bc9a134913d0a2ef631ba37aa5937bca451d499576333b28a7ab7 not found: ID does not exist" containerID="1027f009f85bc9a134913d0a2ef631ba37aa5937bca451d499576333b28a7ab7" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.746137 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.746138 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1027f009f85bc9a134913d0a2ef631ba37aa5937bca451d499576333b28a7ab7"} err="failed to get container status \"1027f009f85bc9a134913d0a2ef631ba37aa5937bca451d499576333b28a7ab7\": rpc error: code = NotFound desc = could not find container \"1027f009f85bc9a134913d0a2ef631ba37aa5937bca451d499576333b28a7ab7\": container with ID starting with 1027f009f85bc9a134913d0a2ef631ba37aa5937bca451d499576333b28a7ab7 not found: ID does not exist" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.746165 4775 scope.go:117] "RemoveContainer" containerID="f21e9214eb3f724ffd1fb23bd48cb675fe7032fdd08ebb34daa2d4f382ca3e63" Oct 02 02:02:51 crc kubenswrapper[4775]: E1002 02:02:51.746395 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f21e9214eb3f724ffd1fb23bd48cb675fe7032fdd08ebb34daa2d4f382ca3e63\": container with ID starting with f21e9214eb3f724ffd1fb23bd48cb675fe7032fdd08ebb34daa2d4f382ca3e63 not found: ID does not exist" containerID="f21e9214eb3f724ffd1fb23bd48cb675fe7032fdd08ebb34daa2d4f382ca3e63" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.746424 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f21e9214eb3f724ffd1fb23bd48cb675fe7032fdd08ebb34daa2d4f382ca3e63"} err="failed to get container status \"f21e9214eb3f724ffd1fb23bd48cb675fe7032fdd08ebb34daa2d4f382ca3e63\": rpc error: code = NotFound desc = could not find container \"f21e9214eb3f724ffd1fb23bd48cb675fe7032fdd08ebb34daa2d4f382ca3e63\": container with ID starting with f21e9214eb3f724ffd1fb23bd48cb675fe7032fdd08ebb34daa2d4f382ca3e63 not found: ID does not exist" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.748588 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.749536 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.776614 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0" path="/var/lib/kubelet/pods/efa9cb7f-f30f-49d9-96f9-f00f9f23dbe0/volumes" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.893154 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-public-tls-certs\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.893224 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.893357 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxqgq\" (UniqueName: \"kubernetes.io/projected/a780ff81-6d87-472f-bd2d-ea144d385d21-kube-api-access-kxqgq\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.893573 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.893692 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-config-data\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.893768 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a780ff81-6d87-472f-bd2d-ea144d385d21-logs\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.995893 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.995969 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-config-data\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.996000 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a780ff81-6d87-472f-bd2d-ea144d385d21-logs\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.996032 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-public-tls-certs\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.996061 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.996121 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxqgq\" (UniqueName: \"kubernetes.io/projected/a780ff81-6d87-472f-bd2d-ea144d385d21-kube-api-access-kxqgq\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:51 crc kubenswrapper[4775]: I1002 02:02:51.996487 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a780ff81-6d87-472f-bd2d-ea144d385d21-logs\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:52 crc kubenswrapper[4775]: I1002 02:02:52.009114 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-config-data\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:52 crc kubenswrapper[4775]: I1002 02:02:52.009681 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:52 crc kubenswrapper[4775]: I1002 02:02:52.010333 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-public-tls-certs\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:52 crc kubenswrapper[4775]: I1002 02:02:52.013515 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:52 crc kubenswrapper[4775]: I1002 02:02:52.046582 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxqgq\" (UniqueName: \"kubernetes.io/projected/a780ff81-6d87-472f-bd2d-ea144d385d21-kube-api-access-kxqgq\") pod \"nova-api-0\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " pod="openstack/nova-api-0" Oct 02 02:02:52 crc kubenswrapper[4775]: I1002 02:02:52.062234 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:02:52 crc kubenswrapper[4775]: I1002 02:02:52.497436 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:02:52 crc kubenswrapper[4775]: W1002 02:02:52.504824 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda780ff81_6d87_472f_bd2d_ea144d385d21.slice/crio-7081390f64d64befdb6b204e02f300bd0e38deeab1f2f19b833a06bdf819df86 WatchSource:0}: Error finding container 7081390f64d64befdb6b204e02f300bd0e38deeab1f2f19b833a06bdf819df86: Status 404 returned error can't find the container with id 7081390f64d64befdb6b204e02f300bd0e38deeab1f2f19b833a06bdf819df86 Oct 02 02:02:52 crc kubenswrapper[4775]: I1002 02:02:52.692111 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a780ff81-6d87-472f-bd2d-ea144d385d21","Type":"ContainerStarted","Data":"7081390f64d64befdb6b204e02f300bd0e38deeab1f2f19b833a06bdf819df86"} Oct 02 02:02:52 crc kubenswrapper[4775]: I1002 02:02:52.695376 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2d45342-c4de-48db-8f6b-23f26103497b","Type":"ContainerStarted","Data":"4feb9154e0329bc25829a25b496b1b2573ea599dcffdc17eb82bbc2b54f58aa4"} Oct 02 02:02:53 crc kubenswrapper[4775]: I1002 02:02:53.712565 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a780ff81-6d87-472f-bd2d-ea144d385d21","Type":"ContainerStarted","Data":"23818f1c63bbf796afbf45847cd4ff13c2176069aea622f683a0012c32f678b6"} Oct 02 02:02:53 crc kubenswrapper[4775]: I1002 02:02:53.713105 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a780ff81-6d87-472f-bd2d-ea144d385d21","Type":"ContainerStarted","Data":"1855ca7be080db382445afa292d71c476f1d607185cc45ff0d377e9af7ca7c21"} Oct 02 02:02:53 crc kubenswrapper[4775]: I1002 02:02:53.717767 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2d45342-c4de-48db-8f6b-23f26103497b","Type":"ContainerStarted","Data":"0a11b7f88b7e2c013d1797c32a7ec78c81f775682c0323ea8b58d4f35f6e04b7"} Oct 02 02:02:53 crc kubenswrapper[4775]: I1002 02:02:53.754799 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.754777189 podStartE2EDuration="2.754777189s" podCreationTimestamp="2025-10-02 02:02:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:53.737507589 +0000 UTC m=+1310.904251669" watchObservedRunningTime="2025-10-02 02:02:53.754777189 +0000 UTC m=+1310.921521239" Oct 02 02:02:54 crc kubenswrapper[4775]: I1002 02:02:54.001251 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:54 crc kubenswrapper[4775]: I1002 02:02:54.019029 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:54 crc kubenswrapper[4775]: I1002 02:02:54.732275 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2d45342-c4de-48db-8f6b-23f26103497b","Type":"ContainerStarted","Data":"1c6c4af35ca368385ff971101cf668bbedef2ac9915012172cc65ff5519d556e"} Oct 02 02:02:54 crc kubenswrapper[4775]: I1002 02:02:54.733855 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 02:02:54 crc kubenswrapper[4775]: I1002 02:02:54.760891 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.005495916 podStartE2EDuration="5.760872631s" podCreationTimestamp="2025-10-02 02:02:49 +0000 UTC" firstStartedPulling="2025-10-02 02:02:50.628055725 +0000 UTC m=+1307.794799775" lastFinishedPulling="2025-10-02 02:02:54.38343244 +0000 UTC m=+1311.550176490" observedRunningTime="2025-10-02 02:02:54.752444046 +0000 UTC m=+1311.919188096" watchObservedRunningTime="2025-10-02 02:02:54.760872631 +0000 UTC m=+1311.927616681" Oct 02 02:02:54 crc kubenswrapper[4775]: I1002 02:02:54.765741 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:02:54 crc kubenswrapper[4775]: I1002 02:02:54.941166 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-dklqq"] Oct 02 02:02:54 crc kubenswrapper[4775]: I1002 02:02:54.942773 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:02:54 crc kubenswrapper[4775]: I1002 02:02:54.946520 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 02 02:02:54 crc kubenswrapper[4775]: I1002 02:02:54.948045 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 02 02:02:54 crc kubenswrapper[4775]: I1002 02:02:54.964264 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dklqq"] Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.071818 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-config-data\") pod \"nova-cell1-cell-mapping-dklqq\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.071913 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dklqq\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.071935 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-scripts\") pod \"nova-cell1-cell-mapping-dklqq\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.071973 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qj5z\" (UniqueName: \"kubernetes.io/projected/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-kube-api-access-7qj5z\") pod \"nova-cell1-cell-mapping-dklqq\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.078138 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.142176 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-6qqkn"] Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.142787 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" podUID="ebf491cd-4ab2-42a8-a53d-94193bfceb88" containerName="dnsmasq-dns" containerID="cri-o://00b230cb5807d0b8515a0dd77cc58a129ee4a0f812e8343c5cd23ba4ec66cfc9" gracePeriod=10 Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.174586 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dklqq\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.174632 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-scripts\") pod \"nova-cell1-cell-mapping-dklqq\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.174669 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qj5z\" (UniqueName: \"kubernetes.io/projected/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-kube-api-access-7qj5z\") pod \"nova-cell1-cell-mapping-dklqq\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.174885 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-config-data\") pod \"nova-cell1-cell-mapping-dklqq\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.184918 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-config-data\") pod \"nova-cell1-cell-mapping-dklqq\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.187463 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-dklqq\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.189239 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-scripts\") pod \"nova-cell1-cell-mapping-dklqq\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.209695 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qj5z\" (UniqueName: \"kubernetes.io/projected/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-kube-api-access-7qj5z\") pod \"nova-cell1-cell-mapping-dklqq\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.269667 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.647991 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.741300 4775 generic.go:334] "Generic (PLEG): container finished" podID="ebf491cd-4ab2-42a8-a53d-94193bfceb88" containerID="00b230cb5807d0b8515a0dd77cc58a129ee4a0f812e8343c5cd23ba4ec66cfc9" exitCode=0 Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.741353 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.741359 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" event={"ID":"ebf491cd-4ab2-42a8-a53d-94193bfceb88","Type":"ContainerDied","Data":"00b230cb5807d0b8515a0dd77cc58a129ee4a0f812e8343c5cd23ba4ec66cfc9"} Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.741391 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-6qqkn" event={"ID":"ebf491cd-4ab2-42a8-a53d-94193bfceb88","Type":"ContainerDied","Data":"4d0984404da4fcd2276b10a3b9ddba0f0e4b1e69eb142099539e0bad2cb966e6"} Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.741409 4775 scope.go:117] "RemoveContainer" containerID="00b230cb5807d0b8515a0dd77cc58a129ee4a0f812e8343c5cd23ba4ec66cfc9" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.764635 4775 scope.go:117] "RemoveContainer" containerID="a3f2f5fb1878415b74883963e59abf3f91c197081a8aec84bd142cc8c640ad9d" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.791133 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-dns-svc\") pod \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.791173 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-dns-swift-storage-0\") pod \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.791207 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-config\") pod \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.791337 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-ovsdbserver-nb\") pod \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.791370 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-ovsdbserver-sb\") pod \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.791405 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f6r4k\" (UniqueName: \"kubernetes.io/projected/ebf491cd-4ab2-42a8-a53d-94193bfceb88-kube-api-access-f6r4k\") pod \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\" (UID: \"ebf491cd-4ab2-42a8-a53d-94193bfceb88\") " Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.797627 4775 scope.go:117] "RemoveContainer" containerID="00b230cb5807d0b8515a0dd77cc58a129ee4a0f812e8343c5cd23ba4ec66cfc9" Oct 02 02:02:55 crc kubenswrapper[4775]: E1002 02:02:55.798308 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00b230cb5807d0b8515a0dd77cc58a129ee4a0f812e8343c5cd23ba4ec66cfc9\": container with ID starting with 00b230cb5807d0b8515a0dd77cc58a129ee4a0f812e8343c5cd23ba4ec66cfc9 not found: ID does not exist" containerID="00b230cb5807d0b8515a0dd77cc58a129ee4a0f812e8343c5cd23ba4ec66cfc9" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.798400 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00b230cb5807d0b8515a0dd77cc58a129ee4a0f812e8343c5cd23ba4ec66cfc9"} err="failed to get container status \"00b230cb5807d0b8515a0dd77cc58a129ee4a0f812e8343c5cd23ba4ec66cfc9\": rpc error: code = NotFound desc = could not find container \"00b230cb5807d0b8515a0dd77cc58a129ee4a0f812e8343c5cd23ba4ec66cfc9\": container with ID starting with 00b230cb5807d0b8515a0dd77cc58a129ee4a0f812e8343c5cd23ba4ec66cfc9 not found: ID does not exist" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.798486 4775 scope.go:117] "RemoveContainer" containerID="a3f2f5fb1878415b74883963e59abf3f91c197081a8aec84bd142cc8c640ad9d" Oct 02 02:02:55 crc kubenswrapper[4775]: E1002 02:02:55.798767 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3f2f5fb1878415b74883963e59abf3f91c197081a8aec84bd142cc8c640ad9d\": container with ID starting with a3f2f5fb1878415b74883963e59abf3f91c197081a8aec84bd142cc8c640ad9d not found: ID does not exist" containerID="a3f2f5fb1878415b74883963e59abf3f91c197081a8aec84bd142cc8c640ad9d" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.798805 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3f2f5fb1878415b74883963e59abf3f91c197081a8aec84bd142cc8c640ad9d"} err="failed to get container status \"a3f2f5fb1878415b74883963e59abf3f91c197081a8aec84bd142cc8c640ad9d\": rpc error: code = NotFound desc = could not find container \"a3f2f5fb1878415b74883963e59abf3f91c197081a8aec84bd142cc8c640ad9d\": container with ID starting with a3f2f5fb1878415b74883963e59abf3f91c197081a8aec84bd142cc8c640ad9d not found: ID does not exist" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.801151 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebf491cd-4ab2-42a8-a53d-94193bfceb88-kube-api-access-f6r4k" (OuterVolumeSpecName: "kube-api-access-f6r4k") pod "ebf491cd-4ab2-42a8-a53d-94193bfceb88" (UID: "ebf491cd-4ab2-42a8-a53d-94193bfceb88"). InnerVolumeSpecName "kube-api-access-f6r4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.846377 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-config" (OuterVolumeSpecName: "config") pod "ebf491cd-4ab2-42a8-a53d-94193bfceb88" (UID: "ebf491cd-4ab2-42a8-a53d-94193bfceb88"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.848879 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ebf491cd-4ab2-42a8-a53d-94193bfceb88" (UID: "ebf491cd-4ab2-42a8-a53d-94193bfceb88"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.864784 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ebf491cd-4ab2-42a8-a53d-94193bfceb88" (UID: "ebf491cd-4ab2-42a8-a53d-94193bfceb88"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.866942 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ebf491cd-4ab2-42a8-a53d-94193bfceb88" (UID: "ebf491cd-4ab2-42a8-a53d-94193bfceb88"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.867192 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-dklqq"] Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.870775 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ebf491cd-4ab2-42a8-a53d-94193bfceb88" (UID: "ebf491cd-4ab2-42a8-a53d-94193bfceb88"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.894504 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f6r4k\" (UniqueName: \"kubernetes.io/projected/ebf491cd-4ab2-42a8-a53d-94193bfceb88-kube-api-access-f6r4k\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.894530 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.894540 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.894550 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.894558 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:55 crc kubenswrapper[4775]: I1002 02:02:55.894568 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ebf491cd-4ab2-42a8-a53d-94193bfceb88-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:02:56 crc kubenswrapper[4775]: I1002 02:02:56.079796 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-6qqkn"] Oct 02 02:02:56 crc kubenswrapper[4775]: I1002 02:02:56.087713 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-6qqkn"] Oct 02 02:02:56 crc kubenswrapper[4775]: I1002 02:02:56.769075 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dklqq" event={"ID":"666eb22c-6c8b-4976-ae93-803c8b7cf7c6","Type":"ContainerStarted","Data":"de21c2812f8579db4805f66ccfa27b1e881d6fc1abd46304b474e22b41658e71"} Oct 02 02:02:56 crc kubenswrapper[4775]: I1002 02:02:56.769426 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dklqq" event={"ID":"666eb22c-6c8b-4976-ae93-803c8b7cf7c6","Type":"ContainerStarted","Data":"3cc8c0c4bda98bac23899a8e35ad612213fc2836c0679cf28178fdb2d637e831"} Oct 02 02:02:56 crc kubenswrapper[4775]: I1002 02:02:56.787234 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-dklqq" podStartSLOduration=2.787216854 podStartE2EDuration="2.787216854s" podCreationTimestamp="2025-10-02 02:02:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:02:56.783642803 +0000 UTC m=+1313.950386863" watchObservedRunningTime="2025-10-02 02:02:56.787216854 +0000 UTC m=+1313.953960894" Oct 02 02:02:57 crc kubenswrapper[4775]: I1002 02:02:57.784412 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebf491cd-4ab2-42a8-a53d-94193bfceb88" path="/var/lib/kubelet/pods/ebf491cd-4ab2-42a8-a53d-94193bfceb88/volumes" Oct 02 02:03:00 crc kubenswrapper[4775]: I1002 02:03:00.830205 4775 generic.go:334] "Generic (PLEG): container finished" podID="666eb22c-6c8b-4976-ae93-803c8b7cf7c6" containerID="de21c2812f8579db4805f66ccfa27b1e881d6fc1abd46304b474e22b41658e71" exitCode=0 Oct 02 02:03:00 crc kubenswrapper[4775]: I1002 02:03:00.830295 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dklqq" event={"ID":"666eb22c-6c8b-4976-ae93-803c8b7cf7c6","Type":"ContainerDied","Data":"de21c2812f8579db4805f66ccfa27b1e881d6fc1abd46304b474e22b41658e71"} Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.063242 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.063357 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.336273 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.439651 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-combined-ca-bundle\") pod \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.439798 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-scripts\") pod \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.439998 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-config-data\") pod \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.440092 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qj5z\" (UniqueName: \"kubernetes.io/projected/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-kube-api-access-7qj5z\") pod \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\" (UID: \"666eb22c-6c8b-4976-ae93-803c8b7cf7c6\") " Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.447575 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-kube-api-access-7qj5z" (OuterVolumeSpecName: "kube-api-access-7qj5z") pod "666eb22c-6c8b-4976-ae93-803c8b7cf7c6" (UID: "666eb22c-6c8b-4976-ae93-803c8b7cf7c6"). InnerVolumeSpecName "kube-api-access-7qj5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.465270 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-scripts" (OuterVolumeSpecName: "scripts") pod "666eb22c-6c8b-4976-ae93-803c8b7cf7c6" (UID: "666eb22c-6c8b-4976-ae93-803c8b7cf7c6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.469495 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-config-data" (OuterVolumeSpecName: "config-data") pod "666eb22c-6c8b-4976-ae93-803c8b7cf7c6" (UID: "666eb22c-6c8b-4976-ae93-803c8b7cf7c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.479004 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "666eb22c-6c8b-4976-ae93-803c8b7cf7c6" (UID: "666eb22c-6c8b-4976-ae93-803c8b7cf7c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.543218 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qj5z\" (UniqueName: \"kubernetes.io/projected/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-kube-api-access-7qj5z\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.543261 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.543273 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.543283 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/666eb22c-6c8b-4976-ae93-803c8b7cf7c6-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.859788 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-dklqq" event={"ID":"666eb22c-6c8b-4976-ae93-803c8b7cf7c6","Type":"ContainerDied","Data":"3cc8c0c4bda98bac23899a8e35ad612213fc2836c0679cf28178fdb2d637e831"} Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.859856 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cc8c0c4bda98bac23899a8e35ad612213fc2836c0679cf28178fdb2d637e831" Oct 02 02:03:02 crc kubenswrapper[4775]: I1002 02:03:02.859867 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-dklqq" Oct 02 02:03:03 crc kubenswrapper[4775]: I1002 02:03:03.045156 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:03:03 crc kubenswrapper[4775]: I1002 02:03:03.045396 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a780ff81-6d87-472f-bd2d-ea144d385d21" containerName="nova-api-log" containerID="cri-o://1855ca7be080db382445afa292d71c476f1d607185cc45ff0d377e9af7ca7c21" gracePeriod=30 Oct 02 02:03:03 crc kubenswrapper[4775]: I1002 02:03:03.045533 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a780ff81-6d87-472f-bd2d-ea144d385d21" containerName="nova-api-api" containerID="cri-o://23818f1c63bbf796afbf45847cd4ff13c2176069aea622f683a0012c32f678b6" gracePeriod=30 Oct 02 02:03:03 crc kubenswrapper[4775]: I1002 02:03:03.055180 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a780ff81-6d87-472f-bd2d-ea144d385d21" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": EOF" Oct 02 02:03:03 crc kubenswrapper[4775]: I1002 02:03:03.055196 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="a780ff81-6d87-472f-bd2d-ea144d385d21" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.200:8774/\": EOF" Oct 02 02:03:03 crc kubenswrapper[4775]: I1002 02:03:03.124106 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:03:03 crc kubenswrapper[4775]: I1002 02:03:03.125402 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" containerName="nova-metadata-log" containerID="cri-o://5de23478e0ccd06671708d54ffe109e061c817a4c1d2f255498395c9c5eda273" gracePeriod=30 Oct 02 02:03:03 crc kubenswrapper[4775]: I1002 02:03:03.125599 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" containerName="nova-metadata-metadata" containerID="cri-o://f68aaa46749cbbb1fff57d01d3426721b116a24cc8c1b3ee14b7fd2f36824d8f" gracePeriod=30 Oct 02 02:03:03 crc kubenswrapper[4775]: I1002 02:03:03.142016 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:03:03 crc kubenswrapper[4775]: I1002 02:03:03.142627 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7203b75a-e8db-46d7-a80d-ae72b3e6b411" containerName="nova-scheduler-scheduler" containerID="cri-o://727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a" gracePeriod=30 Oct 02 02:03:03 crc kubenswrapper[4775]: I1002 02:03:03.870783 4775 generic.go:334] "Generic (PLEG): container finished" podID="a780ff81-6d87-472f-bd2d-ea144d385d21" containerID="1855ca7be080db382445afa292d71c476f1d607185cc45ff0d377e9af7ca7c21" exitCode=143 Oct 02 02:03:03 crc kubenswrapper[4775]: I1002 02:03:03.871056 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a780ff81-6d87-472f-bd2d-ea144d385d21","Type":"ContainerDied","Data":"1855ca7be080db382445afa292d71c476f1d607185cc45ff0d377e9af7ca7c21"} Oct 02 02:03:03 crc kubenswrapper[4775]: I1002 02:03:03.872729 4775 generic.go:334] "Generic (PLEG): container finished" podID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" containerID="5de23478e0ccd06671708d54ffe109e061c817a4c1d2f255498395c9c5eda273" exitCode=143 Oct 02 02:03:03 crc kubenswrapper[4775]: I1002 02:03:03.872826 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f59877f-b40e-4585-a6a0-b22a8cf885eb","Type":"ContainerDied","Data":"5de23478e0ccd06671708d54ffe109e061c817a4c1d2f255498395c9c5eda273"} Oct 02 02:03:04 crc kubenswrapper[4775]: E1002 02:03:04.043186 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 02:03:04 crc kubenswrapper[4775]: E1002 02:03:04.044514 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 02:03:04 crc kubenswrapper[4775]: E1002 02:03:04.045771 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 02:03:04 crc kubenswrapper[4775]: E1002 02:03:04.045861 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7203b75a-e8db-46d7-a80d-ae72b3e6b411" containerName="nova-scheduler-scheduler" Oct 02 02:03:06 crc kubenswrapper[4775]: I1002 02:03:06.271791 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": read tcp 10.217.0.2:34190->10.217.0.193:8775: read: connection reset by peer" Oct 02 02:03:06 crc kubenswrapper[4775]: I1002 02:03:06.272804 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": read tcp 10.217.0.2:34186->10.217.0.193:8775: read: connection reset by peer" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.766036 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.912570 4775 generic.go:334] "Generic (PLEG): container finished" podID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" containerID="f68aaa46749cbbb1fff57d01d3426721b116a24cc8c1b3ee14b7fd2f36824d8f" exitCode=0 Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.912627 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f59877f-b40e-4585-a6a0-b22a8cf885eb","Type":"ContainerDied","Data":"f68aaa46749cbbb1fff57d01d3426721b116a24cc8c1b3ee14b7fd2f36824d8f"} Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.912670 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f59877f-b40e-4585-a6a0-b22a8cf885eb","Type":"ContainerDied","Data":"28ab43a939e893a3d6986354846470f9bd135dfd2190d23da9e2504608e69600"} Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.912667 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.912687 4775 scope.go:117] "RemoveContainer" containerID="f68aaa46749cbbb1fff57d01d3426721b116a24cc8c1b3ee14b7fd2f36824d8f" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.937060 4775 scope.go:117] "RemoveContainer" containerID="5de23478e0ccd06671708d54ffe109e061c817a4c1d2f255498395c9c5eda273" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.955074 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-nova-metadata-tls-certs\") pod \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.955128 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7rtk\" (UniqueName: \"kubernetes.io/projected/6f59877f-b40e-4585-a6a0-b22a8cf885eb-kube-api-access-l7rtk\") pod \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.955230 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-config-data\") pod \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.955277 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f59877f-b40e-4585-a6a0-b22a8cf885eb-logs\") pod \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.955307 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-combined-ca-bundle\") pod \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\" (UID: \"6f59877f-b40e-4585-a6a0-b22a8cf885eb\") " Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.956187 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6f59877f-b40e-4585-a6a0-b22a8cf885eb-logs" (OuterVolumeSpecName: "logs") pod "6f59877f-b40e-4585-a6a0-b22a8cf885eb" (UID: "6f59877f-b40e-4585-a6a0-b22a8cf885eb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.961612 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f59877f-b40e-4585-a6a0-b22a8cf885eb-kube-api-access-l7rtk" (OuterVolumeSpecName: "kube-api-access-l7rtk") pod "6f59877f-b40e-4585-a6a0-b22a8cf885eb" (UID: "6f59877f-b40e-4585-a6a0-b22a8cf885eb"). InnerVolumeSpecName "kube-api-access-l7rtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.961836 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7rtk\" (UniqueName: \"kubernetes.io/projected/6f59877f-b40e-4585-a6a0-b22a8cf885eb-kube-api-access-l7rtk\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.961849 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f59877f-b40e-4585-a6a0-b22a8cf885eb-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.970996 4775 scope.go:117] "RemoveContainer" containerID="f68aaa46749cbbb1fff57d01d3426721b116a24cc8c1b3ee14b7fd2f36824d8f" Oct 02 02:03:07 crc kubenswrapper[4775]: E1002 02:03:06.971537 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f68aaa46749cbbb1fff57d01d3426721b116a24cc8c1b3ee14b7fd2f36824d8f\": container with ID starting with f68aaa46749cbbb1fff57d01d3426721b116a24cc8c1b3ee14b7fd2f36824d8f not found: ID does not exist" containerID="f68aaa46749cbbb1fff57d01d3426721b116a24cc8c1b3ee14b7fd2f36824d8f" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.971579 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f68aaa46749cbbb1fff57d01d3426721b116a24cc8c1b3ee14b7fd2f36824d8f"} err="failed to get container status \"f68aaa46749cbbb1fff57d01d3426721b116a24cc8c1b3ee14b7fd2f36824d8f\": rpc error: code = NotFound desc = could not find container \"f68aaa46749cbbb1fff57d01d3426721b116a24cc8c1b3ee14b7fd2f36824d8f\": container with ID starting with f68aaa46749cbbb1fff57d01d3426721b116a24cc8c1b3ee14b7fd2f36824d8f not found: ID does not exist" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.971599 4775 scope.go:117] "RemoveContainer" containerID="5de23478e0ccd06671708d54ffe109e061c817a4c1d2f255498395c9c5eda273" Oct 02 02:03:07 crc kubenswrapper[4775]: E1002 02:03:06.971941 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5de23478e0ccd06671708d54ffe109e061c817a4c1d2f255498395c9c5eda273\": container with ID starting with 5de23478e0ccd06671708d54ffe109e061c817a4c1d2f255498395c9c5eda273 not found: ID does not exist" containerID="5de23478e0ccd06671708d54ffe109e061c817a4c1d2f255498395c9c5eda273" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.971979 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5de23478e0ccd06671708d54ffe109e061c817a4c1d2f255498395c9c5eda273"} err="failed to get container status \"5de23478e0ccd06671708d54ffe109e061c817a4c1d2f255498395c9c5eda273\": rpc error: code = NotFound desc = could not find container \"5de23478e0ccd06671708d54ffe109e061c817a4c1d2f255498395c9c5eda273\": container with ID starting with 5de23478e0ccd06671708d54ffe109e061c817a4c1d2f255498395c9c5eda273 not found: ID does not exist" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:06.995842 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f59877f-b40e-4585-a6a0-b22a8cf885eb" (UID: "6f59877f-b40e-4585-a6a0-b22a8cf885eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.011173 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-config-data" (OuterVolumeSpecName: "config-data") pod "6f59877f-b40e-4585-a6a0-b22a8cf885eb" (UID: "6f59877f-b40e-4585-a6a0-b22a8cf885eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.039012 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "6f59877f-b40e-4585-a6a0-b22a8cf885eb" (UID: "6f59877f-b40e-4585-a6a0-b22a8cf885eb"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.063647 4775 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.063675 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.063686 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f59877f-b40e-4585-a6a0-b22a8cf885eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.233357 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.233490 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.272878 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.285422 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.299817 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:03:07 crc kubenswrapper[4775]: E1002 02:03:07.300700 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf491cd-4ab2-42a8-a53d-94193bfceb88" containerName="init" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.300740 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf491cd-4ab2-42a8-a53d-94193bfceb88" containerName="init" Oct 02 02:03:07 crc kubenswrapper[4775]: E1002 02:03:07.300802 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" containerName="nova-metadata-metadata" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.300819 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" containerName="nova-metadata-metadata" Oct 02 02:03:07 crc kubenswrapper[4775]: E1002 02:03:07.300856 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="666eb22c-6c8b-4976-ae93-803c8b7cf7c6" containerName="nova-manage" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.300873 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="666eb22c-6c8b-4976-ae93-803c8b7cf7c6" containerName="nova-manage" Oct 02 02:03:07 crc kubenswrapper[4775]: E1002 02:03:07.300908 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" containerName="nova-metadata-log" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.300925 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" containerName="nova-metadata-log" Oct 02 02:03:07 crc kubenswrapper[4775]: E1002 02:03:07.300993 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebf491cd-4ab2-42a8-a53d-94193bfceb88" containerName="dnsmasq-dns" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.301009 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebf491cd-4ab2-42a8-a53d-94193bfceb88" containerName="dnsmasq-dns" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.301483 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebf491cd-4ab2-42a8-a53d-94193bfceb88" containerName="dnsmasq-dns" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.301528 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="666eb22c-6c8b-4976-ae93-803c8b7cf7c6" containerName="nova-manage" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.301567 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" containerName="nova-metadata-metadata" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.301599 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" containerName="nova-metadata-log" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.304448 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.308661 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.309049 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.326463 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.369922 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzh7k\" (UniqueName: \"kubernetes.io/projected/5e322679-b95d-4328-b03b-0be7f0e82ac3-kube-api-access-hzh7k\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.370005 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e322679-b95d-4328-b03b-0be7f0e82ac3-logs\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.370040 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.370124 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-config-data\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.370217 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.471215 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-config-data\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.471647 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.471817 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzh7k\" (UniqueName: \"kubernetes.io/projected/5e322679-b95d-4328-b03b-0be7f0e82ac3-kube-api-access-hzh7k\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.471924 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e322679-b95d-4328-b03b-0be7f0e82ac3-logs\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.472042 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.472702 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e322679-b95d-4328-b03b-0be7f0e82ac3-logs\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.477156 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.477408 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-config-data\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.479835 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.499592 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzh7k\" (UniqueName: \"kubernetes.io/projected/5e322679-b95d-4328-b03b-0be7f0e82ac3-kube-api-access-hzh7k\") pod \"nova-metadata-0\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.630586 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.780019 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f59877f-b40e-4585-a6a0-b22a8cf885eb" path="/var/lib/kubelet/pods/6f59877f-b40e-4585-a6a0-b22a8cf885eb/volumes" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.829869 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.937049 4775 generic.go:334] "Generic (PLEG): container finished" podID="7203b75a-e8db-46d7-a80d-ae72b3e6b411" containerID="727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a" exitCode=0 Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.937118 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.937137 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7203b75a-e8db-46d7-a80d-ae72b3e6b411","Type":"ContainerDied","Data":"727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a"} Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.937184 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7203b75a-e8db-46d7-a80d-ae72b3e6b411","Type":"ContainerDied","Data":"d5f747ad678cec38c9271d36ae4713e9188419662aeb3a1c5d570ca9f39c59d8"} Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.937205 4775 scope.go:117] "RemoveContainer" containerID="727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.965418 4775 scope.go:117] "RemoveContainer" containerID="727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a" Oct 02 02:03:07 crc kubenswrapper[4775]: E1002 02:03:07.965805 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a\": container with ID starting with 727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a not found: ID does not exist" containerID="727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.965869 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a"} err="failed to get container status \"727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a\": rpc error: code = NotFound desc = could not find container \"727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a\": container with ID starting with 727bc61f20aff7986016157d3a62caa84aeec122e5922f7aa43c8d34ad7c5c2a not found: ID does not exist" Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.982693 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7203b75a-e8db-46d7-a80d-ae72b3e6b411-config-data\") pod \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\" (UID: \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\") " Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.982857 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6zlx\" (UniqueName: \"kubernetes.io/projected/7203b75a-e8db-46d7-a80d-ae72b3e6b411-kube-api-access-b6zlx\") pod \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\" (UID: \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\") " Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.982994 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7203b75a-e8db-46d7-a80d-ae72b3e6b411-combined-ca-bundle\") pod \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\" (UID: \"7203b75a-e8db-46d7-a80d-ae72b3e6b411\") " Oct 02 02:03:07 crc kubenswrapper[4775]: I1002 02:03:07.988668 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7203b75a-e8db-46d7-a80d-ae72b3e6b411-kube-api-access-b6zlx" (OuterVolumeSpecName: "kube-api-access-b6zlx") pod "7203b75a-e8db-46d7-a80d-ae72b3e6b411" (UID: "7203b75a-e8db-46d7-a80d-ae72b3e6b411"). InnerVolumeSpecName "kube-api-access-b6zlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.025796 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7203b75a-e8db-46d7-a80d-ae72b3e6b411-config-data" (OuterVolumeSpecName: "config-data") pod "7203b75a-e8db-46d7-a80d-ae72b3e6b411" (UID: "7203b75a-e8db-46d7-a80d-ae72b3e6b411"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.029946 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7203b75a-e8db-46d7-a80d-ae72b3e6b411-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7203b75a-e8db-46d7-a80d-ae72b3e6b411" (UID: "7203b75a-e8db-46d7-a80d-ae72b3e6b411"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.086913 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7203b75a-e8db-46d7-a80d-ae72b3e6b411-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.086974 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6zlx\" (UniqueName: \"kubernetes.io/projected/7203b75a-e8db-46d7-a80d-ae72b3e6b411-kube-api-access-b6zlx\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.086986 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7203b75a-e8db-46d7-a80d-ae72b3e6b411-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.125313 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.314649 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.372280 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.380318 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:03:08 crc kubenswrapper[4775]: E1002 02:03:08.380987 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7203b75a-e8db-46d7-a80d-ae72b3e6b411" containerName="nova-scheduler-scheduler" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.381018 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7203b75a-e8db-46d7-a80d-ae72b3e6b411" containerName="nova-scheduler-scheduler" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.381308 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7203b75a-e8db-46d7-a80d-ae72b3e6b411" containerName="nova-scheduler-scheduler" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.382386 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.389135 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.395360 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.396244 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp9x5\" (UniqueName: \"kubernetes.io/projected/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-kube-api-access-tp9x5\") pod \"nova-scheduler-0\" (UID: \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\") " pod="openstack/nova-scheduler-0" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.396340 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\") " pod="openstack/nova-scheduler-0" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.396454 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-config-data\") pod \"nova-scheduler-0\" (UID: \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\") " pod="openstack/nova-scheduler-0" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.497755 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp9x5\" (UniqueName: \"kubernetes.io/projected/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-kube-api-access-tp9x5\") pod \"nova-scheduler-0\" (UID: \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\") " pod="openstack/nova-scheduler-0" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.497832 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\") " pod="openstack/nova-scheduler-0" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.497897 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-config-data\") pod \"nova-scheduler-0\" (UID: \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\") " pod="openstack/nova-scheduler-0" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.501813 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-config-data\") pod \"nova-scheduler-0\" (UID: \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\") " pod="openstack/nova-scheduler-0" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.505333 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\") " pod="openstack/nova-scheduler-0" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.519076 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp9x5\" (UniqueName: \"kubernetes.io/projected/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-kube-api-access-tp9x5\") pod \"nova-scheduler-0\" (UID: \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\") " pod="openstack/nova-scheduler-0" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.680534 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.770396 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.803258 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-public-tls-certs\") pod \"a780ff81-6d87-472f-bd2d-ea144d385d21\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.803692 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-config-data\") pod \"a780ff81-6d87-472f-bd2d-ea144d385d21\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.803746 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a780ff81-6d87-472f-bd2d-ea144d385d21-logs\") pod \"a780ff81-6d87-472f-bd2d-ea144d385d21\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.803783 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-internal-tls-certs\") pod \"a780ff81-6d87-472f-bd2d-ea144d385d21\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.803870 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kxqgq\" (UniqueName: \"kubernetes.io/projected/a780ff81-6d87-472f-bd2d-ea144d385d21-kube-api-access-kxqgq\") pod \"a780ff81-6d87-472f-bd2d-ea144d385d21\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.803903 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-combined-ca-bundle\") pod \"a780ff81-6d87-472f-bd2d-ea144d385d21\" (UID: \"a780ff81-6d87-472f-bd2d-ea144d385d21\") " Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.806210 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a780ff81-6d87-472f-bd2d-ea144d385d21-logs" (OuterVolumeSpecName: "logs") pod "a780ff81-6d87-472f-bd2d-ea144d385d21" (UID: "a780ff81-6d87-472f-bd2d-ea144d385d21"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.808736 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a780ff81-6d87-472f-bd2d-ea144d385d21-kube-api-access-kxqgq" (OuterVolumeSpecName: "kube-api-access-kxqgq") pod "a780ff81-6d87-472f-bd2d-ea144d385d21" (UID: "a780ff81-6d87-472f-bd2d-ea144d385d21"). InnerVolumeSpecName "kube-api-access-kxqgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.836524 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a780ff81-6d87-472f-bd2d-ea144d385d21" (UID: "a780ff81-6d87-472f-bd2d-ea144d385d21"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.851283 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a780ff81-6d87-472f-bd2d-ea144d385d21" (UID: "a780ff81-6d87-472f-bd2d-ea144d385d21"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.851320 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-config-data" (OuterVolumeSpecName: "config-data") pod "a780ff81-6d87-472f-bd2d-ea144d385d21" (UID: "a780ff81-6d87-472f-bd2d-ea144d385d21"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.870825 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a780ff81-6d87-472f-bd2d-ea144d385d21" (UID: "a780ff81-6d87-472f-bd2d-ea144d385d21"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.906672 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kxqgq\" (UniqueName: \"kubernetes.io/projected/a780ff81-6d87-472f-bd2d-ea144d385d21-kube-api-access-kxqgq\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.906709 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.906724 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.906766 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.906778 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a780ff81-6d87-472f-bd2d-ea144d385d21-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.906789 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a780ff81-6d87-472f-bd2d-ea144d385d21-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.951733 4775 generic.go:334] "Generic (PLEG): container finished" podID="a780ff81-6d87-472f-bd2d-ea144d385d21" containerID="23818f1c63bbf796afbf45847cd4ff13c2176069aea622f683a0012c32f678b6" exitCode=0 Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.951805 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a780ff81-6d87-472f-bd2d-ea144d385d21","Type":"ContainerDied","Data":"23818f1c63bbf796afbf45847cd4ff13c2176069aea622f683a0012c32f678b6"} Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.951822 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.951851 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a780ff81-6d87-472f-bd2d-ea144d385d21","Type":"ContainerDied","Data":"7081390f64d64befdb6b204e02f300bd0e38deeab1f2f19b833a06bdf819df86"} Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.951870 4775 scope.go:117] "RemoveContainer" containerID="23818f1c63bbf796afbf45847cd4ff13c2176069aea622f683a0012c32f678b6" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.953138 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e322679-b95d-4328-b03b-0be7f0e82ac3","Type":"ContainerStarted","Data":"33e28213f9c089646224401e310e6c70d1efdff65261c89fe379e544e09ae98f"} Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.953169 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e322679-b95d-4328-b03b-0be7f0e82ac3","Type":"ContainerStarted","Data":"0d717b8fb0cfdfcf5e5314de128b353f94395b4da93ebdbec0e1f666b4b81244"} Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.953178 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e322679-b95d-4328-b03b-0be7f0e82ac3","Type":"ContainerStarted","Data":"0b25a17cee3661f96b7c2e3eb90aced2226298c80f4b5f4c4e1f0f3062398aee"} Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.984178 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.9841559960000001 podStartE2EDuration="1.984155996s" podCreationTimestamp="2025-10-02 02:03:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:03:08.968340094 +0000 UTC m=+1326.135084144" watchObservedRunningTime="2025-10-02 02:03:08.984155996 +0000 UTC m=+1326.150900036" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.985384 4775 scope.go:117] "RemoveContainer" containerID="1855ca7be080db382445afa292d71c476f1d607185cc45ff0d377e9af7ca7c21" Oct 02 02:03:08 crc kubenswrapper[4775]: I1002 02:03:08.994621 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.004715 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.010654 4775 scope.go:117] "RemoveContainer" containerID="23818f1c63bbf796afbf45847cd4ff13c2176069aea622f683a0012c32f678b6" Oct 02 02:03:09 crc kubenswrapper[4775]: E1002 02:03:09.012126 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23818f1c63bbf796afbf45847cd4ff13c2176069aea622f683a0012c32f678b6\": container with ID starting with 23818f1c63bbf796afbf45847cd4ff13c2176069aea622f683a0012c32f678b6 not found: ID does not exist" containerID="23818f1c63bbf796afbf45847cd4ff13c2176069aea622f683a0012c32f678b6" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.012164 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23818f1c63bbf796afbf45847cd4ff13c2176069aea622f683a0012c32f678b6"} err="failed to get container status \"23818f1c63bbf796afbf45847cd4ff13c2176069aea622f683a0012c32f678b6\": rpc error: code = NotFound desc = could not find container \"23818f1c63bbf796afbf45847cd4ff13c2176069aea622f683a0012c32f678b6\": container with ID starting with 23818f1c63bbf796afbf45847cd4ff13c2176069aea622f683a0012c32f678b6 not found: ID does not exist" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.012192 4775 scope.go:117] "RemoveContainer" containerID="1855ca7be080db382445afa292d71c476f1d607185cc45ff0d377e9af7ca7c21" Oct 02 02:03:09 crc kubenswrapper[4775]: E1002 02:03:09.012768 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1855ca7be080db382445afa292d71c476f1d607185cc45ff0d377e9af7ca7c21\": container with ID starting with 1855ca7be080db382445afa292d71c476f1d607185cc45ff0d377e9af7ca7c21 not found: ID does not exist" containerID="1855ca7be080db382445afa292d71c476f1d607185cc45ff0d377e9af7ca7c21" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.012815 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1855ca7be080db382445afa292d71c476f1d607185cc45ff0d377e9af7ca7c21"} err="failed to get container status \"1855ca7be080db382445afa292d71c476f1d607185cc45ff0d377e9af7ca7c21\": rpc error: code = NotFound desc = could not find container \"1855ca7be080db382445afa292d71c476f1d607185cc45ff0d377e9af7ca7c21\": container with ID starting with 1855ca7be080db382445afa292d71c476f1d607185cc45ff0d377e9af7ca7c21 not found: ID does not exist" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.015462 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 02:03:09 crc kubenswrapper[4775]: E1002 02:03:09.015836 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a780ff81-6d87-472f-bd2d-ea144d385d21" containerName="nova-api-api" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.015852 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a780ff81-6d87-472f-bd2d-ea144d385d21" containerName="nova-api-api" Oct 02 02:03:09 crc kubenswrapper[4775]: E1002 02:03:09.015876 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a780ff81-6d87-472f-bd2d-ea144d385d21" containerName="nova-api-log" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.015882 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a780ff81-6d87-472f-bd2d-ea144d385d21" containerName="nova-api-log" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.016127 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a780ff81-6d87-472f-bd2d-ea144d385d21" containerName="nova-api-api" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.016151 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a780ff81-6d87-472f-bd2d-ea144d385d21" containerName="nova-api-log" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.017452 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.021139 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.021350 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.021424 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.034562 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.106248 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:03:09 crc kubenswrapper[4775]: W1002 02:03:09.106314 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda64f8ba3_65b9_489f_ace1_757f3ecce6b4.slice/crio-e3bfc96e50f98ba49539fa2c9cb90f53ad1f2a6bf089d7ba9643db31fd98274d WatchSource:0}: Error finding container e3bfc96e50f98ba49539fa2c9cb90f53ad1f2a6bf089d7ba9643db31fd98274d: Status 404 returned error can't find the container with id e3bfc96e50f98ba49539fa2c9cb90f53ad1f2a6bf089d7ba9643db31fd98274d Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.112538 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.112606 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmdwd\" (UniqueName: \"kubernetes.io/projected/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-kube-api-access-dmdwd\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.112641 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-public-tls-certs\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.112670 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-config-data\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.112718 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.112759 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-logs\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.214877 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.214940 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmdwd\" (UniqueName: \"kubernetes.io/projected/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-kube-api-access-dmdwd\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.214986 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-public-tls-certs\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.215016 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-config-data\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.215079 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.215113 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-logs\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.215649 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-logs\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.218725 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-public-tls-certs\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.223183 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-internal-tls-certs\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.223364 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.223479 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-config-data\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.233047 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmdwd\" (UniqueName: \"kubernetes.io/projected/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-kube-api-access-dmdwd\") pod \"nova-api-0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.335165 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.780405 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7203b75a-e8db-46d7-a80d-ae72b3e6b411" path="/var/lib/kubelet/pods/7203b75a-e8db-46d7-a80d-ae72b3e6b411/volumes" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.781523 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a780ff81-6d87-472f-bd2d-ea144d385d21" path="/var/lib/kubelet/pods/a780ff81-6d87-472f-bd2d-ea144d385d21/volumes" Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.840187 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.966221 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a64f8ba3-65b9-489f-ace1-757f3ecce6b4","Type":"ContainerStarted","Data":"cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa"} Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.966272 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a64f8ba3-65b9-489f-ace1-757f3ecce6b4","Type":"ContainerStarted","Data":"e3bfc96e50f98ba49539fa2c9cb90f53ad1f2a6bf089d7ba9643db31fd98274d"} Oct 02 02:03:09 crc kubenswrapper[4775]: I1002 02:03:09.967242 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0","Type":"ContainerStarted","Data":"bb2c755284b81a31e41cade55e5b66ce01b89e599614cc72c51530d3b154f91e"} Oct 02 02:03:10 crc kubenswrapper[4775]: I1002 02:03:10.023592 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.023561395 podStartE2EDuration="2.023561395s" podCreationTimestamp="2025-10-02 02:03:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:03:09.978755955 +0000 UTC m=+1327.145500035" watchObservedRunningTime="2025-10-02 02:03:10.023561395 +0000 UTC m=+1327.190305485" Oct 02 02:03:10 crc kubenswrapper[4775]: I1002 02:03:10.983913 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0","Type":"ContainerStarted","Data":"9d068bf0ca5c469b57dcad75c4e6ebff62f9badd694815d6349f36cd138a20a8"} Oct 02 02:03:10 crc kubenswrapper[4775]: I1002 02:03:10.986168 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0","Type":"ContainerStarted","Data":"303fb7d6dfef981bea20389d78f64a2cf286846122e5c5d3878e6b65fef9edb7"} Oct 02 02:03:11 crc kubenswrapper[4775]: I1002 02:03:11.012406 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.012390448 podStartE2EDuration="3.012390448s" podCreationTimestamp="2025-10-02 02:03:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:03:11.006915699 +0000 UTC m=+1328.173659739" watchObservedRunningTime="2025-10-02 02:03:11.012390448 +0000 UTC m=+1328.179134488" Oct 02 02:03:12 crc kubenswrapper[4775]: I1002 02:03:12.631689 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 02:03:12 crc kubenswrapper[4775]: I1002 02:03:12.632060 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 02:03:13 crc kubenswrapper[4775]: I1002 02:03:13.681564 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 02:03:17 crc kubenswrapper[4775]: I1002 02:03:17.631567 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 02:03:17 crc kubenswrapper[4775]: I1002 02:03:17.632228 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 02:03:18 crc kubenswrapper[4775]: I1002 02:03:18.654182 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5e322679-b95d-4328-b03b-0be7f0e82ac3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 02:03:18 crc kubenswrapper[4775]: I1002 02:03:18.654357 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="5e322679-b95d-4328-b03b-0be7f0e82ac3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 02:03:18 crc kubenswrapper[4775]: I1002 02:03:18.684336 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 02:03:18 crc kubenswrapper[4775]: I1002 02:03:18.715217 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 02:03:19 crc kubenswrapper[4775]: I1002 02:03:19.110522 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 02:03:19 crc kubenswrapper[4775]: I1002 02:03:19.336433 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 02:03:19 crc kubenswrapper[4775]: I1002 02:03:19.338244 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 02:03:20 crc kubenswrapper[4775]: I1002 02:03:20.095732 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 02:03:20 crc kubenswrapper[4775]: I1002 02:03:20.348100 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 02:03:20 crc kubenswrapper[4775]: I1002 02:03:20.348135 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.204:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 02 02:03:27 crc kubenswrapper[4775]: I1002 02:03:27.640665 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 02:03:27 crc kubenswrapper[4775]: I1002 02:03:27.641501 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 02:03:27 crc kubenswrapper[4775]: I1002 02:03:27.647491 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 02:03:27 crc kubenswrapper[4775]: I1002 02:03:27.650747 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 02:03:29 crc kubenswrapper[4775]: I1002 02:03:29.343178 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 02:03:29 crc kubenswrapper[4775]: I1002 02:03:29.344929 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 02:03:29 crc kubenswrapper[4775]: I1002 02:03:29.346893 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 02:03:29 crc kubenswrapper[4775]: I1002 02:03:29.353091 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 02:03:30 crc kubenswrapper[4775]: I1002 02:03:30.212995 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 02:03:30 crc kubenswrapper[4775]: I1002 02:03:30.234694 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 02:03:37 crc kubenswrapper[4775]: I1002 02:03:37.261728 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:03:37 crc kubenswrapper[4775]: I1002 02:03:37.262509 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:03:37 crc kubenswrapper[4775]: I1002 02:03:37.262600 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 02:03:37 crc kubenswrapper[4775]: I1002 02:03:37.263787 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"08d56b32ae1f08ce86ec95ee88e05d04f2a7f9064777b463fc71e568251c9583"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:03:37 crc kubenswrapper[4775]: I1002 02:03:37.265230 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://08d56b32ae1f08ce86ec95ee88e05d04f2a7f9064777b463fc71e568251c9583" gracePeriod=600 Oct 02 02:03:38 crc kubenswrapper[4775]: I1002 02:03:38.303235 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="08d56b32ae1f08ce86ec95ee88e05d04f2a7f9064777b463fc71e568251c9583" exitCode=0 Oct 02 02:03:38 crc kubenswrapper[4775]: I1002 02:03:38.303348 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"08d56b32ae1f08ce86ec95ee88e05d04f2a7f9064777b463fc71e568251c9583"} Oct 02 02:03:38 crc kubenswrapper[4775]: I1002 02:03:38.303694 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99"} Oct 02 02:03:38 crc kubenswrapper[4775]: I1002 02:03:38.303720 4775 scope.go:117] "RemoveContainer" containerID="61c8c995ae4362ef1e859023c2c2ffc70bf9e259d47718d71624878bd8dc02ec" Oct 02 02:03:48 crc kubenswrapper[4775]: I1002 02:03:48.810414 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 02 02:03:48 crc kubenswrapper[4775]: I1002 02:03:48.810968 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="764d109a-a65d-4fa6-ab71-a8ffb49e442e" containerName="openstackclient" containerID="cri-o://ce4365b06788ec5e2d1ab7f03c97b339cfee51b36bf1b4f570178d210ed939dc" gracePeriod=2 Oct 02 02:03:48 crc kubenswrapper[4775]: I1002 02:03:48.821816 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.016346 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.016630 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="7da01b78-a318-448e-8a4c-7ad3d9e833c0" containerName="openstack-network-exporter" containerID="cri-o://5b7338db5965862d0736735b7d766b4ed845f3e4cf8b4b81acd87174941ecfa6" gracePeriod=30 Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.016766 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="7da01b78-a318-448e-8a4c-7ad3d9e833c0" containerName="ovn-northd" containerID="cri-o://39f92c334c8a479d69a76f6ee477dc9e38b2f2fb13b731fddbffda7d7dac3511" gracePeriod=30 Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.041823 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placementf75b-account-delete-dl78c"] Oct 02 02:03:49 crc kubenswrapper[4775]: E1002 02:03:49.042244 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="764d109a-a65d-4fa6-ab71-a8ffb49e442e" containerName="openstackclient" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.042265 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="764d109a-a65d-4fa6-ab71-a8ffb49e442e" containerName="openstackclient" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.042465 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="764d109a-a65d-4fa6-ab71-a8ffb49e442e" containerName="openstackclient" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.043268 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementf75b-account-delete-dl78c" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.100196 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placementf75b-account-delete-dl78c"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.108449 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.109033 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="f7685ff7-607b-416c-9382-2c00edecaa54" containerName="openstack-network-exporter" containerID="cri-o://c0f5e7ac59fadc846940a5bf746aaf3da2e4edb2f52a02d0bbdc57d8847e6271" gracePeriod=300 Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.163319 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87xmq\" (UniqueName: \"kubernetes.io/projected/1e5e6844-45f6-4fd9-ac4f-3df2086164eb-kube-api-access-87xmq\") pod \"placementf75b-account-delete-dl78c\" (UID: \"1e5e6844-45f6-4fd9-ac4f-3df2086164eb\") " pod="openstack/placementf75b-account-delete-dl78c" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.231016 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.259760 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="f7685ff7-607b-416c-9382-2c00edecaa54" containerName="ovsdbserver-sb" containerID="cri-o://669d1080c1ef87bdc5aa946628176e085a6c1ef96db7059d070d3a1cdd097266" gracePeriod=300 Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.266915 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87xmq\" (UniqueName: \"kubernetes.io/projected/1e5e6844-45f6-4fd9-ac4f-3df2086164eb-kube-api-access-87xmq\") pod \"placementf75b-account-delete-dl78c\" (UID: \"1e5e6844-45f6-4fd9-ac4f-3df2086164eb\") " pod="openstack/placementf75b-account-delete-dl78c" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.284883 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance95b1-account-delete-4jqdz"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.297318 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance95b1-account-delete-4jqdz" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.300610 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87xmq\" (UniqueName: \"kubernetes.io/projected/1e5e6844-45f6-4fd9-ac4f-3df2086164eb-kube-api-access-87xmq\") pod \"placementf75b-account-delete-dl78c\" (UID: \"1e5e6844-45f6-4fd9-ac4f-3df2086164eb\") " pod="openstack/placementf75b-account-delete-dl78c" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.303511 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance95b1-account-delete-4jqdz"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.324982 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder5592-account-delete-mhfmd"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.326197 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5592-account-delete-mhfmd" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.353580 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder5592-account-delete-mhfmd"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.363020 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbicanb963-account-delete-z5bdz"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.364194 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanb963-account-delete-z5bdz" Oct 02 02:03:49 crc kubenswrapper[4775]: E1002 02:03:49.369834 4775 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 02 02:03:49 crc kubenswrapper[4775]: E1002 02:03:49.369891 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data podName:6638baa9-b367-40d1-8111-673dca6434ad nodeName:}" failed. No retries permitted until 2025-10-02 02:03:49.869873571 +0000 UTC m=+1367.036617611 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data") pod "rabbitmq-server-0" (UID: "6638baa9-b367-40d1-8111-673dca6434ad") : configmap "rabbitmq-config-data" not found Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.370648 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placementf75b-account-delete-dl78c" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.388090 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicanb963-account-delete-z5bdz"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.422099 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-hdtb5"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.432324 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-q2bd5"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.457336 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-q2bd5"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.457752 4775 generic.go:334] "Generic (PLEG): container finished" podID="7da01b78-a318-448e-8a4c-7ad3d9e833c0" containerID="5b7338db5965862d0736735b7d766b4ed845f3e4cf8b4b81acd87174941ecfa6" exitCode=2 Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.457808 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7da01b78-a318-448e-8a4c-7ad3d9e833c0","Type":"ContainerDied","Data":"5b7338db5965862d0736735b7d766b4ed845f3e4cf8b4b81acd87174941ecfa6"} Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.493656 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-hdtb5"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.556023 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lzz9\" (UniqueName: \"kubernetes.io/projected/cdea921f-d632-4463-a011-2e6f23f70546-kube-api-access-5lzz9\") pod \"glance95b1-account-delete-4jqdz\" (UID: \"cdea921f-d632-4463-a011-2e6f23f70546\") " pod="openstack/glance95b1-account-delete-4jqdz" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.559783 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2klh\" (UniqueName: \"kubernetes.io/projected/e3732523-d2de-45c9-ac69-770197f413f3-kube-api-access-g2klh\") pod \"cinder5592-account-delete-mhfmd\" (UID: \"e3732523-d2de-45c9-ac69-770197f413f3\") " pod="openstack/cinder5592-account-delete-mhfmd" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.563831 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd47g\" (UniqueName: \"kubernetes.io/projected/ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a-kube-api-access-jd47g\") pod \"barbicanb963-account-delete-z5bdz\" (UID: \"ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a\") " pod="openstack/barbicanb963-account-delete-z5bdz" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.585904 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f7685ff7-607b-416c-9382-2c00edecaa54/ovsdbserver-sb/0.log" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.587187 4775 generic.go:334] "Generic (PLEG): container finished" podID="f7685ff7-607b-416c-9382-2c00edecaa54" containerID="c0f5e7ac59fadc846940a5bf746aaf3da2e4edb2f52a02d0bbdc57d8847e6271" exitCode=2 Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.587281 4775 generic.go:334] "Generic (PLEG): container finished" podID="f7685ff7-607b-416c-9382-2c00edecaa54" containerID="669d1080c1ef87bdc5aa946628176e085a6c1ef96db7059d070d3a1cdd097266" exitCode=143 Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.587354 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f7685ff7-607b-416c-9382-2c00edecaa54","Type":"ContainerDied","Data":"c0f5e7ac59fadc846940a5bf746aaf3da2e4edb2f52a02d0bbdc57d8847e6271"} Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.587440 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f7685ff7-607b-416c-9382-2c00edecaa54","Type":"ContainerDied","Data":"669d1080c1ef87bdc5aa946628176e085a6c1ef96db7059d070d3a1cdd097266"} Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.634673 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-x8pvr"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.642062 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-x8pvr"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.663008 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-7szxv"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.665128 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-7szxv" podUID="e9813c97-677a-42c2-b526-576a4c5f2968" containerName="openstack-network-exporter" containerID="cri-o://47cbd1a6f145ee46322b61858213b665fd7c5043d5451925bf26dd59dec44238" gracePeriod=30 Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.670021 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jdbs4"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.683897 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2klh\" (UniqueName: \"kubernetes.io/projected/e3732523-d2de-45c9-ac69-770197f413f3-kube-api-access-g2klh\") pod \"cinder5592-account-delete-mhfmd\" (UID: \"e3732523-d2de-45c9-ac69-770197f413f3\") " pod="openstack/cinder5592-account-delete-mhfmd" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.684078 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd47g\" (UniqueName: \"kubernetes.io/projected/ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a-kube-api-access-jd47g\") pod \"barbicanb963-account-delete-z5bdz\" (UID: \"ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a\") " pod="openstack/barbicanb963-account-delete-z5bdz" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.684150 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lzz9\" (UniqueName: \"kubernetes.io/projected/cdea921f-d632-4463-a011-2e6f23f70546-kube-api-access-5lzz9\") pod \"glance95b1-account-delete-4jqdz\" (UID: \"cdea921f-d632-4463-a011-2e6f23f70546\") " pod="openstack/glance95b1-account-delete-4jqdz" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.687870 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-xwnrv"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.718370 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2klh\" (UniqueName: \"kubernetes.io/projected/e3732523-d2de-45c9-ac69-770197f413f3-kube-api-access-g2klh\") pod \"cinder5592-account-delete-mhfmd\" (UID: \"e3732523-d2de-45c9-ac69-770197f413f3\") " pod="openstack/cinder5592-account-delete-mhfmd" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.736518 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lzz9\" (UniqueName: \"kubernetes.io/projected/cdea921f-d632-4463-a011-2e6f23f70546-kube-api-access-5lzz9\") pod \"glance95b1-account-delete-4jqdz\" (UID: \"cdea921f-d632-4463-a011-2e6f23f70546\") " pod="openstack/glance95b1-account-delete-4jqdz" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.737848 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd47g\" (UniqueName: \"kubernetes.io/projected/ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a-kube-api-access-jd47g\") pod \"barbicanb963-account-delete-z5bdz\" (UID: \"ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a\") " pod="openstack/barbicanb963-account-delete-z5bdz" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.820667 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8df5f08f-cb36-4414-adb4-9bd68db8a41c" path="/var/lib/kubelet/pods/8df5f08f-cb36-4414-adb4-9bd68db8a41c/volumes" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.821458 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a3f9c1f-1298-4835-9935-56b808aae84d" path="/var/lib/kubelet/pods/9a3f9c1f-1298-4835-9935-56b808aae84d/volumes" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.822126 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9" path="/var/lib/kubelet/pods/fcfec5fd-0af5-44b8-8ba3-890e6a3cb6a9/volumes" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.845506 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.856286 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-dw9gs"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.871874 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-dw9gs"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.892718 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-hjvr2"] Oct 02 02:03:49 crc kubenswrapper[4775]: E1002 02:03:49.893978 4775 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 02 02:03:49 crc kubenswrapper[4775]: E1002 02:03:49.894023 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data podName:6638baa9-b367-40d1-8111-673dca6434ad nodeName:}" failed. No retries permitted until 2025-10-02 02:03:50.894006868 +0000 UTC m=+1368.060750908 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data") pod "rabbitmq-server-0" (UID: "6638baa9-b367-40d1-8111-673dca6434ad") : configmap "rabbitmq-config-data" not found Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.907756 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-hjvr2"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.911523 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance95b1-account-delete-4jqdz" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.927185 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5592-account-delete-mhfmd" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.944927 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-n65mg"] Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.963923 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanb963-account-delete-z5bdz" Oct 02 02:03:49 crc kubenswrapper[4775]: I1002 02:03:49.980163 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-n65mg"] Oct 02 02:03:49 crc kubenswrapper[4775]: E1002 02:03:49.996849 4775 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 02 02:03:49 crc kubenswrapper[4775]: E1002 02:03:49.996912 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data podName:fb0a7ddd-1558-4ad8-a8f4-e140306d19f9 nodeName:}" failed. No retries permitted until 2025-10-02 02:03:50.496898881 +0000 UTC m=+1367.663642921 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data") pod "rabbitmq-cell1-server-0" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9") : configmap "rabbitmq-cell1-config-data" not found Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.001930 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-8jbsr"] Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.002228 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" podUID="1364df9e-151f-40d0-9a5a-606a54021ef7" containerName="dnsmasq-dns" containerID="cri-o://9bdbd014e9940be2ba8f1e074e43a05086b36eec14f5338e99629708c2efdca9" gracePeriod=10 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.024039 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-777987c7d4-d5vzz"] Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.024295 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-777987c7d4-d5vzz" podUID="56813183-3dfc-430c-aaac-bda27d91340b" containerName="placement-log" containerID="cri-o://709d8a7bd6eb1d47bf0d40f08c279c162aa59c94b3583a614ae61982e1165eea" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.024827 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-777987c7d4-d5vzz" podUID="56813183-3dfc-430c-aaac-bda27d91340b" containerName="placement-api" containerID="cri-o://d46af6901507bf3d2c6f5cd7933a7df11f703991822ecd42a838e772bb645206" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.060381 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.061272 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" containerName="openstack-network-exporter" containerID="cri-o://b8d49f290ce08384af0005110e4c784aaece1c7d6ad52d6aaa1e3385f7367c8c" gracePeriod=300 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.077763 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" podUID="1364df9e-151f-40d0-9a5a-606a54021ef7" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.198:5353: connect: connection refused" Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.317053 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-fxjf5"] Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.428024 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-fxjf5"] Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.482030 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-dklqq"] Oct 02 02:03:50 crc kubenswrapper[4775]: E1002 02:03:50.516347 4775 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 02 02:03:50 crc kubenswrapper[4775]: E1002 02:03:50.516657 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data podName:fb0a7ddd-1558-4ad8-a8f4-e140306d19f9 nodeName:}" failed. No retries permitted until 2025-10-02 02:03:51.516640835 +0000 UTC m=+1368.683384875 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data") pod "rabbitmq-cell1-server-0" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9") : configmap "rabbitmq-cell1-config-data" not found Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.551598 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-dklqq"] Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.565077 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.565488 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-server" containerID="cri-o://e1f6027f6175a22e8ea34c7faa9698a5ad269b7fc4b557121c200890ba72ee33" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.565994 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="swift-recon-cron" containerID="cri-o://9a1e1ae6edc64baff01fb577c5fdff389a505836706e4e8e364c628542190596" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.566054 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="rsync" containerID="cri-o://e43595485c4ef19d17323cd2f15cc8c6c01cfd72d63f9027002a311666221ea2" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.566086 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-expirer" containerID="cri-o://a9fa0f88f8c28ceafea5f7d19fe1fa35b3b7c2e9511b51b17451d5c799b2adc6" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.566117 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-updater" containerID="cri-o://3b361b7fc7887886b0530a00fe1e6d2dd5a5e58cc96a870064fbcbeb95903b7b" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.566146 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-auditor" containerID="cri-o://6772443aac50d3396138519e7b7ffbb69baa9434eb913eb475d376b8df66e3ec" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.566177 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-replicator" containerID="cri-o://ad68cb16ec4f02e54fb4a47b88804b69e1f0edd04ff7e5ec3f86cdfe3bdfc45d" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.566210 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-server" containerID="cri-o://b14587fcd25cde3d5786c84529e1eac9820c2a094fc6d465d49b957d0ca1f117" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.566251 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-updater" containerID="cri-o://783ca618d590525d6b1848438fe648b59eddb9e00181a39a7cabe0cd1d417d06" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.566287 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-auditor" containerID="cri-o://05475c086b8f31273f54dfe829e118fed85817c9d16d5585b851871fc39726bb" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.566317 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-replicator" containerID="cri-o://24cb4406a78e877a5465e016c64de5577c4354d83b07c019e2d33609646a1f9a" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.566343 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-server" containerID="cri-o://a994b8ddd7e1ce4fa5946caa21707a15ffb193c4b308890c750b2ed3745c88d6" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.566371 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-reaper" containerID="cri-o://35e88601383e15eea377e15e1d17c552b0acdc4bf1bb0f64c98b395f45c898cf" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.566400 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-auditor" containerID="cri-o://1ad3d27e30ff5a6104b653354bc8348baa64801d396c0e15ef454a96004528cf" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.566431 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-replicator" containerID="cri-o://93013dec09b53cd82ac1d950f89f6d21acde0be93880dc71f932f03315d26749" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.583512 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.583797 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" containerName="glance-log" containerID="cri-o://12f66def0a878a3dc74caa0f75419c15f378c889154267260d12b5003e5ebee3" gracePeriod=30 Oct 02 02:03:50 crc kubenswrapper[4775]: I1002 02:03:50.584387 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" containerName="glance-httpd" containerID="cri-o://63252743cedda5425e3430e58790235c337b97aaa0683e299068d0092eca5e7f" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.642451 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.642872 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="24e08fdf-b02d-4d6b-b451-8aa967ff04f6" containerName="cinder-scheduler" containerID="cri-o://86650058d32beec95536f3008ef303ebaebff3385b901bf98a315478321c1bce" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.643546 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="24e08fdf-b02d-4d6b-b451-8aa967ff04f6" containerName="probe" containerID="cri-o://ecfa4d3c77b7cfaba81f2e5e014ad3a08c3704a721e8b2bcc3173012b42a0aff" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.652354 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.652584 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8f286b47-509d-479d-bad5-0bbec930558a" containerName="glance-log" containerID="cri-o://35fba6a1f824b8a5a30f3f0d8b52922737c0bdb3b40b6987fcc54b88ccfc67ab" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.652717 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8f286b47-509d-479d-bad5-0bbec930558a" containerName="glance-httpd" containerID="cri-o://db8059deef67f936d22e07a37161aed4eda44bdc81845a2918822f9946fd9c00" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.665496 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.665753 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5bca47d0-79b7-4992-867b-d5e05e02cdc8" containerName="cinder-api-log" containerID="cri-o://04d6bdf03305dd966a2ca95b6a3220b41a5add02a16becda008422ee10720237" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.665872 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="5bca47d0-79b7-4992-867b-d5e05e02cdc8" containerName="cinder-api" containerID="cri-o://e38c74d33b3ad7b8f747e53560c787d296c45bf29b3227467260f620932b62a9" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.717023 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-546f7b9c49-h6rcs"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.717415 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-546f7b9c49-h6rcs" podUID="2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" containerName="neutron-api" containerID="cri-o://8c9ca1dbcfc1426fde26f45e447e342f40b762f7d726a5c8194371f054ef7feb" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.717571 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-546f7b9c49-h6rcs" podUID="2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" containerName="neutron-httpd" containerID="cri-o://85353a3b8a9a2c508bfbb1d8f15bdea36723aba6e9d1ddda8f826588f795b512" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.806396 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-582w8"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.837324 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-582w8"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.860292 4775 generic.go:334] "Generic (PLEG): container finished" podID="c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" containerID="b8d49f290ce08384af0005110e4c784aaece1c7d6ad52d6aaa1e3385f7367c8c" exitCode=2 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.860401 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92","Type":"ContainerDied","Data":"b8d49f290ce08384af0005110e4c784aaece1c7d6ad52d6aaa1e3385f7367c8c"} Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.882295 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-f75b-account-create-6t2rc"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.906411 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-f75b-account-create-6t2rc"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.910699 4775 generic.go:334] "Generic (PLEG): container finished" podID="56813183-3dfc-430c-aaac-bda27d91340b" containerID="709d8a7bd6eb1d47bf0d40f08c279c162aa59c94b3583a614ae61982e1165eea" exitCode=143 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.910799 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-777987c7d4-d5vzz" event={"ID":"56813183-3dfc-430c-aaac-bda27d91340b","Type":"ContainerDied","Data":"709d8a7bd6eb1d47bf0d40f08c279c162aa59c94b3583a614ae61982e1165eea"} Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.918121 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7szxv_e9813c97-677a-42c2-b526-576a4c5f2968/openstack-network-exporter/0.log" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.918160 4775 generic.go:334] "Generic (PLEG): container finished" podID="e9813c97-677a-42c2-b526-576a4c5f2968" containerID="47cbd1a6f145ee46322b61858213b665fd7c5043d5451925bf26dd59dec44238" exitCode=2 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.918206 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7szxv" event={"ID":"e9813c97-677a-42c2-b526-576a4c5f2968","Type":"ContainerDied","Data":"47cbd1a6f145ee46322b61858213b665fd7c5043d5451925bf26dd59dec44238"} Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.922006 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementf75b-account-delete-dl78c"] Oct 02 02:03:51 crc kubenswrapper[4775]: E1002 02:03:50.937121 4775 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 02 02:03:51 crc kubenswrapper[4775]: E1002 02:03:50.937171 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data podName:6638baa9-b367-40d1-8111-673dca6434ad nodeName:}" failed. No retries permitted until 2025-10-02 02:03:52.93715536 +0000 UTC m=+1370.103899400 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data") pod "rabbitmq-server-0" (UID: "6638baa9-b367-40d1-8111-673dca6434ad") : configmap "rabbitmq-config-data" not found Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.949423 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-x8t6d"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.949564 4775 generic.go:334] "Generic (PLEG): container finished" podID="1364df9e-151f-40d0-9a5a-606a54021ef7" containerID="9bdbd014e9940be2ba8f1e074e43a05086b36eec14f5338e99629708c2efdca9" exitCode=0 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.949603 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" event={"ID":"1364df9e-151f-40d0-9a5a-606a54021ef7","Type":"ContainerDied","Data":"9bdbd014e9940be2ba8f1e074e43a05086b36eec14f5338e99629708c2efdca9"} Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:50.976139 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-19ef-account-create-n7j6v"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.018048 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-x8t6d"] Oct 02 02:03:51 crc kubenswrapper[4775]: E1002 02:03:51.046206 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="39f92c334c8a479d69a76f6ee477dc9e38b2f2fb13b731fddbffda7d7dac3511" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 02 02:03:51 crc kubenswrapper[4775]: E1002 02:03:51.058303 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="39f92c334c8a479d69a76f6ee477dc9e38b2f2fb13b731fddbffda7d7dac3511" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.062021 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-19ef-account-create-n7j6v"] Oct 02 02:03:51 crc kubenswrapper[4775]: E1002 02:03:51.081309 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="39f92c334c8a479d69a76f6ee477dc9e38b2f2fb13b731fddbffda7d7dac3511" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 02 02:03:51 crc kubenswrapper[4775]: E1002 02:03:51.081377 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="7da01b78-a318-448e-8a4c-7ad3d9e833c0" containerName="ovn-northd" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.087251 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-m576z"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.123909 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-m576z"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.134305 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.134547 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" containerName="nova-api-log" containerID="cri-o://303fb7d6dfef981bea20389d78f64a2cf286846122e5c5d3878e6b65fef9edb7" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.134936 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" containerName="nova-api-api" containerID="cri-o://9d068bf0ca5c469b57dcad75c4e6ebff62f9badd694815d6349f36cd138a20a8" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.157515 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-95b1-account-create-57xcl"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.169103 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-95b1-account-create-57xcl"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.177100 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance95b1-account-delete-4jqdz"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.187175 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder5592-account-delete-mhfmd"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.190319 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-5592-account-create-f4hz6"] Oct 02 02:03:51 crc kubenswrapper[4775]: E1002 02:03:51.196866 4775 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 02 02:03:51 crc kubenswrapper[4775]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 02 02:03:51 crc kubenswrapper[4775]: + source /usr/local/bin/container-scripts/functions Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNBridge=br-int Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNRemote=tcp:localhost:6642 Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNEncapType=geneve Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNAvailabilityZones= Oct 02 02:03:51 crc kubenswrapper[4775]: ++ EnableChassisAsGateway=true Oct 02 02:03:51 crc kubenswrapper[4775]: ++ PhysicalNetworks= Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNHostName= Oct 02 02:03:51 crc kubenswrapper[4775]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 02 02:03:51 crc kubenswrapper[4775]: ++ ovs_dir=/var/lib/openvswitch Oct 02 02:03:51 crc kubenswrapper[4775]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 02 02:03:51 crc kubenswrapper[4775]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 02 02:03:51 crc kubenswrapper[4775]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 02 02:03:51 crc kubenswrapper[4775]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 02:03:51 crc kubenswrapper[4775]: + sleep 0.5 Oct 02 02:03:51 crc kubenswrapper[4775]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 02:03:51 crc kubenswrapper[4775]: + sleep 0.5 Oct 02 02:03:51 crc kubenswrapper[4775]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 02:03:51 crc kubenswrapper[4775]: + cleanup_ovsdb_server_semaphore Oct 02 02:03:51 crc kubenswrapper[4775]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 02 02:03:51 crc kubenswrapper[4775]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 02 02:03:51 crc kubenswrapper[4775]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-xwnrv" message=< Oct 02 02:03:51 crc kubenswrapper[4775]: Exiting ovsdb-server (5) [ OK ] Oct 02 02:03:51 crc kubenswrapper[4775]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 02 02:03:51 crc kubenswrapper[4775]: + source /usr/local/bin/container-scripts/functions Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNBridge=br-int Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNRemote=tcp:localhost:6642 Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNEncapType=geneve Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNAvailabilityZones= Oct 02 02:03:51 crc kubenswrapper[4775]: ++ EnableChassisAsGateway=true Oct 02 02:03:51 crc kubenswrapper[4775]: ++ PhysicalNetworks= Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNHostName= Oct 02 02:03:51 crc kubenswrapper[4775]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 02 02:03:51 crc kubenswrapper[4775]: ++ ovs_dir=/var/lib/openvswitch Oct 02 02:03:51 crc kubenswrapper[4775]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 02 02:03:51 crc kubenswrapper[4775]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 02 02:03:51 crc kubenswrapper[4775]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 02 02:03:51 crc kubenswrapper[4775]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 02:03:51 crc kubenswrapper[4775]: + sleep 0.5 Oct 02 02:03:51 crc kubenswrapper[4775]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 02:03:51 crc kubenswrapper[4775]: + sleep 0.5 Oct 02 02:03:51 crc kubenswrapper[4775]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 02:03:51 crc kubenswrapper[4775]: + cleanup_ovsdb_server_semaphore Oct 02 02:03:51 crc kubenswrapper[4775]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 02 02:03:51 crc kubenswrapper[4775]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 02 02:03:51 crc kubenswrapper[4775]: > Oct 02 02:03:51 crc kubenswrapper[4775]: E1002 02:03:51.196898 4775 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 02 02:03:51 crc kubenswrapper[4775]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Oct 02 02:03:51 crc kubenswrapper[4775]: + source /usr/local/bin/container-scripts/functions Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNBridge=br-int Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNRemote=tcp:localhost:6642 Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNEncapType=geneve Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNAvailabilityZones= Oct 02 02:03:51 crc kubenswrapper[4775]: ++ EnableChassisAsGateway=true Oct 02 02:03:51 crc kubenswrapper[4775]: ++ PhysicalNetworks= Oct 02 02:03:51 crc kubenswrapper[4775]: ++ OVNHostName= Oct 02 02:03:51 crc kubenswrapper[4775]: ++ DB_FILE=/etc/openvswitch/conf.db Oct 02 02:03:51 crc kubenswrapper[4775]: ++ ovs_dir=/var/lib/openvswitch Oct 02 02:03:51 crc kubenswrapper[4775]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Oct 02 02:03:51 crc kubenswrapper[4775]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Oct 02 02:03:51 crc kubenswrapper[4775]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 02 02:03:51 crc kubenswrapper[4775]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 02:03:51 crc kubenswrapper[4775]: + sleep 0.5 Oct 02 02:03:51 crc kubenswrapper[4775]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 02:03:51 crc kubenswrapper[4775]: + sleep 0.5 Oct 02 02:03:51 crc kubenswrapper[4775]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Oct 02 02:03:51 crc kubenswrapper[4775]: + cleanup_ovsdb_server_semaphore Oct 02 02:03:51 crc kubenswrapper[4775]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Oct 02 02:03:51 crc kubenswrapper[4775]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Oct 02 02:03:51 crc kubenswrapper[4775]: > pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovsdb-server" containerID="cri-o://19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.196926 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovsdb-server" containerID="cri-o://19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" gracePeriod=29 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.199711 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-t6skn"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.215037 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-5592-account-create-f4hz6"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.224428 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-t6skn"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.236313 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-pfvt8"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.236352 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-pfvt8"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.246887 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicanb963-account-delete-z5bdz"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.263465 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-b963-account-create-zbgmn"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.270748 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-b963-account-create-zbgmn"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.277715 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.282811 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-rjs9s"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.291822 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-rjs9s"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.299816 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.300100 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5e322679-b95d-4328-b03b-0be7f0e82ac3" containerName="nova-metadata-log" containerID="cri-o://0d717b8fb0cfdfcf5e5314de128b353f94395b4da93ebdbec0e1f666b4b81244" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.300408 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5e322679-b95d-4328-b03b-0be7f0e82ac3" containerName="nova-metadata-metadata" containerID="cri-o://33e28213f9c089646224401e310e6c70d1efdff65261c89fe379e544e09ae98f" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.305499 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-5350-account-create-lk56t"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.313135 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-0729-account-create-v8vz9"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.321129 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-5350-account-create-lk56t"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.329709 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-0729-account-create-v8vz9"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.350270 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovs-vswitchd" probeResult="failure" output=< Oct 02 02:03:51 crc kubenswrapper[4775]: 2025-10-02T02:03:51Z|00001|jsonrpc|WARN|unix:/var/run/openvswitch/ovs-vswitchd.11.ctl: receive error: Connection reset by peer Oct 02 02:03:51 crc kubenswrapper[4775]: 2025-10-02T02:03:51Z|00002|unixctl|WARN|error communicating with unix:/var/run/openvswitch/ovs-vswitchd.11.ctl: Connection reset by peer Oct 02 02:03:51 crc kubenswrapper[4775]: ovs-appctl: /var/run/openvswitch/ovs-vswitchd.11.ctl: transaction error (Connection reset by peer) Oct 02 02:03:51 crc kubenswrapper[4775]: ERROR - Failed retrieving ofproto/list from ovs-vswitchd Oct 02 02:03:51 crc kubenswrapper[4775]: > Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.350369 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-ddr42"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.350408 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-b8mb2"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.351225 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" containerName="ovsdbserver-nb" containerID="cri-o://0089ce10498396392f25a55b07f3d03f418bbb5e5d51703da8bdb90f1f8cc858" gracePeriod=299 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.424537 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-f3f8-account-create-86qvz"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.454202 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-b8mb2"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.472903 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-ddr42"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.482869 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-f3f8-account-create-86qvz"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.496395 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.514144 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-d79dd6fcb-lzscb"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.514465 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" podUID="af16c5f4-fcd0-4450-a8b7-789ff104e477" containerName="barbican-keystone-listener-log" containerID="cri-o://3aff4b0f97413fafd4a12944ee3201d623169d99e066f7bf1d2aa5a8977df17f" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.515196 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" podUID="af16c5f4-fcd0-4450-a8b7-789ff104e477" containerName="barbican-keystone-listener" containerID="cri-o://ae9bd041c2c057dee311d85415e1bfd222d0109f3f01cb2fcd2149362f273f6e" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.535315 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-79594d69d8-x75fk"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.535554 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-79594d69d8-x75fk" podUID="8d6bb525-8c94-461f-84c8-9e466b789706" containerName="barbican-api-log" containerID="cri-o://9f162e9f65c919c062e4c98961fb915d6d73a9e2d76b98f4844b90766369c972" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.535762 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-79594d69d8-x75fk" podUID="8d6bb525-8c94-461f-84c8-9e466b789706" containerName="barbican-api" containerID="cri-o://ddc807916bf2152781e0b263b81d2ba2f8ff95a514f51a7d7918774142370f47" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.543818 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovs-vswitchd" containerID="cri-o://2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" gracePeriod=29 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.557007 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-547cbfbdcc-nk6pz"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.557215 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-547cbfbdcc-nk6pz" podUID="163a6d17-be4c-47fc-8835-e218e8a803f1" containerName="barbican-worker-log" containerID="cri-o://9d74e2868485fd65eb7dd4d89c912bc745152afc0102e87eaff58d02c1858850" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.557516 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-547cbfbdcc-nk6pz" podUID="163a6d17-be4c-47fc-8835-e218e8a803f1" containerName="barbican-worker" containerID="cri-o://979478a6b7d6bf80d7823b67cd948eb55c9e92647ad8b39e7e0c75474a16acc7" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.571596 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.572293 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="a5c60336-6e25-4c68-98e6-52801f34578b" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://38c453bdb403040517a2812e8643acfa55f8124baa338e04c4028a69bd4cbbd1" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: E1002 02:03:51.587334 4775 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 02 02:03:51 crc kubenswrapper[4775]: E1002 02:03:51.587461 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data podName:fb0a7ddd-1558-4ad8-a8f4-e140306d19f9 nodeName:}" failed. No retries permitted until 2025-10-02 02:03:53.587388786 +0000 UTC m=+1370.754132826 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data") pod "rabbitmq-cell1-server-0" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9") : configmap "rabbitmq-cell1-config-data" not found Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.588639 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.588877 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="46453bb2-6af0-4dd4-83ba-b14bf3f96310" containerName="nova-cell1-conductor-conductor" containerID="cri-o://b66fc6b4d3be69dcb9c2d949e6dfe8d6ab6ddb59f7b8ad17c8380c0b6034ccf6" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.617901 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7jsjh"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.634665 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-7jsjh"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.651108 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="6638baa9-b367-40d1-8111-673dca6434ad" containerName="rabbitmq" containerID="cri-o://836e86acc5faf2a3e9656e8d8d2d493dd928f3335bd576fc6d293e8ae72cc25c" gracePeriod=604800 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.657009 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7jgdw"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.673002 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-7jgdw"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.682400 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.682609 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="c74e0a1d-125d-43f7-9ba3-cea70453fc01" containerName="nova-cell0-conductor-conductor" containerID="cri-o://2cf05392cd7bff8e546c2fbad0965375d71482567f75af41133d05802bad7f02" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.688553 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.752345 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="3adc5647-1145-4c44-9ecc-66d1d9a19023" containerName="galera" containerID="cri-o://1209a9ffa5fb6c6952b1b8effd2ebc2e2fb407ab840357e9b3b1f508504a915a" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.779144 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" containerName="rabbitmq" containerID="cri-o://c4f574fe362f1c98b06c39ae135acf138f8c156d889aa930a36b192af34be971" gracePeriod=604800 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.810151 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b46397b-0c28-46ab-8f20-1426466202b3" path="/var/lib/kubelet/pods/0b46397b-0c28-46ab-8f20-1426466202b3/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.810744 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d054374-ba0f-41f4-8894-f1181d003bb3" path="/var/lib/kubelet/pods/1d054374-ba0f-41f4-8894-f1181d003bb3/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.811255 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="666eb22c-6c8b-4976-ae93-803c8b7cf7c6" path="/var/lib/kubelet/pods/666eb22c-6c8b-4976-ae93-803c8b7cf7c6/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.811773 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77750932-5356-469a-ad3d-640973d830d9" path="/var/lib/kubelet/pods/77750932-5356-469a-ad3d-640973d830d9/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.813003 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="777deed6-1940-4ae7-b4bb-e9f71fa4429e" path="/var/lib/kubelet/pods/777deed6-1940-4ae7-b4bb-e9f71fa4429e/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.813566 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="833fd434-3bdc-4be7-be6f-f06915509129" path="/var/lib/kubelet/pods/833fd434-3bdc-4be7-be6f-f06915509129/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.814154 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9" path="/var/lib/kubelet/pods/8d7fca32-3e7f-4b54-96ae-8bc0dbf311c9/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.815868 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ee94784-1d45-48fa-b7ad-81f12013c44c" path="/var/lib/kubelet/pods/8ee94784-1d45-48fa-b7ad-81f12013c44c/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.816472 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a120b1c5-156f-432c-96ca-f9a4dfc6dfa7" path="/var/lib/kubelet/pods/a120b1c5-156f-432c-96ca-f9a4dfc6dfa7/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.816986 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a34ad439-41ea-4ee1-a7b3-4075d9923915" path="/var/lib/kubelet/pods/a34ad439-41ea-4ee1-a7b3-4075d9923915/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.817445 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b01ed253-a630-4b6c-9282-929b34b69b6d" path="/var/lib/kubelet/pods/b01ed253-a630-4b6c-9282-929b34b69b6d/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.818389 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7f333d7-e144-4ad1-be2f-aae61372eab5" path="/var/lib/kubelet/pods/b7f333d7-e144-4ad1-be2f-aae61372eab5/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.818880 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcde2ae4-f3e6-4a25-a9d9-ede53f848a44" path="/var/lib/kubelet/pods/bcde2ae4-f3e6-4a25-a9d9-ede53f848a44/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.819381 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c20e73b6-3188-49c6-836d-e518d93d9457" path="/var/lib/kubelet/pods/c20e73b6-3188-49c6-836d-e518d93d9457/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.819855 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2862399-bef9-4b91-9376-3ceff3377234" path="/var/lib/kubelet/pods/c2862399-bef9-4b91-9376-3ceff3377234/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.820908 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2d22306-6f52-462a-a980-210c208c595e" path="/var/lib/kubelet/pods/c2d22306-6f52-462a-a980-210c208c595e/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.821467 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c34b79a8-c071-4fff-8dbc-cab3601cc3f8" path="/var/lib/kubelet/pods/c34b79a8-c071-4fff-8dbc-cab3601cc3f8/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.828522 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdaf4602-ca11-419d-846c-425d3d5bc766" path="/var/lib/kubelet/pods/cdaf4602-ca11-419d-846c-425d3d5bc766/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.829050 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d01b3c76-8f88-4fc6-be13-4f8150a14949" path="/var/lib/kubelet/pods/d01b3c76-8f88-4fc6-be13-4f8150a14949/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.829566 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d05cdd83-5605-46ed-88e8-168edbbcf7c1" path="/var/lib/kubelet/pods/d05cdd83-5605-46ed-88e8-168edbbcf7c1/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.830564 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e19ef24d-2e7c-47a9-943c-a49f650df3a7" path="/var/lib/kubelet/pods/e19ef24d-2e7c-47a9-943c-a49f650df3a7/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.831145 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f14a7a78-6b40-4c82-8dc6-64cc069cb1fd" path="/var/lib/kubelet/pods/f14a7a78-6b40-4c82-8dc6-64cc069cb1fd/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.831902 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdb63c2b-eade-4e1d-a62f-24475df065e2" path="/var/lib/kubelet/pods/fdb63c2b-eade-4e1d-a62f-24475df065e2/volumes" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.833789 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.833829 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementf75b-account-delete-dl78c"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.834081 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="a64f8ba3-65b9-489f-ace1-757f3ecce6b4" containerName="nova-scheduler-scheduler" containerID="cri-o://cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa" gracePeriod=30 Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.952065 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance95b1-account-delete-4jqdz"] Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.976313 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7szxv_e9813c97-677a-42c2-b526-576a4c5f2968/openstack-network-exporter/0.log" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.976442 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7szxv" event={"ID":"e9813c97-677a-42c2-b526-576a4c5f2968","Type":"ContainerDied","Data":"824fea2e009c65cbc6e17e23f3849349ab3e9f4be398ec804d6d7dec322e2ce1"} Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.976527 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="824fea2e009c65cbc6e17e23f3849349ab3e9f4be398ec804d6d7dec322e2ce1" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.983287 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" event={"ID":"1364df9e-151f-40d0-9a5a-606a54021ef7","Type":"ContainerDied","Data":"5da6a552854ebc54e69c6124436167ea7d5496a4afc985a7bb27725f03a1dc6e"} Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.983374 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5da6a552854ebc54e69c6124436167ea7d5496a4afc985a7bb27725f03a1dc6e" Oct 02 02:03:51 crc kubenswrapper[4775]: I1002 02:03:51.995687 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder5592-account-delete-mhfmd"] Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.013110 4775 generic.go:334] "Generic (PLEG): container finished" podID="2a54aa55-e3e2-4844-b175-e68442220b92" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.013333 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xwnrv" event={"ID":"2a54aa55-e3e2-4844-b175-e68442220b92","Type":"ContainerDied","Data":"19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.030253 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicanb963-account-delete-z5bdz"] Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.045113 4775 generic.go:334] "Generic (PLEG): container finished" podID="5e322679-b95d-4328-b03b-0be7f0e82ac3" containerID="0d717b8fb0cfdfcf5e5314de128b353f94395b4da93ebdbec0e1f666b4b81244" exitCode=143 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.045468 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e322679-b95d-4328-b03b-0be7f0e82ac3","Type":"ContainerDied","Data":"0d717b8fb0cfdfcf5e5314de128b353f94395b4da93ebdbec0e1f666b4b81244"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.059555 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f7685ff7-607b-416c-9382-2c00edecaa54/ovsdbserver-sb/0.log" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.059634 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"f7685ff7-607b-416c-9382-2c00edecaa54","Type":"ContainerDied","Data":"e41ea836798b128f59bc00526da81b9cd7e04b857cb248a53de519eb40c2a59e"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.059657 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e41ea836798b128f59bc00526da81b9cd7e04b857cb248a53de519eb40c2a59e" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.072921 4775 generic.go:334] "Generic (PLEG): container finished" podID="8d6bb525-8c94-461f-84c8-9e466b789706" containerID="9f162e9f65c919c062e4c98961fb915d6d73a9e2d76b98f4844b90766369c972" exitCode=143 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.072990 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79594d69d8-x75fk" event={"ID":"8d6bb525-8c94-461f-84c8-9e466b789706","Type":"ContainerDied","Data":"9f162e9f65c919c062e4c98961fb915d6d73a9e2d76b98f4844b90766369c972"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.074595 4775 generic.go:334] "Generic (PLEG): container finished" podID="5bca47d0-79b7-4992-867b-d5e05e02cdc8" containerID="04d6bdf03305dd966a2ca95b6a3220b41a5add02a16becda008422ee10720237" exitCode=143 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.074632 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5bca47d0-79b7-4992-867b-d5e05e02cdc8","Type":"ContainerDied","Data":"04d6bdf03305dd966a2ca95b6a3220b41a5add02a16becda008422ee10720237"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.076095 4775 generic.go:334] "Generic (PLEG): container finished" podID="8f286b47-509d-479d-bad5-0bbec930558a" containerID="35fba6a1f824b8a5a30f3f0d8b52922737c0bdb3b40b6987fcc54b88ccfc67ab" exitCode=143 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.076133 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f286b47-509d-479d-bad5-0bbec930558a","Type":"ContainerDied","Data":"35fba6a1f824b8a5a30f3f0d8b52922737c0bdb3b40b6987fcc54b88ccfc67ab"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082170 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="e43595485c4ef19d17323cd2f15cc8c6c01cfd72d63f9027002a311666221ea2" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082194 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="a9fa0f88f8c28ceafea5f7d19fe1fa35b3b7c2e9511b51b17451d5c799b2adc6" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082203 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="3b361b7fc7887886b0530a00fe1e6d2dd5a5e58cc96a870064fbcbeb95903b7b" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082212 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="6772443aac50d3396138519e7b7ffbb69baa9434eb913eb475d376b8df66e3ec" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082219 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="ad68cb16ec4f02e54fb4a47b88804b69e1f0edd04ff7e5ec3f86cdfe3bdfc45d" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082276 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="b14587fcd25cde3d5786c84529e1eac9820c2a094fc6d465d49b957d0ca1f117" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082283 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="783ca618d590525d6b1848438fe648b59eddb9e00181a39a7cabe0cd1d417d06" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082288 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="05475c086b8f31273f54dfe829e118fed85817c9d16d5585b851871fc39726bb" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082294 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="24cb4406a78e877a5465e016c64de5577c4354d83b07c019e2d33609646a1f9a" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082301 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="a994b8ddd7e1ce4fa5946caa21707a15ffb193c4b308890c750b2ed3745c88d6" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082317 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="35e88601383e15eea377e15e1d17c552b0acdc4bf1bb0f64c98b395f45c898cf" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082322 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="1ad3d27e30ff5a6104b653354bc8348baa64801d396c0e15ef454a96004528cf" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082328 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="93013dec09b53cd82ac1d950f89f6d21acde0be93880dc71f932f03315d26749" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082335 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="e1f6027f6175a22e8ea34c7faa9698a5ad269b7fc4b557121c200890ba72ee33" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082370 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"e43595485c4ef19d17323cd2f15cc8c6c01cfd72d63f9027002a311666221ea2"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082389 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"a9fa0f88f8c28ceafea5f7d19fe1fa35b3b7c2e9511b51b17451d5c799b2adc6"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082399 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"3b361b7fc7887886b0530a00fe1e6d2dd5a5e58cc96a870064fbcbeb95903b7b"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082407 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"6772443aac50d3396138519e7b7ffbb69baa9434eb913eb475d376b8df66e3ec"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082415 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"ad68cb16ec4f02e54fb4a47b88804b69e1f0edd04ff7e5ec3f86cdfe3bdfc45d"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082424 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"b14587fcd25cde3d5786c84529e1eac9820c2a094fc6d465d49b957d0ca1f117"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082432 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"783ca618d590525d6b1848438fe648b59eddb9e00181a39a7cabe0cd1d417d06"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082440 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"05475c086b8f31273f54dfe829e118fed85817c9d16d5585b851871fc39726bb"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082449 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"24cb4406a78e877a5465e016c64de5577c4354d83b07c019e2d33609646a1f9a"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082457 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"a994b8ddd7e1ce4fa5946caa21707a15ffb193c4b308890c750b2ed3745c88d6"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082465 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"35e88601383e15eea377e15e1d17c552b0acdc4bf1bb0f64c98b395f45c898cf"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082473 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"1ad3d27e30ff5a6104b653354bc8348baa64801d396c0e15ef454a96004528cf"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082492 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"93013dec09b53cd82ac1d950f89f6d21acde0be93880dc71f932f03315d26749"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.082501 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"e1f6027f6175a22e8ea34c7faa9698a5ad269b7fc4b557121c200890ba72ee33"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.084566 4775 generic.go:334] "Generic (PLEG): container finished" podID="2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" containerID="85353a3b8a9a2c508bfbb1d8f15bdea36723aba6e9d1ddda8f826588f795b512" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.084605 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-546f7b9c49-h6rcs" event={"ID":"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab","Type":"ContainerDied","Data":"85353a3b8a9a2c508bfbb1d8f15bdea36723aba6e9d1ddda8f826588f795b512"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.139833 4775 generic.go:334] "Generic (PLEG): container finished" podID="8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" containerID="303fb7d6dfef981bea20389d78f64a2cf286846122e5c5d3878e6b65fef9edb7" exitCode=143 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.139891 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0","Type":"ContainerDied","Data":"303fb7d6dfef981bea20389d78f64a2cf286846122e5c5d3878e6b65fef9edb7"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.150980 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder5592-account-delete-mhfmd" event={"ID":"e3732523-d2de-45c9-ac69-770197f413f3","Type":"ContainerStarted","Data":"a069d853f5876b595def267dae48810cc164f576380990b4bbc0c9615bdfff3a"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.153237 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementf75b-account-delete-dl78c" event={"ID":"1e5e6844-45f6-4fd9-ac4f-3df2086164eb","Type":"ContainerStarted","Data":"aa6896692c152e1046b85695e2663ae1516576a62c99a10f86e65ba674a54200"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.165237 4775 generic.go:334] "Generic (PLEG): container finished" podID="163a6d17-be4c-47fc-8835-e218e8a803f1" containerID="9d74e2868485fd65eb7dd4d89c912bc745152afc0102e87eaff58d02c1858850" exitCode=143 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.165295 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-547cbfbdcc-nk6pz" event={"ID":"163a6d17-be4c-47fc-8835-e218e8a803f1","Type":"ContainerDied","Data":"9d74e2868485fd65eb7dd4d89c912bc745152afc0102e87eaff58d02c1858850"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.168986 4775 generic.go:334] "Generic (PLEG): container finished" podID="764d109a-a65d-4fa6-ab71-a8ffb49e442e" containerID="ce4365b06788ec5e2d1ab7f03c97b339cfee51b36bf1b4f570178d210ed939dc" exitCode=137 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.169385 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9935899c8ea57c04dd3323283b4595948225cf06d8dc3c4b280ffb5b2632bae4" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.175202 4775 generic.go:334] "Generic (PLEG): container finished" podID="af16c5f4-fcd0-4450-a8b7-789ff104e477" containerID="3aff4b0f97413fafd4a12944ee3201d623169d99e066f7bf1d2aa5a8977df17f" exitCode=143 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.175257 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" event={"ID":"af16c5f4-fcd0-4450-a8b7-789ff104e477","Type":"ContainerDied","Data":"3aff4b0f97413fafd4a12944ee3201d623169d99e066f7bf1d2aa5a8977df17f"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.177162 4775 generic.go:334] "Generic (PLEG): container finished" podID="c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" containerID="0089ce10498396392f25a55b07f3d03f418bbb5e5d51703da8bdb90f1f8cc858" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.177235 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92","Type":"ContainerDied","Data":"0089ce10498396392f25a55b07f3d03f418bbb5e5d51703da8bdb90f1f8cc858"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.179917 4775 generic.go:334] "Generic (PLEG): container finished" podID="8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" containerID="12f66def0a878a3dc74caa0f75419c15f378c889154267260d12b5003e5ebee3" exitCode=143 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.179995 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51","Type":"ContainerDied","Data":"12f66def0a878a3dc74caa0f75419c15f378c889154267260d12b5003e5ebee3"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.181989 4775 generic.go:334] "Generic (PLEG): container finished" podID="24e08fdf-b02d-4d6b-b451-8aa967ff04f6" containerID="ecfa4d3c77b7cfaba81f2e5e014ad3a08c3704a721e8b2bcc3173012b42a0aff" exitCode=0 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.182015 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"24e08fdf-b02d-4d6b-b451-8aa967ff04f6","Type":"ContainerDied","Data":"ecfa4d3c77b7cfaba81f2e5e014ad3a08c3704a721e8b2bcc3173012b42a0aff"} Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.206180 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7szxv_e9813c97-677a-42c2-b526-576a4c5f2968/openstack-network-exporter/0.log" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.206243 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7szxv" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.268100 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_f7685ff7-607b-416c-9382-2c00edecaa54/ovsdbserver-sb/0.log" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.268167 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.317342 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9813c97-677a-42c2-b526-576a4c5f2968-config\") pod \"e9813c97-677a-42c2-b526-576a4c5f2968\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.317402 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9813c97-677a-42c2-b526-576a4c5f2968-metrics-certs-tls-certs\") pod \"e9813c97-677a-42c2-b526-576a4c5f2968\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.317436 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e9813c97-677a-42c2-b526-576a4c5f2968-ovn-rundir\") pod \"e9813c97-677a-42c2-b526-576a4c5f2968\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.317492 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9813c97-677a-42c2-b526-576a4c5f2968-combined-ca-bundle\") pod \"e9813c97-677a-42c2-b526-576a4c5f2968\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.317598 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97bcs\" (UniqueName: \"kubernetes.io/projected/e9813c97-677a-42c2-b526-576a4c5f2968-kube-api-access-97bcs\") pod \"e9813c97-677a-42c2-b526-576a4c5f2968\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.317625 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e9813c97-677a-42c2-b526-576a4c5f2968-ovs-rundir\") pod \"e9813c97-677a-42c2-b526-576a4c5f2968\" (UID: \"e9813c97-677a-42c2-b526-576a4c5f2968\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.317753 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9813c97-677a-42c2-b526-576a4c5f2968-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "e9813c97-677a-42c2-b526-576a4c5f2968" (UID: "e9813c97-677a-42c2-b526-576a4c5f2968"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.319891 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9813c97-677a-42c2-b526-576a4c5f2968-config" (OuterVolumeSpecName: "config") pod "e9813c97-677a-42c2-b526-576a4c5f2968" (UID: "e9813c97-677a-42c2-b526-576a4c5f2968"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.320919 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9813c97-677a-42c2-b526-576a4c5f2968-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "e9813c97-677a-42c2-b526-576a4c5f2968" (UID: "e9813c97-677a-42c2-b526-576a4c5f2968"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.331116 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9813c97-677a-42c2-b526-576a4c5f2968-kube-api-access-97bcs" (OuterVolumeSpecName: "kube-api-access-97bcs") pod "e9813c97-677a-42c2-b526-576a4c5f2968" (UID: "e9813c97-677a-42c2-b526-576a4c5f2968"). InnerVolumeSpecName "kube-api-access-97bcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.332573 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-585fd6468c-8gx5n"] Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.332801 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-585fd6468c-8gx5n" podUID="446bd539-6bf4-42b1-ac19-40c8c80b2d45" containerName="proxy-httpd" containerID="cri-o://9d2ff242a1e29ad8407f6c796b021ec08d4a487ead665a7032ac8ddbf2177bbc" gracePeriod=30 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.333047 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-585fd6468c-8gx5n" podUID="446bd539-6bf4-42b1-ac19-40c8c80b2d45" containerName="proxy-server" containerID="cri-o://9a86b1d47a3009097a1c9b5e0d10b4d2e4d64194427ec3554b609bf3c481d41c" gracePeriod=30 Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.389116 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.422904 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7685ff7-607b-416c-9382-2c00edecaa54-config\") pod \"f7685ff7-607b-416c-9382-2c00edecaa54\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.423224 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7685ff7-607b-416c-9382-2c00edecaa54-scripts\") pod \"f7685ff7-607b-416c-9382-2c00edecaa54\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.423305 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mm56r\" (UniqueName: \"kubernetes.io/projected/f7685ff7-607b-416c-9382-2c00edecaa54-kube-api-access-mm56r\") pod \"f7685ff7-607b-416c-9382-2c00edecaa54\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.423371 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-metrics-certs-tls-certs\") pod \"f7685ff7-607b-416c-9382-2c00edecaa54\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.423497 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f7685ff7-607b-416c-9382-2c00edecaa54-ovsdb-rundir\") pod \"f7685ff7-607b-416c-9382-2c00edecaa54\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.423576 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-combined-ca-bundle\") pod \"f7685ff7-607b-416c-9382-2c00edecaa54\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.423640 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-ovsdbserver-sb-tls-certs\") pod \"f7685ff7-607b-416c-9382-2c00edecaa54\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.423710 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"f7685ff7-607b-416c-9382-2c00edecaa54\" (UID: \"f7685ff7-607b-416c-9382-2c00edecaa54\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.424171 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9813c97-677a-42c2-b526-576a4c5f2968-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.424240 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/e9813c97-677a-42c2-b526-576a4c5f2968-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.424292 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97bcs\" (UniqueName: \"kubernetes.io/projected/e9813c97-677a-42c2-b526-576a4c5f2968-kube-api-access-97bcs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.424343 4775 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/e9813c97-677a-42c2-b526-576a4c5f2968-ovs-rundir\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.424359 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7685ff7-607b-416c-9382-2c00edecaa54-config" (OuterVolumeSpecName: "config") pod "f7685ff7-607b-416c-9382-2c00edecaa54" (UID: "f7685ff7-607b-416c-9382-2c00edecaa54"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.426462 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7685ff7-607b-416c-9382-2c00edecaa54-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "f7685ff7-607b-416c-9382-2c00edecaa54" (UID: "f7685ff7-607b-416c-9382-2c00edecaa54"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.428397 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7685ff7-607b-416c-9382-2c00edecaa54-scripts" (OuterVolumeSpecName: "scripts") pod "f7685ff7-607b-416c-9382-2c00edecaa54" (UID: "f7685ff7-607b-416c-9382-2c00edecaa54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.435971 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.448519 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7685ff7-607b-416c-9382-2c00edecaa54-kube-api-access-mm56r" (OuterVolumeSpecName: "kube-api-access-mm56r") pod "f7685ff7-607b-416c-9382-2c00edecaa54" (UID: "f7685ff7-607b-416c-9382-2c00edecaa54"). InnerVolumeSpecName "kube-api-access-mm56r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.448742 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "f7685ff7-607b-416c-9382-2c00edecaa54" (UID: "f7685ff7-607b-416c-9382-2c00edecaa54"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.466729 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.515121 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.534764 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/764d109a-a65d-4fa6-ab71-a8ffb49e442e-openstack-config-secret\") pod \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.534804 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt4xd\" (UniqueName: \"kubernetes.io/projected/1364df9e-151f-40d0-9a5a-606a54021ef7-kube-api-access-rt4xd\") pod \"1364df9e-151f-40d0-9a5a-606a54021ef7\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.534901 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-config\") pod \"1364df9e-151f-40d0-9a5a-606a54021ef7\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.534933 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-ovsdbserver-sb\") pod \"1364df9e-151f-40d0-9a5a-606a54021ef7\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.534973 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcvnx\" (UniqueName: \"kubernetes.io/projected/764d109a-a65d-4fa6-ab71-a8ffb49e442e-kube-api-access-vcvnx\") pod \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.535018 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764d109a-a65d-4fa6-ab71-a8ffb49e442e-combined-ca-bundle\") pod \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.535039 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-ovsdbserver-nb\") pod \"1364df9e-151f-40d0-9a5a-606a54021ef7\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.535141 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/764d109a-a65d-4fa6-ab71-a8ffb49e442e-openstack-config\") pod \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\" (UID: \"764d109a-a65d-4fa6-ab71-a8ffb49e442e\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.535178 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-dns-swift-storage-0\") pod \"1364df9e-151f-40d0-9a5a-606a54021ef7\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.535222 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-dns-svc\") pod \"1364df9e-151f-40d0-9a5a-606a54021ef7\" (UID: \"1364df9e-151f-40d0-9a5a-606a54021ef7\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.535655 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f7685ff7-607b-416c-9382-2c00edecaa54-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.535667 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f7685ff7-607b-416c-9382-2c00edecaa54-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.535679 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mm56r\" (UniqueName: \"kubernetes.io/projected/f7685ff7-607b-416c-9382-2c00edecaa54-kube-api-access-mm56r\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.535687 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f7685ff7-607b-416c-9382-2c00edecaa54-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.535706 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.564014 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/764d109a-a65d-4fa6-ab71-a8ffb49e442e-kube-api-access-vcvnx" (OuterVolumeSpecName: "kube-api-access-vcvnx") pod "764d109a-a65d-4fa6-ab71-a8ffb49e442e" (UID: "764d109a-a65d-4fa6-ab71-a8ffb49e442e"). InnerVolumeSpecName "kube-api-access-vcvnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.581502 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1364df9e-151f-40d0-9a5a-606a54021ef7-kube-api-access-rt4xd" (OuterVolumeSpecName: "kube-api-access-rt4xd") pod "1364df9e-151f-40d0-9a5a-606a54021ef7" (UID: "1364df9e-151f-40d0-9a5a-606a54021ef7"). InnerVolumeSpecName "kube-api-access-rt4xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.636775 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-config-data\") pod \"163a6d17-be4c-47fc-8835-e218e8a803f1\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.636864 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-metrics-certs-tls-certs\") pod \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.636896 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.636969 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-config-data-custom\") pod \"163a6d17-be4c-47fc-8835-e218e8a803f1\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.637015 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/163a6d17-be4c-47fc-8835-e218e8a803f1-logs\") pod \"163a6d17-be4c-47fc-8835-e218e8a803f1\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.637123 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27jdw\" (UniqueName: \"kubernetes.io/projected/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-kube-api-access-27jdw\") pod \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.637146 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-ovsdb-rundir\") pod \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.637178 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-ovsdbserver-nb-tls-certs\") pod \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.637309 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64lhb\" (UniqueName: \"kubernetes.io/projected/163a6d17-be4c-47fc-8835-e218e8a803f1-kube-api-access-64lhb\") pod \"163a6d17-be4c-47fc-8835-e218e8a803f1\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.637374 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-config\") pod \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.637390 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-scripts\") pod \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.637493 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-combined-ca-bundle\") pod \"163a6d17-be4c-47fc-8835-e218e8a803f1\" (UID: \"163a6d17-be4c-47fc-8835-e218e8a803f1\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.637554 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-combined-ca-bundle\") pod \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\" (UID: \"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92\") " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.637910 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt4xd\" (UniqueName: \"kubernetes.io/projected/1364df9e-151f-40d0-9a5a-606a54021ef7-kube-api-access-rt4xd\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.637926 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcvnx\" (UniqueName: \"kubernetes.io/projected/764d109a-a65d-4fa6-ab71-a8ffb49e442e-kube-api-access-vcvnx\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.639555 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" (UID: "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.640167 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-config" (OuterVolumeSpecName: "config") pod "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" (UID: "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.649652 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/163a6d17-be4c-47fc-8835-e218e8a803f1-logs" (OuterVolumeSpecName: "logs") pod "163a6d17-be4c-47fc-8835-e218e8a803f1" (UID: "163a6d17-be4c-47fc-8835-e218e8a803f1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.666392 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-scripts" (OuterVolumeSpecName: "scripts") pod "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" (UID: "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.719908 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" (UID: "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.736708 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/163a6d17-be4c-47fc-8835-e218e8a803f1-kube-api-access-64lhb" (OuterVolumeSpecName: "kube-api-access-64lhb") pod "163a6d17-be4c-47fc-8835-e218e8a803f1" (UID: "163a6d17-be4c-47fc-8835-e218e8a803f1"). InnerVolumeSpecName "kube-api-access-64lhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.738098 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/764d109a-a65d-4fa6-ab71-a8ffb49e442e-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "764d109a-a65d-4fa6-ab71-a8ffb49e442e" (UID: "764d109a-a65d-4fa6-ab71-a8ffb49e442e"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.739271 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/163a6d17-be4c-47fc-8835-e218e8a803f1-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.739283 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.739293 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64lhb\" (UniqueName: \"kubernetes.io/projected/163a6d17-be4c-47fc-8835-e218e8a803f1-kube-api-access-64lhb\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.739302 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.739310 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.739328 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.739336 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/764d109a-a65d-4fa6-ab71-a8ffb49e442e-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.745348 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-kube-api-access-27jdw" (OuterVolumeSpecName: "kube-api-access-27jdw") pod "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" (UID: "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92"). InnerVolumeSpecName "kube-api-access-27jdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.774896 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "163a6d17-be4c-47fc-8835-e218e8a803f1" (UID: "163a6d17-be4c-47fc-8835-e218e8a803f1"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.840735 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.840791 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27jdw\" (UniqueName: \"kubernetes.io/projected/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-kube-api-access-27jdw\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:52 crc kubenswrapper[4775]: E1002 02:03:52.942344 4775 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 02 02:03:52 crc kubenswrapper[4775]: E1002 02:03:52.942684 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data podName:6638baa9-b367-40d1-8111-673dca6434ad nodeName:}" failed. No retries permitted until 2025-10-02 02:03:56.942669637 +0000 UTC m=+1374.109413677 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data") pod "rabbitmq-server-0" (UID: "6638baa9-b367-40d1-8111-673dca6434ad") : configmap "rabbitmq-config-data" not found Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.958496 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9813c97-677a-42c2-b526-576a4c5f2968-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9813c97-677a-42c2-b526-576a4c5f2968" (UID: "e9813c97-677a-42c2-b526-576a4c5f2968"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:52 crc kubenswrapper[4775]: I1002 02:03:52.990928 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.000638 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f7685ff7-607b-416c-9382-2c00edecaa54" (UID: "f7685ff7-607b-416c-9382-2c00edecaa54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.044823 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9813c97-677a-42c2-b526-576a4c5f2968-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.044857 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.044897 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.046140 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/764d109a-a65d-4fa6-ab71-a8ffb49e442e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "764d109a-a65d-4fa6-ab71-a8ffb49e442e" (UID: "764d109a-a65d-4fa6-ab71-a8ffb49e442e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.073929 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1364df9e-151f-40d0-9a5a-606a54021ef7" (UID: "1364df9e-151f-40d0-9a5a-606a54021ef7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.100840 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.102090 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1364df9e-151f-40d0-9a5a-606a54021ef7" (UID: "1364df9e-151f-40d0-9a5a-606a54021ef7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.117428 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" (UID: "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.129572 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1364df9e-151f-40d0-9a5a-606a54021ef7" (UID: "1364df9e-151f-40d0-9a5a-606a54021ef7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.153056 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.153083 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.153092 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/764d109a-a65d-4fa6-ab71-a8ffb49e442e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.153106 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.153117 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.153125 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.216902 4775 generic.go:334] "Generic (PLEG): container finished" podID="af16c5f4-fcd0-4450-a8b7-789ff104e477" containerID="ae9bd041c2c057dee311d85415e1bfd222d0109f3f01cb2fcd2149362f273f6e" exitCode=0 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.217096 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" event={"ID":"af16c5f4-fcd0-4450-a8b7-789ff104e477","Type":"ContainerDied","Data":"ae9bd041c2c057dee311d85415e1bfd222d0109f3f01cb2fcd2149362f273f6e"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.217124 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" event={"ID":"af16c5f4-fcd0-4450-a8b7-789ff104e477","Type":"ContainerDied","Data":"4c3c3ac65982b5585d40b0110d0e93c428516101d184f5a470d6b08eb9b85bfb"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.217136 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c3c3ac65982b5585d40b0110d0e93c428516101d184f5a470d6b08eb9b85bfb" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.219656 4775 generic.go:334] "Generic (PLEG): container finished" podID="1e5e6844-45f6-4fd9-ac4f-3df2086164eb" containerID="4112b6fd527ff9fca7a2399175d5a6860d1e392cc729ed9cbc4ea16d1712bb78" exitCode=0 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.219716 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementf75b-account-delete-dl78c" event={"ID":"1e5e6844-45f6-4fd9-ac4f-3df2086164eb","Type":"ContainerDied","Data":"4112b6fd527ff9fca7a2399175d5a6860d1e392cc729ed9cbc4ea16d1712bb78"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.225175 4775 generic.go:334] "Generic (PLEG): container finished" podID="cdea921f-d632-4463-a011-2e6f23f70546" containerID="a430bd7d375548fc7aa2abbd9b702aea48844495f11cfa9ca7aff1e02c52b8b9" exitCode=0 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.225364 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance95b1-account-delete-4jqdz" event={"ID":"cdea921f-d632-4463-a011-2e6f23f70546","Type":"ContainerDied","Data":"a430bd7d375548fc7aa2abbd9b702aea48844495f11cfa9ca7aff1e02c52b8b9"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.225529 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance95b1-account-delete-4jqdz" event={"ID":"cdea921f-d632-4463-a011-2e6f23f70546","Type":"ContainerStarted","Data":"0fbfb28794714072a7ba90e82444620bf54fb99d422e573e75332f9090b40ee9"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.232010 4775 generic.go:334] "Generic (PLEG): container finished" podID="ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a" containerID="cda556d95d519500729d0be04b15c0114590b6b3efb514dfac80d925c08696b9" exitCode=0 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.232073 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanb963-account-delete-z5bdz" event={"ID":"ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a","Type":"ContainerDied","Data":"cda556d95d519500729d0be04b15c0114590b6b3efb514dfac80d925c08696b9"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.232097 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanb963-account-delete-z5bdz" event={"ID":"ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a","Type":"ContainerStarted","Data":"a3934827a0b3db7d670af62754d194702fd060d18de214524d551c1e2f125eae"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.233581 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "163a6d17-be4c-47fc-8835-e218e8a803f1" (UID: "163a6d17-be4c-47fc-8835-e218e8a803f1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.244712 4775 generic.go:334] "Generic (PLEG): container finished" podID="46453bb2-6af0-4dd4-83ba-b14bf3f96310" containerID="b66fc6b4d3be69dcb9c2d949e6dfe8d6ab6ddb59f7b8ad17c8380c0b6034ccf6" exitCode=0 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.244796 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"46453bb2-6af0-4dd4-83ba-b14bf3f96310","Type":"ContainerDied","Data":"b66fc6b4d3be69dcb9c2d949e6dfe8d6ab6ddb59f7b8ad17c8380c0b6034ccf6"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.244842 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"46453bb2-6af0-4dd4-83ba-b14bf3f96310","Type":"ContainerDied","Data":"e79277171e00d6730ea668e1d3a8b2e2495caf310572d88755b3b04fcb9a6364"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.244855 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e79277171e00d6730ea668e1d3a8b2e2495caf310572d88755b3b04fcb9a6364" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.248984 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.249011 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92","Type":"ContainerDied","Data":"e194882b58a7d56c86b9f7a5e4ff2fc880c4254d1368412ee77182aacb63d04d"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.250283 4775 scope.go:117] "RemoveContainer" containerID="b8d49f290ce08384af0005110e4c784aaece1c7d6ad52d6aaa1e3385f7367c8c" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.255418 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3732523-d2de-45c9-ac69-770197f413f3" containerID="ddabebd6b3c747fa7b6ca77959080b27a169f7f5ace3d87005ed768fd6e84492" exitCode=0 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.255535 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder5592-account-delete-mhfmd" event={"ID":"e3732523-d2de-45c9-ac69-770197f413f3","Type":"ContainerDied","Data":"ddabebd6b3c747fa7b6ca77959080b27a169f7f5ace3d87005ed768fd6e84492"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.258390 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.264683 4775 generic.go:334] "Generic (PLEG): container finished" podID="a5c60336-6e25-4c68-98e6-52801f34578b" containerID="38c453bdb403040517a2812e8643acfa55f8124baa338e04c4028a69bd4cbbd1" exitCode=0 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.264746 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a5c60336-6e25-4c68-98e6-52801f34578b","Type":"ContainerDied","Data":"38c453bdb403040517a2812e8643acfa55f8124baa338e04c4028a69bd4cbbd1"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.264777 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a5c60336-6e25-4c68-98e6-52801f34578b","Type":"ContainerDied","Data":"dbe64d055f38f78c1200ca47b2599fd920af93a48d37676c014b7689ac274217"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.264788 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbe64d055f38f78c1200ca47b2599fd920af93a48d37676c014b7689ac274217" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.279296 4775 generic.go:334] "Generic (PLEG): container finished" podID="163a6d17-be4c-47fc-8835-e218e8a803f1" containerID="979478a6b7d6bf80d7823b67cd948eb55c9e92647ad8b39e7e0c75474a16acc7" exitCode=0 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.279356 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-547cbfbdcc-nk6pz" event={"ID":"163a6d17-be4c-47fc-8835-e218e8a803f1","Type":"ContainerDied","Data":"979478a6b7d6bf80d7823b67cd948eb55c9e92647ad8b39e7e0c75474a16acc7"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.279381 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-547cbfbdcc-nk6pz" event={"ID":"163a6d17-be4c-47fc-8835-e218e8a803f1","Type":"ContainerDied","Data":"3ce9875a5a256a84090325bfbb15e5ce36ac47673c839e8cef7c8c64a9706094"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.279435 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-547cbfbdcc-nk6pz" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.291518 4775 generic.go:334] "Generic (PLEG): container finished" podID="446bd539-6bf4-42b1-ac19-40c8c80b2d45" containerID="9a86b1d47a3009097a1c9b5e0d10b4d2e4d64194427ec3554b609bf3c481d41c" exitCode=0 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.291679 4775 generic.go:334] "Generic (PLEG): container finished" podID="446bd539-6bf4-42b1-ac19-40c8c80b2d45" containerID="9d2ff242a1e29ad8407f6c796b021ec08d4a487ead665a7032ac8ddbf2177bbc" exitCode=0 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.291826 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-585fd6468c-8gx5n" event={"ID":"446bd539-6bf4-42b1-ac19-40c8c80b2d45","Type":"ContainerDied","Data":"9a86b1d47a3009097a1c9b5e0d10b4d2e4d64194427ec3554b609bf3c481d41c"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.291923 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-585fd6468c-8gx5n" event={"ID":"446bd539-6bf4-42b1-ac19-40c8c80b2d45","Type":"ContainerDied","Data":"9d2ff242a1e29ad8407f6c796b021ec08d4a487ead665a7032ac8ddbf2177bbc"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.318184 4775 generic.go:334] "Generic (PLEG): container finished" podID="3adc5647-1145-4c44-9ecc-66d1d9a19023" containerID="1209a9ffa5fb6c6952b1b8effd2ebc2e2fb407ab840357e9b3b1f508504a915a" exitCode=0 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.331486 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7szxv" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.331500 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.332987 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/764d109a-a65d-4fa6-ab71-a8ffb49e442e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "764d109a-a65d-4fa6-ab71-a8ffb49e442e" (UID: "764d109a-a65d-4fa6-ab71-a8ffb49e442e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.331546 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.318404 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3adc5647-1145-4c44-9ecc-66d1d9a19023","Type":"ContainerDied","Data":"1209a9ffa5fb6c6952b1b8effd2ebc2e2fb407ab840357e9b3b1f508504a915a"} Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.331584 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-8jbsr" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.333160 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" (UID: "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.333421 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1364df9e-151f-40d0-9a5a-606a54021ef7" (UID: "1364df9e-151f-40d0-9a5a-606a54021ef7"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.359517 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.359546 4775 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.359556 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/764d109a-a65d-4fa6-ab71-a8ffb49e442e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.359911 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9813c97-677a-42c2-b526-576a4c5f2968-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "e9813c97-677a-42c2-b526-576a4c5f2968" (UID: "e9813c97-677a-42c2-b526-576a4c5f2968"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.363471 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-config" (OuterVolumeSpecName: "config") pod "1364df9e-151f-40d0-9a5a-606a54021ef7" (UID: "1364df9e-151f-40d0-9a5a-606a54021ef7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.399446 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-config-data" (OuterVolumeSpecName: "config-data") pod "163a6d17-be4c-47fc-8835-e218e8a803f1" (UID: "163a6d17-be4c-47fc-8835-e218e8a803f1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.427158 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "f7685ff7-607b-416c-9382-2c00edecaa54" (UID: "f7685ff7-607b-416c-9382-2c00edecaa54"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.428739 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "f7685ff7-607b-416c-9382-2c00edecaa54" (UID: "f7685ff7-607b-416c-9382-2c00edecaa54"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.452762 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" (UID: "c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.461677 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9813c97-677a-42c2-b526-576a4c5f2968-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.461708 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.461717 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1364df9e-151f-40d0-9a5a-606a54021ef7-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.461725 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/163a6d17-be4c-47fc-8835-e218e8a803f1-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.461734 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f7685ff7-607b-416c-9382-2c00edecaa54-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.461744 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:53 crc kubenswrapper[4775]: E1002 02:03:53.565044 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b66fc6b4d3be69dcb9c2d949e6dfe8d6ab6ddb59f7b8ad17c8380c0b6034ccf6 is running failed: container process not found" containerID="b66fc6b4d3be69dcb9c2d949e6dfe8d6ab6ddb59f7b8ad17c8380c0b6034ccf6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 02:03:53 crc kubenswrapper[4775]: E1002 02:03:53.565526 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b66fc6b4d3be69dcb9c2d949e6dfe8d6ab6ddb59f7b8ad17c8380c0b6034ccf6 is running failed: container process not found" containerID="b66fc6b4d3be69dcb9c2d949e6dfe8d6ab6ddb59f7b8ad17c8380c0b6034ccf6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 02:03:53 crc kubenswrapper[4775]: E1002 02:03:53.566083 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b66fc6b4d3be69dcb9c2d949e6dfe8d6ab6ddb59f7b8ad17c8380c0b6034ccf6 is running failed: container process not found" containerID="b66fc6b4d3be69dcb9c2d949e6dfe8d6ab6ddb59f7b8ad17c8380c0b6034ccf6" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 02:03:53 crc kubenswrapper[4775]: E1002 02:03:53.566123 4775 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of b66fc6b4d3be69dcb9c2d949e6dfe8d6ab6ddb59f7b8ad17c8380c0b6034ccf6 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="46453bb2-6af0-4dd4-83ba-b14bf3f96310" containerName="nova-cell1-conductor-conductor" Oct 02 02:03:53 crc kubenswrapper[4775]: E1002 02:03:53.624567 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1209a9ffa5fb6c6952b1b8effd2ebc2e2fb407ab840357e9b3b1f508504a915a is running failed: container process not found" containerID="1209a9ffa5fb6c6952b1b8effd2ebc2e2fb407ab840357e9b3b1f508504a915a" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 02 02:03:53 crc kubenswrapper[4775]: E1002 02:03:53.624832 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1209a9ffa5fb6c6952b1b8effd2ebc2e2fb407ab840357e9b3b1f508504a915a is running failed: container process not found" containerID="1209a9ffa5fb6c6952b1b8effd2ebc2e2fb407ab840357e9b3b1f508504a915a" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 02 02:03:53 crc kubenswrapper[4775]: E1002 02:03:53.625052 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1209a9ffa5fb6c6952b1b8effd2ebc2e2fb407ab840357e9b3b1f508504a915a is running failed: container process not found" containerID="1209a9ffa5fb6c6952b1b8effd2ebc2e2fb407ab840357e9b3b1f508504a915a" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Oct 02 02:03:53 crc kubenswrapper[4775]: E1002 02:03:53.625078 4775 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 1209a9ffa5fb6c6952b1b8effd2ebc2e2fb407ab840357e9b3b1f508504a915a is running failed: container process not found" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="3adc5647-1145-4c44-9ecc-66d1d9a19023" containerName="galera" Oct 02 02:03:53 crc kubenswrapper[4775]: E1002 02:03:53.665598 4775 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 02 02:03:53 crc kubenswrapper[4775]: E1002 02:03:53.665667 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data podName:fb0a7ddd-1558-4ad8-a8f4-e140306d19f9 nodeName:}" failed. No retries permitted until 2025-10-02 02:03:57.665650532 +0000 UTC m=+1374.832394572 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data") pod "rabbitmq-cell1-server-0" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9") : configmap "rabbitmq-cell1-config-data" not found Oct 02 02:03:53 crc kubenswrapper[4775]: E1002 02:03:53.697123 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 02:03:53 crc kubenswrapper[4775]: E1002 02:03:53.701466 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 02:03:53 crc kubenswrapper[4775]: E1002 02:03:53.707603 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 02:03:53 crc kubenswrapper[4775]: E1002 02:03:53.707650 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a64f8ba3-65b9-489f-ace1-757f3ecce6b4" containerName="nova-scheduler-scheduler" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.803284 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="764d109a-a65d-4fa6-ab71-a8ffb49e442e" path="/var/lib/kubelet/pods/764d109a-a65d-4fa6-ab71-a8ffb49e442e/volumes" Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.810262 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.810589 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="ceilometer-central-agent" containerID="cri-o://8df9de2a0e96e291b49efa835c2476332824829ee66a90f08c5fa720d9e412f8" gracePeriod=30 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.810775 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="proxy-httpd" containerID="cri-o://1c6c4af35ca368385ff971101cf668bbedef2ac9915012172cc65ff5519d556e" gracePeriod=30 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.810825 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="sg-core" containerID="cri-o://0a11b7f88b7e2c013d1797c32a7ec78c81f775682c0323ea8b58d4f35f6e04b7" gracePeriod=30 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.810866 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="ceilometer-notification-agent" containerID="cri-o://4feb9154e0329bc25829a25b496b1b2573ea599dcffdc17eb82bbc2b54f58aa4" gracePeriod=30 Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.853778 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:03:53 crc kubenswrapper[4775]: I1002 02:03:53.854029 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="24ef8940-eb43-459c-9cfe-854df57cfe74" containerName="kube-state-metrics" containerID="cri-o://7faaf4aaa749333ce8970b0b3bfce0b23a77558417d2d73450a187551b1a850f" gracePeriod=30 Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.139050 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.139371 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="406dd24f-d800-44be-b33d-019c532c2feb" containerName="memcached" containerID="cri-o://062b09ca95edb1ddddcb2de55757014ea328d800d380273e8fb5c4e3bb488c11" gracePeriod=30 Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.235711 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2smp7"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.242894 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2smp7"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.317490 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-vqhvp"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.340257 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-vqhvp"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.340542 4775 scope.go:117] "RemoveContainer" containerID="0089ce10498396392f25a55b07f3d03f418bbb5e5d51703da8bdb90f1f8cc858" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.397725 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-cron-29322841-j7zjm"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.411686 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.431755 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-cron-29322841-j7zjm"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.434153 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.434399 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placementf75b-account-delete-dl78c" event={"ID":"1e5e6844-45f6-4fd9-ac4f-3df2086164eb","Type":"ContainerDied","Data":"aa6896692c152e1046b85695e2663ae1516576a62c99a10f86e65ba674a54200"} Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.434419 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa6896692c152e1046b85695e2663ae1516576a62c99a10f86e65ba674a54200" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.451765 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-564487c87-zppvw"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.451945 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-564487c87-zppvw" podUID="1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e" containerName="keystone-api" containerID="cri-o://7ad48d14db57c9f5d0c7e7d3fcaf64adb99c760e739a396090a74eb51c2a9919" gracePeriod=30 Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.460213 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.463140 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance95b1-account-delete-4jqdz" event={"ID":"cdea921f-d632-4463-a011-2e6f23f70546","Type":"ContainerDied","Data":"0fbfb28794714072a7ba90e82444620bf54fb99d422e573e75332f9090b40ee9"} Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.463172 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fbfb28794714072a7ba90e82444620bf54fb99d422e573e75332f9090b40ee9" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.477127 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-j4qvx"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.483345 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.484930 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanb963-account-delete-z5bdz" event={"ID":"ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a","Type":"ContainerDied","Data":"a3934827a0b3db7d670af62754d194702fd060d18de214524d551c1e2f125eae"} Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.484985 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3934827a0b3db7d670af62754d194702fd060d18de214524d551c1e2f125eae" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.485066 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance95b1-account-delete-4jqdz" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.485165 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.485752 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-j4qvx"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.495978 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7e50-account-create-nlkzk"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.508033 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanb963-account-delete-z5bdz" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.510490 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7e50-account-create-nlkzk"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.520113 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="5bca47d0-79b7-4992-867b-d5e05e02cdc8" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.163:8776/healthcheck\": read tcp 10.217.0.2:43436->10.217.0.163:8776: read: connection reset by peer" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.520238 4775 scope.go:117] "RemoveContainer" containerID="979478a6b7d6bf80d7823b67cd948eb55c9e92647ad8b39e7e0c75474a16acc7" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.520492 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder5592-account-delete-mhfmd" event={"ID":"e3732523-d2de-45c9-ac69-770197f413f3","Type":"ContainerDied","Data":"a069d853f5876b595def267dae48810cc164f576380990b4bbc0c9615bdfff3a"} Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.520515 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a069d853f5876b595def267dae48810cc164f576380990b4bbc0c9615bdfff3a" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.537595 4775 generic.go:334] "Generic (PLEG): container finished" podID="56813183-3dfc-430c-aaac-bda27d91340b" containerID="d46af6901507bf3d2c6f5cd7933a7df11f703991822ecd42a838e772bb645206" exitCode=0 Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.537649 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-777987c7d4-d5vzz" event={"ID":"56813183-3dfc-430c-aaac-bda27d91340b","Type":"ContainerDied","Data":"d46af6901507bf3d2c6f5cd7933a7df11f703991822ecd42a838e772bb645206"} Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.537813 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.541787 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-8jbsr"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.543609 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5592-account-delete-mhfmd" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545078 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-config-data\") pod \"a5c60336-6e25-4c68-98e6-52801f34578b\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545130 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qfwn\" (UniqueName: \"kubernetes.io/projected/a5c60336-6e25-4c68-98e6-52801f34578b-kube-api-access-2qfwn\") pod \"a5c60336-6e25-4c68-98e6-52801f34578b\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545172 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lzz9\" (UniqueName: \"kubernetes.io/projected/cdea921f-d632-4463-a011-2e6f23f70546-kube-api-access-5lzz9\") pod \"cdea921f-d632-4463-a011-2e6f23f70546\" (UID: \"cdea921f-d632-4463-a011-2e6f23f70546\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545194 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-secrets\") pod \"3adc5647-1145-4c44-9ecc-66d1d9a19023\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545239 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-config-data-default\") pod \"3adc5647-1145-4c44-9ecc-66d1d9a19023\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545277 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3adc5647-1145-4c44-9ecc-66d1d9a19023-config-data-generated\") pod \"3adc5647-1145-4c44-9ecc-66d1d9a19023\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545310 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-vencrypt-tls-certs\") pod \"a5c60336-6e25-4c68-98e6-52801f34578b\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545341 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-config-data\") pod \"af16c5f4-fcd0-4450-a8b7-789ff104e477\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545361 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-config-data-custom\") pod \"af16c5f4-fcd0-4450-a8b7-789ff104e477\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545383 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-galera-tls-certs\") pod \"3adc5647-1145-4c44-9ecc-66d1d9a19023\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545417 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-kolla-config\") pod \"3adc5647-1145-4c44-9ecc-66d1d9a19023\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545440 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af16c5f4-fcd0-4450-a8b7-789ff104e477-logs\") pod \"af16c5f4-fcd0-4450-a8b7-789ff104e477\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545464 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-combined-ca-bundle\") pod \"3adc5647-1145-4c44-9ecc-66d1d9a19023\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545499 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-operator-scripts\") pod \"3adc5647-1145-4c44-9ecc-66d1d9a19023\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545521 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-nova-novncproxy-tls-certs\") pod \"a5c60336-6e25-4c68-98e6-52801f34578b\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545702 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqjt7\" (UniqueName: \"kubernetes.io/projected/af16c5f4-fcd0-4450-a8b7-789ff104e477-kube-api-access-dqjt7\") pod \"af16c5f4-fcd0-4450-a8b7-789ff104e477\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545752 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-combined-ca-bundle\") pod \"af16c5f4-fcd0-4450-a8b7-789ff104e477\" (UID: \"af16c5f4-fcd0-4450-a8b7-789ff104e477\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545772 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46453bb2-6af0-4dd4-83ba-b14bf3f96310-config-data\") pod \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\" (UID: \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545795 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"3adc5647-1145-4c44-9ecc-66d1d9a19023\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545813 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46453bb2-6af0-4dd4-83ba-b14bf3f96310-combined-ca-bundle\") pod \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\" (UID: \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545846 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rglgf\" (UniqueName: \"kubernetes.io/projected/46453bb2-6af0-4dd4-83ba-b14bf3f96310-kube-api-access-rglgf\") pod \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\" (UID: \"46453bb2-6af0-4dd4-83ba-b14bf3f96310\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545869 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-combined-ca-bundle\") pod \"a5c60336-6e25-4c68-98e6-52801f34578b\" (UID: \"a5c60336-6e25-4c68-98e6-52801f34578b\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.545884 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wph9j\" (UniqueName: \"kubernetes.io/projected/3adc5647-1145-4c44-9ecc-66d1d9a19023-kube-api-access-wph9j\") pod \"3adc5647-1145-4c44-9ecc-66d1d9a19023\" (UID: \"3adc5647-1145-4c44-9ecc-66d1d9a19023\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.547459 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3adc5647-1145-4c44-9ecc-66d1d9a19023","Type":"ContainerDied","Data":"7fa5b0f45bdda8d329e73a560098dae21985e99f3cfcb9a5ed11b49f0a4f88f7"} Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.547557 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.548322 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "3adc5647-1145-4c44-9ecc-66d1d9a19023" (UID: "3adc5647-1145-4c44-9ecc-66d1d9a19023"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.551110 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3adc5647-1145-4c44-9ecc-66d1d9a19023" (UID: "3adc5647-1145-4c44-9ecc-66d1d9a19023"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.560865 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-8jbsr"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.561499 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af16c5f4-fcd0-4450-a8b7-789ff104e477-logs" (OuterVolumeSpecName: "logs") pod "af16c5f4-fcd0-4450-a8b7-789ff104e477" (UID: "af16c5f4-fcd0-4450-a8b7-789ff104e477"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.563801 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3adc5647-1145-4c44-9ecc-66d1d9a19023-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "3adc5647-1145-4c44-9ecc-66d1d9a19023" (UID: "3adc5647-1145-4c44-9ecc-66d1d9a19023"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.573405 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-585fd6468c-8gx5n" event={"ID":"446bd539-6bf4-42b1-ac19-40c8c80b2d45","Type":"ContainerDied","Data":"66aa2c1759ddb8693183396b39f057b3cb58c4c2ae61f8f748b5bf95acd9d362"} Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.573599 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-585fd6468c-8gx5n" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.575341 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdea921f-d632-4463-a011-2e6f23f70546-kube-api-access-5lzz9" (OuterVolumeSpecName: "kube-api-access-5lzz9") pod "cdea921f-d632-4463-a011-2e6f23f70546" (UID: "cdea921f-d632-4463-a011-2e6f23f70546"). InnerVolumeSpecName "kube-api-access-5lzz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.578990 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "3adc5647-1145-4c44-9ecc-66d1d9a19023" (UID: "3adc5647-1145-4c44-9ecc-66d1d9a19023"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.588072 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.591174 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af16c5f4-fcd0-4450-a8b7-789ff104e477-kube-api-access-dqjt7" (OuterVolumeSpecName: "kube-api-access-dqjt7") pod "af16c5f4-fcd0-4450-a8b7-789ff104e477" (UID: "af16c5f4-fcd0-4450-a8b7-789ff104e477"). InnerVolumeSpecName "kube-api-access-dqjt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.601856 4775 generic.go:334] "Generic (PLEG): container finished" podID="d2d45342-c4de-48db-8f6b-23f26103497b" containerID="1c6c4af35ca368385ff971101cf668bbedef2ac9915012172cc65ff5519d556e" exitCode=0 Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.601883 4775 generic.go:334] "Generic (PLEG): container finished" podID="d2d45342-c4de-48db-8f6b-23f26103497b" containerID="0a11b7f88b7e2c013d1797c32a7ec78c81f775682c0323ea8b58d4f35f6e04b7" exitCode=2 Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.601920 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2d45342-c4de-48db-8f6b-23f26103497b","Type":"ContainerDied","Data":"1c6c4af35ca368385ff971101cf668bbedef2ac9915012172cc65ff5519d556e"} Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.601986 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2d45342-c4de-48db-8f6b-23f26103497b","Type":"ContainerDied","Data":"0a11b7f88b7e2c013d1797c32a7ec78c81f775682c0323ea8b58d4f35f6e04b7"} Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.602056 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.602747 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-secrets" (OuterVolumeSpecName: "secrets") pod "3adc5647-1145-4c44-9ecc-66d1d9a19023" (UID: "3adc5647-1145-4c44-9ecc-66d1d9a19023"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.607552 4775 generic.go:334] "Generic (PLEG): container finished" podID="24ef8940-eb43-459c-9cfe-854df57cfe74" containerID="7faaf4aaa749333ce8970b0b3bfce0b23a77558417d2d73450a187551b1a850f" exitCode=2 Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.607593 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"24ef8940-eb43-459c-9cfe-854df57cfe74","Type":"ContainerDied","Data":"7faaf4aaa749333ce8970b0b3bfce0b23a77558417d2d73450a187551b1a850f"} Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.608408 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "mysql-db") pod "3adc5647-1145-4c44-9ecc-66d1d9a19023" (UID: "3adc5647-1145-4c44-9ecc-66d1d9a19023"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.647308 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/446bd539-6bf4-42b1-ac19-40c8c80b2d45-log-httpd\") pod \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.647354 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2klh\" (UniqueName: \"kubernetes.io/projected/e3732523-d2de-45c9-ac69-770197f413f3-kube-api-access-g2klh\") pod \"e3732523-d2de-45c9-ac69-770197f413f3\" (UID: \"e3732523-d2de-45c9-ac69-770197f413f3\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.647465 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-public-tls-certs\") pod \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.647538 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/446bd539-6bf4-42b1-ac19-40c8c80b2d45-run-httpd\") pod \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.647558 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/446bd539-6bf4-42b1-ac19-40c8c80b2d45-etc-swift\") pod \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.647621 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dsqxs\" (UniqueName: \"kubernetes.io/projected/446bd539-6bf4-42b1-ac19-40c8c80b2d45-kube-api-access-dsqxs\") pod \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.647638 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-config-data\") pod \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.647692 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-combined-ca-bundle\") pod \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.647724 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd47g\" (UniqueName: \"kubernetes.io/projected/ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a-kube-api-access-jd47g\") pod \"ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a\" (UID: \"ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.647746 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-internal-tls-certs\") pod \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\" (UID: \"446bd539-6bf4-42b1-ac19-40c8c80b2d45\") " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.648477 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.648509 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lzz9\" (UniqueName: \"kubernetes.io/projected/cdea921f-d632-4463-a011-2e6f23f70546-kube-api-access-5lzz9\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.648537 4775 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.648549 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.648560 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3adc5647-1145-4c44-9ecc-66d1d9a19023-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.648571 4775 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.648579 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af16c5f4-fcd0-4450-a8b7-789ff104e477-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.648587 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3adc5647-1145-4c44-9ecc-66d1d9a19023-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.648598 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqjt7\" (UniqueName: \"kubernetes.io/projected/af16c5f4-fcd0-4450-a8b7-789ff104e477-kube-api-access-dqjt7\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.660863 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46453bb2-6af0-4dd4-83ba-b14bf3f96310-kube-api-access-rglgf" (OuterVolumeSpecName: "kube-api-access-rglgf") pod "46453bb2-6af0-4dd4-83ba-b14bf3f96310" (UID: "46453bb2-6af0-4dd4-83ba-b14bf3f96310"). InnerVolumeSpecName "kube-api-access-rglgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.661309 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3adc5647-1145-4c44-9ecc-66d1d9a19023-kube-api-access-wph9j" (OuterVolumeSpecName: "kube-api-access-wph9j") pod "3adc5647-1145-4c44-9ecc-66d1d9a19023" (UID: "3adc5647-1145-4c44-9ecc-66d1d9a19023"). InnerVolumeSpecName "kube-api-access-wph9j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.661367 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5c60336-6e25-4c68-98e6-52801f34578b-kube-api-access-2qfwn" (OuterVolumeSpecName: "kube-api-access-2qfwn") pod "a5c60336-6e25-4c68-98e6-52801f34578b" (UID: "a5c60336-6e25-4c68-98e6-52801f34578b"). InnerVolumeSpecName "kube-api-access-2qfwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.663855 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/446bd539-6bf4-42b1-ac19-40c8c80b2d45-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "446bd539-6bf4-42b1-ac19-40c8c80b2d45" (UID: "446bd539-6bf4-42b1-ac19-40c8c80b2d45"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.664251 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/446bd539-6bf4-42b1-ac19-40c8c80b2d45-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "446bd539-6bf4-42b1-ac19-40c8c80b2d45" (UID: "446bd539-6bf4-42b1-ac19-40c8c80b2d45"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.682028 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3732523-d2de-45c9-ac69-770197f413f3-kube-api-access-g2klh" (OuterVolumeSpecName: "kube-api-access-g2klh") pod "e3732523-d2de-45c9-ac69-770197f413f3" (UID: "e3732523-d2de-45c9-ac69-770197f413f3"). InnerVolumeSpecName "kube-api-access-g2klh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.682117 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.699507 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/446bd539-6bf4-42b1-ac19-40c8c80b2d45-kube-api-access-dsqxs" (OuterVolumeSpecName: "kube-api-access-dsqxs") pod "446bd539-6bf4-42b1-ac19-40c8c80b2d45" (UID: "446bd539-6bf4-42b1-ac19-40c8c80b2d45"). InnerVolumeSpecName "kube-api-access-dsqxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.701480 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.708175 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-547cbfbdcc-nk6pz"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.729075 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/446bd539-6bf4-42b1-ac19-40c8c80b2d45-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "446bd539-6bf4-42b1-ac19-40c8c80b2d45" (UID: "446bd539-6bf4-42b1-ac19-40c8c80b2d45"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.729207 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a-kube-api-access-jd47g" (OuterVolumeSpecName: "kube-api-access-jd47g") pod "ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a" (UID: "ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a"). InnerVolumeSpecName "kube-api-access-jd47g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.734304 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "af16c5f4-fcd0-4450-a8b7-789ff104e477" (UID: "af16c5f4-fcd0-4450-a8b7-789ff104e477"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.741103 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-547cbfbdcc-nk6pz"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.775671 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wph9j\" (UniqueName: \"kubernetes.io/projected/3adc5647-1145-4c44-9ecc-66d1d9a19023-kube-api-access-wph9j\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.775695 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qfwn\" (UniqueName: \"kubernetes.io/projected/a5c60336-6e25-4c68-98e6-52801f34578b-kube-api-access-2qfwn\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.775705 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/446bd539-6bf4-42b1-ac19-40c8c80b2d45-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.775814 4775 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/446bd539-6bf4-42b1-ac19-40c8c80b2d45-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.775826 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.775836 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dsqxs\" (UniqueName: \"kubernetes.io/projected/446bd539-6bf4-42b1-ac19-40c8c80b2d45-kube-api-access-dsqxs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.775845 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd47g\" (UniqueName: \"kubernetes.io/projected/ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a-kube-api-access-jd47g\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.775854 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/446bd539-6bf4-42b1-ac19-40c8c80b2d45-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.775862 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2klh\" (UniqueName: \"kubernetes.io/projected/e3732523-d2de-45c9-ac69-770197f413f3-kube-api-access-g2klh\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.775976 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rglgf\" (UniqueName: \"kubernetes.io/projected/46453bb2-6af0-4dd4-83ba-b14bf3f96310-kube-api-access-rglgf\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.802064 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-7szxv"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.808170 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5e322679-b95d-4328-b03b-0be7f0e82ac3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:51682->10.217.0.202:8775: read: connection reset by peer" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.808306 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="5e322679-b95d-4328-b03b-0be7f0e82ac3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:51690->10.217.0.202:8775: read: connection reset by peer" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.815761 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3adc5647-1145-4c44-9ecc-66d1d9a19023" (UID: "3adc5647-1145-4c44-9ecc-66d1d9a19023"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.825135 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-7szxv"] Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.838125 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46453bb2-6af0-4dd4-83ba-b14bf3f96310-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46453bb2-6af0-4dd4-83ba-b14bf3f96310" (UID: "46453bb2-6af0-4dd4-83ba-b14bf3f96310"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.841349 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af16c5f4-fcd0-4450-a8b7-789ff104e477" (UID: "af16c5f4-fcd0-4450-a8b7-789ff104e477"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.852353 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5c60336-6e25-4c68-98e6-52801f34578b" (UID: "a5c60336-6e25-4c68-98e6-52801f34578b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.853806 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.863826 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-config-data" (OuterVolumeSpecName: "config-data") pod "a5c60336-6e25-4c68-98e6-52801f34578b" (UID: "a5c60336-6e25-4c68-98e6-52801f34578b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.876897 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.876923 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.876932 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.876941 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46453bb2-6af0-4dd4-83ba-b14bf3f96310-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.877106 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.877118 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.953750 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="002eb223-af34-4c80-9570-894add9b0e3c" containerName="galera" containerID="cri-o://8569575c44b11cca45a36d7a16be245143dd1895ca868b19ec6f5372e6eb3a4d" gracePeriod=30 Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.973341 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "3adc5647-1145-4c44-9ecc-66d1d9a19023" (UID: "3adc5647-1145-4c44-9ecc-66d1d9a19023"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:54 crc kubenswrapper[4775]: I1002 02:03:54.982430 4775 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3adc5647-1145-4c44-9ecc-66d1d9a19023-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.002383 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-config-data" (OuterVolumeSpecName: "config-data") pod "af16c5f4-fcd0-4450-a8b7-789ff104e477" (UID: "af16c5f4-fcd0-4450-a8b7-789ff104e477"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.018529 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "a5c60336-6e25-4c68-98e6-52801f34578b" (UID: "a5c60336-6e25-4c68-98e6-52801f34578b"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.020140 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "a5c60336-6e25-4c68-98e6-52801f34578b" (UID: "a5c60336-6e25-4c68-98e6-52801f34578b"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.021197 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-config-data" (OuterVolumeSpecName: "config-data") pod "446bd539-6bf4-42b1-ac19-40c8c80b2d45" (UID: "446bd539-6bf4-42b1-ac19-40c8c80b2d45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.043921 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "446bd539-6bf4-42b1-ac19-40c8c80b2d45" (UID: "446bd539-6bf4-42b1-ac19-40c8c80b2d45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.056111 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46453bb2-6af0-4dd4-83ba-b14bf3f96310-config-data" (OuterVolumeSpecName: "config-data") pod "46453bb2-6af0-4dd4-83ba-b14bf3f96310" (UID: "46453bb2-6af0-4dd4-83ba-b14bf3f96310"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.062141 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "446bd539-6bf4-42b1-ac19-40c8c80b2d45" (UID: "446bd539-6bf4-42b1-ac19-40c8c80b2d45"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.065110 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "446bd539-6bf4-42b1-ac19-40c8c80b2d45" (UID: "446bd539-6bf4-42b1-ac19-40c8c80b2d45"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.084488 4775 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.084517 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af16c5f4-fcd0-4450-a8b7-789ff104e477-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.084526 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.084534 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.084542 4775 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5c60336-6e25-4c68-98e6-52801f34578b-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.084552 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.084562 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46453bb2-6af0-4dd4-83ba-b14bf3f96310-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.084570 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/446bd539-6bf4-42b1-ac19-40c8c80b2d45-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.091459 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-79594d69d8-x75fk" podUID="8d6bb525-8c94-461f-84c8-9e466b789706" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": dial tcp 10.217.0.157:9311: connect: connection refused" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.091845 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-79594d69d8-x75fk" podUID="8d6bb525-8c94-461f-84c8-9e466b789706" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.157:9311/healthcheck\": dial tcp 10.217.0.157:9311: connect: connection refused" Oct 02 02:03:55 crc kubenswrapper[4775]: E1002 02:03:55.191511 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:03:55 crc kubenswrapper[4775]: E1002 02:03:55.191778 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:03:55 crc kubenswrapper[4775]: E1002 02:03:55.191903 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:03:55 crc kubenswrapper[4775]: E1002 02:03:55.192140 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:03:55 crc kubenswrapper[4775]: E1002 02:03:55.192160 4775 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovsdb-server" Oct 02 02:03:55 crc kubenswrapper[4775]: E1002 02:03:55.196434 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:03:55 crc kubenswrapper[4775]: E1002 02:03:55.202189 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:03:55 crc kubenswrapper[4775]: E1002 02:03:55.202250 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovs-vswitchd" Oct 02 02:03:55 crc kubenswrapper[4775]: E1002 02:03:55.305352 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2cf05392cd7bff8e546c2fbad0965375d71482567f75af41133d05802bad7f02 is running failed: container process not found" containerID="2cf05392cd7bff8e546c2fbad0965375d71482567f75af41133d05802bad7f02" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 02:03:55 crc kubenswrapper[4775]: E1002 02:03:55.305935 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2cf05392cd7bff8e546c2fbad0965375d71482567f75af41133d05802bad7f02 is running failed: container process not found" containerID="2cf05392cd7bff8e546c2fbad0965375d71482567f75af41133d05802bad7f02" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 02:03:55 crc kubenswrapper[4775]: E1002 02:03:55.306288 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2cf05392cd7bff8e546c2fbad0965375d71482567f75af41133d05802bad7f02 is running failed: container process not found" containerID="2cf05392cd7bff8e546c2fbad0965375d71482567f75af41133d05802bad7f02" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 02:03:55 crc kubenswrapper[4775]: E1002 02:03:55.306522 4775 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2cf05392cd7bff8e546c2fbad0965375d71482567f75af41133d05802bad7f02 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="c74e0a1d-125d-43f7-9ba3-cea70453fc01" containerName="nova-cell0-conductor-conductor" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.413492 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementf75b-account-delete-dl78c" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.424570 4775 scope.go:117] "RemoveContainer" containerID="9d74e2868485fd65eb7dd4d89c912bc745152afc0102e87eaff58d02c1858850" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.436843 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.437348 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.447812 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.462602 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.464968 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.471879 4775 scope.go:117] "RemoveContainer" containerID="979478a6b7d6bf80d7823b67cd948eb55c9e92647ad8b39e7e0c75474a16acc7" Oct 02 02:03:55 crc kubenswrapper[4775]: E1002 02:03:55.473325 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"979478a6b7d6bf80d7823b67cd948eb55c9e92647ad8b39e7e0c75474a16acc7\": container with ID starting with 979478a6b7d6bf80d7823b67cd948eb55c9e92647ad8b39e7e0c75474a16acc7 not found: ID does not exist" containerID="979478a6b7d6bf80d7823b67cd948eb55c9e92647ad8b39e7e0c75474a16acc7" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.473364 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"979478a6b7d6bf80d7823b67cd948eb55c9e92647ad8b39e7e0c75474a16acc7"} err="failed to get container status \"979478a6b7d6bf80d7823b67cd948eb55c9e92647ad8b39e7e0c75474a16acc7\": rpc error: code = NotFound desc = could not find container \"979478a6b7d6bf80d7823b67cd948eb55c9e92647ad8b39e7e0c75474a16acc7\": container with ID starting with 979478a6b7d6bf80d7823b67cd948eb55c9e92647ad8b39e7e0c75474a16acc7 not found: ID does not exist" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.473389 4775 scope.go:117] "RemoveContainer" containerID="9d74e2868485fd65eb7dd4d89c912bc745152afc0102e87eaff58d02c1858850" Oct 02 02:03:55 crc kubenswrapper[4775]: E1002 02:03:55.473676 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d74e2868485fd65eb7dd4d89c912bc745152afc0102e87eaff58d02c1858850\": container with ID starting with 9d74e2868485fd65eb7dd4d89c912bc745152afc0102e87eaff58d02c1858850 not found: ID does not exist" containerID="9d74e2868485fd65eb7dd4d89c912bc745152afc0102e87eaff58d02c1858850" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.473824 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d74e2868485fd65eb7dd4d89c912bc745152afc0102e87eaff58d02c1858850"} err="failed to get container status \"9d74e2868485fd65eb7dd4d89c912bc745152afc0102e87eaff58d02c1858850\": rpc error: code = NotFound desc = could not find container \"9d74e2868485fd65eb7dd4d89c912bc745152afc0102e87eaff58d02c1858850\": container with ID starting with 9d74e2868485fd65eb7dd4d89c912bc745152afc0102e87eaff58d02c1858850 not found: ID does not exist" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.477096 4775 scope.go:117] "RemoveContainer" containerID="1209a9ffa5fb6c6952b1b8effd2ebc2e2fb407ab840357e9b3b1f508504a915a" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.474774 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.476559 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-585fd6468c-8gx5n"] Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.489223 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-585fd6468c-8gx5n"] Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.493307 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87xmq\" (UniqueName: \"kubernetes.io/projected/1e5e6844-45f6-4fd9-ac4f-3df2086164eb-kube-api-access-87xmq\") pod \"1e5e6844-45f6-4fd9-ac4f-3df2086164eb\" (UID: \"1e5e6844-45f6-4fd9-ac4f-3df2086164eb\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.497603 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e5e6844-45f6-4fd9-ac4f-3df2086164eb-kube-api-access-87xmq" (OuterVolumeSpecName: "kube-api-access-87xmq") pod "1e5e6844-45f6-4fd9-ac4f-3df2086164eb" (UID: "1e5e6844-45f6-4fd9-ac4f-3df2086164eb"). InnerVolumeSpecName "kube-api-access-87xmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.513172 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.524562 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.533000 4775 scope.go:117] "RemoveContainer" containerID="42d91714451f6055dca4d72c0b7726b16d6c992e3e11e19af8e604d77f38a639" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.564236 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-546f7b9c49-h6rcs" podUID="2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.164:9696/\": dial tcp 10.217.0.164:9696: connect: connection refused" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.579938 4775 scope.go:117] "RemoveContainer" containerID="9a86b1d47a3009097a1c9b5e0d10b4d2e4d64194427ec3554b609bf3c481d41c" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594655 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-public-tls-certs\") pod \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594697 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-config-data-custom\") pod \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594723 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-combined-ca-bundle\") pod \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594742 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb6zc\" (UniqueName: \"kubernetes.io/projected/56813183-3dfc-430c-aaac-bda27d91340b-kube-api-access-cb6zc\") pod \"56813183-3dfc-430c-aaac-bda27d91340b\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594764 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-internal-tls-certs\") pod \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594796 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5bca47d0-79b7-4992-867b-d5e05e02cdc8-etc-machine-id\") pod \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594825 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f286b47-509d-479d-bad5-0bbec930558a-logs\") pod \"8f286b47-509d-479d-bad5-0bbec930558a\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594845 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-internal-tls-certs\") pod \"56813183-3dfc-430c-aaac-bda27d91340b\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594867 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr8rv\" (UniqueName: \"kubernetes.io/projected/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-api-access-pr8rv\") pod \"24ef8940-eb43-459c-9cfe-854df57cfe74\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594887 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-combined-ca-bundle\") pod \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594909 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-config-data\") pod \"8f286b47-509d-479d-bad5-0bbec930558a\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594925 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-state-metrics-tls-certs\") pod \"24ef8940-eb43-459c-9cfe-854df57cfe74\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594944 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-public-tls-certs\") pod \"56813183-3dfc-430c-aaac-bda27d91340b\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594972 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-internal-tls-certs\") pod \"8f286b47-509d-479d-bad5-0bbec930558a\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.594989 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-combined-ca-bundle\") pod \"24ef8940-eb43-459c-9cfe-854df57cfe74\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595007 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-logs\") pod \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595034 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56813183-3dfc-430c-aaac-bda27d91340b-logs\") pod \"56813183-3dfc-430c-aaac-bda27d91340b\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595054 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-scripts\") pod \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595070 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-scripts\") pod \"56813183-3dfc-430c-aaac-bda27d91340b\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595104 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-internal-tls-certs\") pod \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595125 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-combined-ca-bundle\") pod \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595144 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-combined-ca-bundle\") pod \"56813183-3dfc-430c-aaac-bda27d91340b\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595177 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bca47d0-79b7-4992-867b-d5e05e02cdc8-logs\") pod \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595197 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfkzl\" (UniqueName: \"kubernetes.io/projected/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-kube-api-access-kfkzl\") pod \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595226 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-config-data\") pod \"56813183-3dfc-430c-aaac-bda27d91340b\" (UID: \"56813183-3dfc-430c-aaac-bda27d91340b\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595256 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-logs\") pod \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595334 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-scripts\") pod \"8f286b47-509d-479d-bad5-0bbec930558a\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595350 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sv9w\" (UniqueName: \"kubernetes.io/projected/8f286b47-509d-479d-bad5-0bbec930558a-kube-api-access-5sv9w\") pod \"8f286b47-509d-479d-bad5-0bbec930558a\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595369 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-config-data\") pod \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595390 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-config-data\") pod \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595405 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f286b47-509d-479d-bad5-0bbec930558a-httpd-run\") pod \"8f286b47-509d-479d-bad5-0bbec930558a\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595421 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"8f286b47-509d-479d-bad5-0bbec930558a\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595437 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-httpd-run\") pod \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595456 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-scripts\") pod \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595478 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-public-tls-certs\") pod \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595497 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-config-data\") pod \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595519 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmdwd\" (UniqueName: \"kubernetes.io/projected/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-kube-api-access-dmdwd\") pod \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\" (UID: \"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595535 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595553 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-combined-ca-bundle\") pod \"8f286b47-509d-479d-bad5-0bbec930558a\" (UID: \"8f286b47-509d-479d-bad5-0bbec930558a\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595570 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7cl5\" (UniqueName: \"kubernetes.io/projected/5bca47d0-79b7-4992-867b-d5e05e02cdc8-kube-api-access-x7cl5\") pod \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\" (UID: \"5bca47d0-79b7-4992-867b-d5e05e02cdc8\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595586 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-state-metrics-tls-config\") pod \"24ef8940-eb43-459c-9cfe-854df57cfe74\" (UID: \"24ef8940-eb43-459c-9cfe-854df57cfe74\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.595608 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-public-tls-certs\") pod \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\" (UID: \"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51\") " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.596077 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87xmq\" (UniqueName: \"kubernetes.io/projected/1e5e6844-45f6-4fd9-ac4f-3df2086164eb-kube-api-access-87xmq\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.598569 4775 scope.go:117] "RemoveContainer" containerID="9d2ff242a1e29ad8407f6c796b021ec08d4a487ead665a7032ac8ddbf2177bbc" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.600854 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5bca47d0-79b7-4992-867b-d5e05e02cdc8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5bca47d0-79b7-4992-867b-d5e05e02cdc8" (UID: "5bca47d0-79b7-4992-867b-d5e05e02cdc8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.610002 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5bca47d0-79b7-4992-867b-d5e05e02cdc8" (UID: "5bca47d0-79b7-4992-867b-d5e05e02cdc8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.612765 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56813183-3dfc-430c-aaac-bda27d91340b-kube-api-access-cb6zc" (OuterVolumeSpecName: "kube-api-access-cb6zc") pod "56813183-3dfc-430c-aaac-bda27d91340b" (UID: "56813183-3dfc-430c-aaac-bda27d91340b"). InnerVolumeSpecName "kube-api-access-cb6zc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.612911 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f286b47-509d-479d-bad5-0bbec930558a-logs" (OuterVolumeSpecName: "logs") pod "8f286b47-509d-479d-bad5-0bbec930558a" (UID: "8f286b47-509d-479d-bad5-0bbec930558a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.614835 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f286b47-509d-479d-bad5-0bbec930558a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8f286b47-509d-479d-bad5-0bbec930558a" (UID: "8f286b47-509d-479d-bad5-0bbec930558a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.615994 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" (UID: "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.617523 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56813183-3dfc-430c-aaac-bda27d91340b-logs" (OuterVolumeSpecName: "logs") pod "56813183-3dfc-430c-aaac-bda27d91340b" (UID: "56813183-3dfc-430c-aaac-bda27d91340b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.617918 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5bca47d0-79b7-4992-867b-d5e05e02cdc8-logs" (OuterVolumeSpecName: "logs") pod "5bca47d0-79b7-4992-867b-d5e05e02cdc8" (UID: "5bca47d0-79b7-4992-867b-d5e05e02cdc8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.618146 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-logs" (OuterVolumeSpecName: "logs") pod "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" (UID: "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.623706 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-logs" (OuterVolumeSpecName: "logs") pod "8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" (UID: "8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.629905 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"24ef8940-eb43-459c-9cfe-854df57cfe74","Type":"ContainerDied","Data":"4588a06ce5aa5eae7ebbeddab0850d8aaef02484e59ea3376959ccc6a6facdbe"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.630012 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.641170 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-777987c7d4-d5vzz" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.641176 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-777987c7d4-d5vzz" event={"ID":"56813183-3dfc-430c-aaac-bda27d91340b","Type":"ContainerDied","Data":"8baa28cbfb413992b0822f7aa986dd152be25ca69358d92c31e610054a7e6493"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.642800 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bca47d0-79b7-4992-867b-d5e05e02cdc8-kube-api-access-x7cl5" (OuterVolumeSpecName: "kube-api-access-x7cl5") pod "5bca47d0-79b7-4992-867b-d5e05e02cdc8" (UID: "5bca47d0-79b7-4992-867b-d5e05e02cdc8"). InnerVolumeSpecName "kube-api-access-x7cl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.647926 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" (UID: "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.650096 4775 generic.go:334] "Generic (PLEG): container finished" podID="5bca47d0-79b7-4992-867b-d5e05e02cdc8" containerID="e38c74d33b3ad7b8f747e53560c787d296c45bf29b3227467260f620932b62a9" exitCode=0 Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.650153 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5bca47d0-79b7-4992-867b-d5e05e02cdc8","Type":"ContainerDied","Data":"e38c74d33b3ad7b8f747e53560c787d296c45bf29b3227467260f620932b62a9"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.650176 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"5bca47d0-79b7-4992-867b-d5e05e02cdc8","Type":"ContainerDied","Data":"0bc9445b2e060fd2d8cafb7bc36635d4aca57867fb9172aed3eebad1b3cd6ef2"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.650222 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.651401 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-scripts" (OuterVolumeSpecName: "scripts") pod "56813183-3dfc-430c-aaac-bda27d91340b" (UID: "56813183-3dfc-430c-aaac-bda27d91340b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.651596 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-scripts" (OuterVolumeSpecName: "scripts") pod "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" (UID: "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.653871 4775 generic.go:334] "Generic (PLEG): container finished" podID="8f286b47-509d-479d-bad5-0bbec930558a" containerID="db8059deef67f936d22e07a37161aed4eda44bdc81845a2918822f9946fd9c00" exitCode=0 Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.653925 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.654054 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f286b47-509d-479d-bad5-0bbec930558a","Type":"ContainerDied","Data":"db8059deef67f936d22e07a37161aed4eda44bdc81845a2918822f9946fd9c00"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.654082 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8f286b47-509d-479d-bad5-0bbec930558a","Type":"ContainerDied","Data":"a8f89d176e72a63eb5c4007babeb5a609d0afffddf3f01497b66de43cffb3e2c"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.654756 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "8f286b47-509d-479d-bad5-0bbec930558a" (UID: "8f286b47-509d-479d-bad5-0bbec930558a"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.655857 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-api-access-pr8rv" (OuterVolumeSpecName: "kube-api-access-pr8rv") pod "24ef8940-eb43-459c-9cfe-854df57cfe74" (UID: "24ef8940-eb43-459c-9cfe-854df57cfe74"). InnerVolumeSpecName "kube-api-access-pr8rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.657101 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-kube-api-access-dmdwd" (OuterVolumeSpecName: "kube-api-access-dmdwd") pod "8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" (UID: "8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0"). InnerVolumeSpecName "kube-api-access-dmdwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.658768 4775 generic.go:334] "Generic (PLEG): container finished" podID="8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" containerID="9d068bf0ca5c469b57dcad75c4e6ebff62f9badd694815d6349f36cd138a20a8" exitCode=0 Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.658847 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-scripts" (OuterVolumeSpecName: "scripts") pod "5bca47d0-79b7-4992-867b-d5e05e02cdc8" (UID: "5bca47d0-79b7-4992-867b-d5e05e02cdc8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.658865 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.658854 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0","Type":"ContainerDied","Data":"9d068bf0ca5c469b57dcad75c4e6ebff62f9badd694815d6349f36cd138a20a8"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.658916 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0","Type":"ContainerDied","Data":"bb2c755284b81a31e41cade55e5b66ce01b89e599614cc72c51530d3b154f91e"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.659693 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f286b47-509d-479d-bad5-0bbec930558a-kube-api-access-5sv9w" (OuterVolumeSpecName: "kube-api-access-5sv9w") pod "8f286b47-509d-479d-bad5-0bbec930558a" (UID: "8f286b47-509d-479d-bad5-0bbec930558a"). InnerVolumeSpecName "kube-api-access-5sv9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.659809 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-scripts" (OuterVolumeSpecName: "scripts") pod "8f286b47-509d-479d-bad5-0bbec930558a" (UID: "8f286b47-509d-479d-bad5-0bbec930558a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.663073 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-kube-api-access-kfkzl" (OuterVolumeSpecName: "kube-api-access-kfkzl") pod "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" (UID: "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51"). InnerVolumeSpecName "kube-api-access-kfkzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.664156 4775 generic.go:334] "Generic (PLEG): container finished" podID="5e322679-b95d-4328-b03b-0be7f0e82ac3" containerID="33e28213f9c089646224401e310e6c70d1efdff65261c89fe379e544e09ae98f" exitCode=0 Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.664212 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e322679-b95d-4328-b03b-0be7f0e82ac3","Type":"ContainerDied","Data":"33e28213f9c089646224401e310e6c70d1efdff65261c89fe379e544e09ae98f"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.669353 4775 generic.go:334] "Generic (PLEG): container finished" podID="8d6bb525-8c94-461f-84c8-9e466b789706" containerID="ddc807916bf2152781e0b263b81d2ba2f8ff95a514f51a7d7918774142370f47" exitCode=0 Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.669375 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79594d69d8-x75fk" event={"ID":"8d6bb525-8c94-461f-84c8-9e466b789706","Type":"ContainerDied","Data":"ddc807916bf2152781e0b263b81d2ba2f8ff95a514f51a7d7918774142370f47"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.669479 4775 scope.go:117] "RemoveContainer" containerID="7faaf4aaa749333ce8970b0b3bfce0b23a77558417d2d73450a187551b1a850f" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.671836 4775 generic.go:334] "Generic (PLEG): container finished" podID="8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" containerID="63252743cedda5425e3430e58790235c337b97aaa0683e299068d0092eca5e7f" exitCode=0 Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.671876 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51","Type":"ContainerDied","Data":"63252743cedda5425e3430e58790235c337b97aaa0683e299068d0092eca5e7f"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.671893 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51","Type":"ContainerDied","Data":"7a38314d1ae75675da87ea02ecbe9d1fac0745fb118de7950280d4535aa753e7"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.671948 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.682640 4775 generic.go:334] "Generic (PLEG): container finished" podID="c74e0a1d-125d-43f7-9ba3-cea70453fc01" containerID="2cf05392cd7bff8e546c2fbad0965375d71482567f75af41133d05802bad7f02" exitCode=0 Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.682711 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c74e0a1d-125d-43f7-9ba3-cea70453fc01","Type":"ContainerDied","Data":"2cf05392cd7bff8e546c2fbad0965375d71482567f75af41133d05802bad7f02"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.684731 4775 generic.go:334] "Generic (PLEG): container finished" podID="406dd24f-d800-44be-b33d-019c532c2feb" containerID="062b09ca95edb1ddddcb2de55757014ea328d800d380273e8fb5c4e3bb488c11" exitCode=0 Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.684771 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"406dd24f-d800-44be-b33d-019c532c2feb","Type":"ContainerDied","Data":"062b09ca95edb1ddddcb2de55757014ea328d800d380273e8fb5c4e3bb488c11"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.690497 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" (UID: "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698224 4775 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5bca47d0-79b7-4992-867b-d5e05e02cdc8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698246 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f286b47-509d-479d-bad5-0bbec930558a-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698256 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr8rv\" (UniqueName: \"kubernetes.io/projected/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-api-access-pr8rv\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698266 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698274 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56813183-3dfc-430c-aaac-bda27d91340b-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698283 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698291 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698298 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5bca47d0-79b7-4992-867b-d5e05e02cdc8-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698306 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfkzl\" (UniqueName: \"kubernetes.io/projected/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-kube-api-access-kfkzl\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698323 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698332 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698340 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sv9w\" (UniqueName: \"kubernetes.io/projected/8f286b47-509d-479d-bad5-0bbec930558a-kube-api-access-5sv9w\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698349 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8f286b47-509d-479d-bad5-0bbec930558a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698373 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698381 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698390 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698398 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmdwd\" (UniqueName: \"kubernetes.io/projected/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-kube-api-access-dmdwd\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698411 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698420 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7cl5\" (UniqueName: \"kubernetes.io/projected/5bca47d0-79b7-4992-867b-d5e05e02cdc8-kube-api-access-x7cl5\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698429 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698439 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.698447 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb6zc\" (UniqueName: \"kubernetes.io/projected/56813183-3dfc-430c-aaac-bda27d91340b-kube-api-access-cb6zc\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.703939 4775 generic.go:334] "Generic (PLEG): container finished" podID="d2d45342-c4de-48db-8f6b-23f26103497b" containerID="4feb9154e0329bc25829a25b496b1b2573ea599dcffdc17eb82bbc2b54f58aa4" exitCode=0 Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.703989 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2d45342-c4de-48db-8f6b-23f26103497b","Type":"ContainerDied","Data":"4feb9154e0329bc25829a25b496b1b2573ea599dcffdc17eb82bbc2b54f58aa4"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.704027 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2d45342-c4de-48db-8f6b-23f26103497b","Type":"ContainerDied","Data":"8df9de2a0e96e291b49efa835c2476332824829ee66a90f08c5fa720d9e412f8"} Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.704002 4775 generic.go:334] "Generic (PLEG): container finished" podID="d2d45342-c4de-48db-8f6b-23f26103497b" containerID="8df9de2a0e96e291b49efa835c2476332824829ee66a90f08c5fa720d9e412f8" exitCode=0 Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.704100 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placementf75b-account-delete-dl78c" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.704155 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.704204 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder5592-account-delete-mhfmd" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.704386 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-d79dd6fcb-lzscb" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.704428 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.704431 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanb963-account-delete-z5bdz" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.704457 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance95b1-account-delete-4jqdz" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.741030 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.742601 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.746381 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-config-data" (OuterVolumeSpecName: "config-data") pod "8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" (UID: "8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.749786 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "24ef8940-eb43-459c-9cfe-854df57cfe74" (UID: "24ef8940-eb43-459c-9cfe-854df57cfe74"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.758235 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "24ef8940-eb43-459c-9cfe-854df57cfe74" (UID: "24ef8940-eb43-459c-9cfe-854df57cfe74"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.758722 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5bca47d0-79b7-4992-867b-d5e05e02cdc8" (UID: "5bca47d0-79b7-4992-867b-d5e05e02cdc8"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.762994 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5bca47d0-79b7-4992-867b-d5e05e02cdc8" (UID: "5bca47d0-79b7-4992-867b-d5e05e02cdc8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.776566 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1364df9e-151f-40d0-9a5a-606a54021ef7" path="/var/lib/kubelet/pods/1364df9e-151f-40d0-9a5a-606a54021ef7/volumes" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.777327 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="163a6d17-be4c-47fc-8835-e218e8a803f1" path="/var/lib/kubelet/pods/163a6d17-be4c-47fc-8835-e218e8a803f1/volumes" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.777978 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="272c8e33-553c-4932-951d-e1944cffc9e7" path="/var/lib/kubelet/pods/272c8e33-553c-4932-951d-e1944cffc9e7/volumes" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.778938 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32229132-5589-46e1-a172-228eaa41bfce" path="/var/lib/kubelet/pods/32229132-5589-46e1-a172-228eaa41bfce/volumes" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.779812 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3adc5647-1145-4c44-9ecc-66d1d9a19023" path="/var/lib/kubelet/pods/3adc5647-1145-4c44-9ecc-66d1d9a19023/volumes" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.780446 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="446bd539-6bf4-42b1-ac19-40c8c80b2d45" path="/var/lib/kubelet/pods/446bd539-6bf4-42b1-ac19-40c8c80b2d45/volumes" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.781456 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d9d080a-bb9e-4394-947f-5fb5df9a52ec" path="/var/lib/kubelet/pods/4d9d080a-bb9e-4394-947f-5fb5df9a52ec/volumes" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.781905 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a95c042-cb96-42be-8192-2af9ad63d43d" path="/var/lib/kubelet/pods/8a95c042-cb96-42be-8192-2af9ad63d43d/volumes" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.782738 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97a109cd-fde0-409f-a0cc-e2e3c78acb47" path="/var/lib/kubelet/pods/97a109cd-fde0-409f-a0cc-e2e3c78acb47/volumes" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.783747 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" path="/var/lib/kubelet/pods/c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92/volumes" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.784414 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9813c97-677a-42c2-b526-576a4c5f2968" path="/var/lib/kubelet/pods/e9813c97-677a-42c2-b526-576a4c5f2968/volumes" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.785156 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7685ff7-607b-416c-9382-2c00edecaa54" path="/var/lib/kubelet/pods/f7685ff7-607b-416c-9382-2c00edecaa54/volumes" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.799474 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" (UID: "8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.800197 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.800296 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.800388 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.800453 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.800516 4775 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.800580 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.800929 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.801048 4775 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.808364 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56813183-3dfc-430c-aaac-bda27d91340b" (UID: "56813183-3dfc-430c-aaac-bda27d91340b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.874058 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-config-data" (OuterVolumeSpecName: "config-data") pod "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" (UID: "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.874062 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f286b47-509d-479d-bad5-0bbec930558a" (UID: "8f286b47-509d-479d-bad5-0bbec930558a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.893346 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24ef8940-eb43-459c-9cfe-854df57cfe74" (UID: "24ef8940-eb43-459c-9cfe-854df57cfe74"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.903918 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24ef8940-eb43-459c-9cfe-854df57cfe74-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.904203 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.904212 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.904223 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.907138 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" (UID: "8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.943024 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-config-data" (OuterVolumeSpecName: "config-data") pod "8f286b47-509d-479d-bad5-0bbec930558a" (UID: "8f286b47-509d-479d-bad5-0bbec930558a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.958736 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" (UID: "8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.981088 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-config-data" (OuterVolumeSpecName: "config-data") pod "56813183-3dfc-430c-aaac-bda27d91340b" (UID: "56813183-3dfc-430c-aaac-bda27d91340b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:55 crc kubenswrapper[4775]: I1002 02:03:55.990820 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5bca47d0-79b7-4992-867b-d5e05e02cdc8" (UID: "5bca47d0-79b7-4992-867b-d5e05e02cdc8"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.006029 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.006056 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.006065 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.006073 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.006081 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.016075 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "56813183-3dfc-430c-aaac-bda27d91340b" (UID: "56813183-3dfc-430c-aaac-bda27d91340b"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: E1002 02:03:56.017489 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="39f92c334c8a479d69a76f6ee477dc9e38b2f2fb13b731fddbffda7d7dac3511" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 02 02:03:56 crc kubenswrapper[4775]: E1002 02:03:56.021027 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="39f92c334c8a479d69a76f6ee477dc9e38b2f2fb13b731fddbffda7d7dac3511" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 02 02:03:56 crc kubenswrapper[4775]: E1002 02:03:56.026490 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="39f92c334c8a479d69a76f6ee477dc9e38b2f2fb13b731fddbffda7d7dac3511" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Oct 02 02:03:56 crc kubenswrapper[4775]: E1002 02:03:56.026552 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="7da01b78-a318-448e-8a4c-7ad3d9e833c0" containerName="ovn-northd" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.030838 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" (UID: "8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.045015 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-config-data" (OuterVolumeSpecName: "config-data") pod "5bca47d0-79b7-4992-867b-d5e05e02cdc8" (UID: "5bca47d0-79b7-4992-867b-d5e05e02cdc8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.048147 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8f286b47-509d-479d-bad5-0bbec930558a" (UID: "8f286b47-509d-479d-bad5-0bbec930558a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.065881 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "56813183-3dfc-430c-aaac-bda27d91340b" (UID: "56813183-3dfc-430c-aaac-bda27d91340b"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.113252 4775 scope.go:117] "RemoveContainer" containerID="d46af6901507bf3d2c6f5cd7933a7df11f703991822ecd42a838e772bb645206" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.113508 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.114322 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.114342 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/56813183-3dfc-430c-aaac-bda27d91340b-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.114351 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8f286b47-509d-479d-bad5-0bbec930558a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.115589 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bca47d0-79b7-4992-867b-d5e05e02cdc8-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.115603 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.216319 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmg46\" (UniqueName: \"kubernetes.io/projected/8d6bb525-8c94-461f-84c8-9e466b789706-kube-api-access-qmg46\") pod \"8d6bb525-8c94-461f-84c8-9e466b789706\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.216608 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-combined-ca-bundle\") pod \"8d6bb525-8c94-461f-84c8-9e466b789706\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.216691 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-config-data\") pod \"8d6bb525-8c94-461f-84c8-9e466b789706\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.216760 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-public-tls-certs\") pod \"8d6bb525-8c94-461f-84c8-9e466b789706\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.217114 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-internal-tls-certs\") pod \"8d6bb525-8c94-461f-84c8-9e466b789706\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.217208 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d6bb525-8c94-461f-84c8-9e466b789706-logs\") pod \"8d6bb525-8c94-461f-84c8-9e466b789706\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.217240 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-config-data-custom\") pod \"8d6bb525-8c94-461f-84c8-9e466b789706\" (UID: \"8d6bb525-8c94-461f-84c8-9e466b789706\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.218489 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d6bb525-8c94-461f-84c8-9e466b789706-logs" (OuterVolumeSpecName: "logs") pod "8d6bb525-8c94-461f-84c8-9e466b789706" (UID: "8d6bb525-8c94-461f-84c8-9e466b789706"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.221143 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8d6bb525-8c94-461f-84c8-9e466b789706" (UID: "8d6bb525-8c94-461f-84c8-9e466b789706"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.224596 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d6bb525-8c94-461f-84c8-9e466b789706-kube-api-access-qmg46" (OuterVolumeSpecName: "kube-api-access-qmg46") pod "8d6bb525-8c94-461f-84c8-9e466b789706" (UID: "8d6bb525-8c94-461f-84c8-9e466b789706"). InnerVolumeSpecName "kube-api-access-qmg46". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.280186 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8d6bb525-8c94-461f-84c8-9e466b789706" (UID: "8d6bb525-8c94-461f-84c8-9e466b789706"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.284576 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-config-data" (OuterVolumeSpecName: "config-data") pod "8d6bb525-8c94-461f-84c8-9e466b789706" (UID: "8d6bb525-8c94-461f-84c8-9e466b789706"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.285054 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8d6bb525-8c94-461f-84c8-9e466b789706" (UID: "8d6bb525-8c94-461f-84c8-9e466b789706"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.299052 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8d6bb525-8c94-461f-84c8-9e466b789706" (UID: "8d6bb525-8c94-461f-84c8-9e466b789706"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.318896 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.318917 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmg46\" (UniqueName: \"kubernetes.io/projected/8d6bb525-8c94-461f-84c8-9e466b789706-kube-api-access-qmg46\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.318929 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.318938 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.318963 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.318971 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8d6bb525-8c94-461f-84c8-9e466b789706-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.318979 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d6bb525-8c94-461f-84c8-9e466b789706-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.347476 4775 scope.go:117] "RemoveContainer" containerID="709d8a7bd6eb1d47bf0d40f08c279c162aa59c94b3583a614ae61982e1165eea" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.528522 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.554568 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.560835 4775 scope.go:117] "RemoveContainer" containerID="e38c74d33b3ad7b8f747e53560c787d296c45bf29b3227467260f620932b62a9" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.577636 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.607730 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.617476 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.618140 4775 scope.go:117] "RemoveContainer" containerID="04d6bdf03305dd966a2ca95b6a3220b41a5add02a16becda008422ee10720237" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.621905 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.624099 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-nova-metadata-tls-certs\") pod \"5e322679-b95d-4328-b03b-0be7f0e82ac3\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.627262 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzh7k\" (UniqueName: \"kubernetes.io/projected/5e322679-b95d-4328-b03b-0be7f0e82ac3-kube-api-access-hzh7k\") pod \"5e322679-b95d-4328-b03b-0be7f0e82ac3\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.627508 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-config-data\") pod \"5e322679-b95d-4328-b03b-0be7f0e82ac3\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.627612 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e322679-b95d-4328-b03b-0be7f0e82ac3-logs\") pod \"5e322679-b95d-4328-b03b-0be7f0e82ac3\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.627722 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-combined-ca-bundle\") pod \"5e322679-b95d-4328-b03b-0be7f0e82ac3\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.629625 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placementf75b-account-delete-dl78c"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.631242 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e322679-b95d-4328-b03b-0be7f0e82ac3-logs" (OuterVolumeSpecName: "logs") pod "5e322679-b95d-4328-b03b-0be7f0e82ac3" (UID: "5e322679-b95d-4328-b03b-0be7f0e82ac3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.633185 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placementf75b-account-delete-dl78c"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.640681 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicanb963-account-delete-z5bdz"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.643643 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbicanb963-account-delete-z5bdz"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.647991 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e322679-b95d-4328-b03b-0be7f0e82ac3-kube-api-access-hzh7k" (OuterVolumeSpecName: "kube-api-access-hzh7k") pod "5e322679-b95d-4328-b03b-0be7f0e82ac3" (UID: "5e322679-b95d-4328-b03b-0be7f0e82ac3"). InnerVolumeSpecName "kube-api-access-hzh7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.650030 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-d79dd6fcb-lzscb"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.656714 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-d79dd6fcb-lzscb"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.666147 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance95b1-account-delete-4jqdz"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.666854 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e322679-b95d-4328-b03b-0be7f0e82ac3" (UID: "5e322679-b95d-4328-b03b-0be7f0e82ac3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.734307 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "5e322679-b95d-4328-b03b-0be7f0e82ac3" (UID: "5e322679-b95d-4328-b03b-0be7f0e82ac3"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.734560 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jdbs4" podUID="6efbe4b3-953d-4927-95c6-9b92708f51eb" containerName="ovn-controller" probeResult="failure" output="command timed out" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.734946 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb4p8\" (UniqueName: \"kubernetes.io/projected/406dd24f-d800-44be-b33d-019c532c2feb-kube-api-access-zb4p8\") pod \"406dd24f-d800-44be-b33d-019c532c2feb\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.735050 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-ceilometer-tls-certs\") pod \"d2d45342-c4de-48db-8f6b-23f26103497b\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.735113 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-config-data\") pod \"d2d45342-c4de-48db-8f6b-23f26103497b\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.735175 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/406dd24f-d800-44be-b33d-019c532c2feb-kolla-config\") pod \"406dd24f-d800-44be-b33d-019c532c2feb\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.735257 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-nova-metadata-tls-certs\") pod \"5e322679-b95d-4328-b03b-0be7f0e82ac3\" (UID: \"5e322679-b95d-4328-b03b-0be7f0e82ac3\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.735350 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-scripts\") pod \"d2d45342-c4de-48db-8f6b-23f26103497b\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.735436 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rk2vw\" (UniqueName: \"kubernetes.io/projected/d2d45342-c4de-48db-8f6b-23f26103497b-kube-api-access-rk2vw\") pod \"d2d45342-c4de-48db-8f6b-23f26103497b\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.735503 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406dd24f-d800-44be-b33d-019c532c2feb-combined-ca-bundle\") pod \"406dd24f-d800-44be-b33d-019c532c2feb\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.735601 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74e0a1d-125d-43f7-9ba3-cea70453fc01-combined-ca-bundle\") pod \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\" (UID: \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.735751 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvwrj\" (UniqueName: \"kubernetes.io/projected/c74e0a1d-125d-43f7-9ba3-cea70453fc01-kube-api-access-lvwrj\") pod \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\" (UID: \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.735851 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-sg-core-conf-yaml\") pod \"d2d45342-c4de-48db-8f6b-23f26103497b\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.735920 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2d45342-c4de-48db-8f6b-23f26103497b-log-httpd\") pod \"d2d45342-c4de-48db-8f6b-23f26103497b\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.736019 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-combined-ca-bundle\") pod \"d2d45342-c4de-48db-8f6b-23f26103497b\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.736096 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/406dd24f-d800-44be-b33d-019c532c2feb-config-data\") pod \"406dd24f-d800-44be-b33d-019c532c2feb\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.736163 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2d45342-c4de-48db-8f6b-23f26103497b-run-httpd\") pod \"d2d45342-c4de-48db-8f6b-23f26103497b\" (UID: \"d2d45342-c4de-48db-8f6b-23f26103497b\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.736252 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/406dd24f-d800-44be-b33d-019c532c2feb-memcached-tls-certs\") pod \"406dd24f-d800-44be-b33d-019c532c2feb\" (UID: \"406dd24f-d800-44be-b33d-019c532c2feb\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.736364 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74e0a1d-125d-43f7-9ba3-cea70453fc01-config-data\") pod \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\" (UID: \"c74e0a1d-125d-43f7-9ba3-cea70453fc01\") " Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.736753 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5e322679-b95d-4328-b03b-0be7f0e82ac3-logs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.738309 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.738433 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzh7k\" (UniqueName: \"kubernetes.io/projected/5e322679-b95d-4328-b03b-0be7f0e82ac3-kube-api-access-hzh7k\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: W1002 02:03:56.736219 4775 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/5e322679-b95d-4328-b03b-0be7f0e82ac3/volumes/kubernetes.io~secret/nova-metadata-tls-certs Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.741199 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "5e322679-b95d-4328-b03b-0be7f0e82ac3" (UID: "5e322679-b95d-4328-b03b-0be7f0e82ac3"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.738166 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/406dd24f-d800-44be-b33d-019c532c2feb-kube-api-access-zb4p8" (OuterVolumeSpecName: "kube-api-access-zb4p8") pod "406dd24f-d800-44be-b33d-019c532c2feb" (UID: "406dd24f-d800-44be-b33d-019c532c2feb"). InnerVolumeSpecName "kube-api-access-zb4p8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.739336 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/406dd24f-d800-44be-b33d-019c532c2feb-config-data" (OuterVolumeSpecName: "config-data") pod "406dd24f-d800-44be-b33d-019c532c2feb" (UID: "406dd24f-d800-44be-b33d-019c532c2feb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.739809 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2d45342-c4de-48db-8f6b-23f26103497b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d2d45342-c4de-48db-8f6b-23f26103497b" (UID: "d2d45342-c4de-48db-8f6b-23f26103497b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.741479 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2d45342-c4de-48db-8f6b-23f26103497b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d2d45342-c4de-48db-8f6b-23f26103497b" (UID: "d2d45342-c4de-48db-8f6b-23f26103497b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.741798 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/406dd24f-d800-44be-b33d-019c532c2feb-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "406dd24f-d800-44be-b33d-019c532c2feb" (UID: "406dd24f-d800-44be-b33d-019c532c2feb"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.757145 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance95b1-account-delete-4jqdz"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.767616 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"406dd24f-d800-44be-b33d-019c532c2feb","Type":"ContainerDied","Data":"fe62491b1f99d977bbcdd66d6c9f7ce148e13044528e146f34627800b201ae60"} Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.767810 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.770015 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.770145 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5e322679-b95d-4328-b03b-0be7f0e82ac3","Type":"ContainerDied","Data":"0b25a17cee3661f96b7c2e3eb90aced2226298c80f4b5f4c4e1f0f3062398aee"} Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.784122 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c74e0a1d-125d-43f7-9ba3-cea70453fc01-kube-api-access-lvwrj" (OuterVolumeSpecName: "kube-api-access-lvwrj") pod "c74e0a1d-125d-43f7-9ba3-cea70453fc01" (UID: "c74e0a1d-125d-43f7-9ba3-cea70453fc01"). InnerVolumeSpecName "kube-api-access-lvwrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.784156 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-scripts" (OuterVolumeSpecName: "scripts") pod "d2d45342-c4de-48db-8f6b-23f26103497b" (UID: "d2d45342-c4de-48db-8f6b-23f26103497b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.784715 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2d45342-c4de-48db-8f6b-23f26103497b-kube-api-access-rk2vw" (OuterVolumeSpecName: "kube-api-access-rk2vw") pod "d2d45342-c4de-48db-8f6b-23f26103497b" (UID: "d2d45342-c4de-48db-8f6b-23f26103497b"). InnerVolumeSpecName "kube-api-access-rk2vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.792595 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79594d69d8-x75fk" event={"ID":"8d6bb525-8c94-461f-84c8-9e466b789706","Type":"ContainerDied","Data":"107e97586776a360c4d800e0a787832481a9cc510a49a9cdb3db505fa31fdb98"} Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.793763 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79594d69d8-x75fk" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.795686 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7da01b78-a318-448e-8a4c-7ad3d9e833c0/ovn-northd/0.log" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.795729 4775 generic.go:334] "Generic (PLEG): container finished" podID="7da01b78-a318-448e-8a4c-7ad3d9e833c0" containerID="39f92c334c8a479d69a76f6ee477dc9e38b2f2fb13b731fddbffda7d7dac3511" exitCode=139 Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.795688 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.795807 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7da01b78-a318-448e-8a4c-7ad3d9e833c0","Type":"ContainerDied","Data":"39f92c334c8a479d69a76f6ee477dc9e38b2f2fb13b731fddbffda7d7dac3511"} Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.802611 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c74e0a1d-125d-43f7-9ba3-cea70453fc01","Type":"ContainerDied","Data":"436ed97e2972853f43721bb36b94dc770df3bfe0aed42fb78e37eb4ddd990d75"} Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.802691 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.812286 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.813271 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d2d45342-c4de-48db-8f6b-23f26103497b","Type":"ContainerDied","Data":"251614ec9eafc9d0dc22c0b2e4a21f7e76e5fea8e660ef31288764935585dffd"} Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.813365 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.821477 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.826991 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.831634 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder5592-account-delete-mhfmd"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.835170 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder5592-account-delete-mhfmd"] Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.839706 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/406dd24f-d800-44be-b33d-019c532c2feb-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.839731 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2d45342-c4de-48db-8f6b-23f26103497b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.839740 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb4p8\" (UniqueName: \"kubernetes.io/projected/406dd24f-d800-44be-b33d-019c532c2feb-kube-api-access-zb4p8\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.839750 4775 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/406dd24f-d800-44be-b33d-019c532c2feb-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.839759 4775 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.839766 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.839776 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rk2vw\" (UniqueName: \"kubernetes.io/projected/d2d45342-c4de-48db-8f6b-23f26103497b-kube-api-access-rk2vw\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.839784 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvwrj\" (UniqueName: \"kubernetes.io/projected/c74e0a1d-125d-43f7-9ba3-cea70453fc01-kube-api-access-lvwrj\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.839791 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d2d45342-c4de-48db-8f6b-23f26103497b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.865794 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jdbs4" podUID="6efbe4b3-953d-4927-95c6-9b92708f51eb" containerName="ovn-controller" probeResult="failure" output=< Oct 02 02:03:56 crc kubenswrapper[4775]: ERROR - Failed to get connection status from ovn-controller, ovn-appctl exit status: 0 Oct 02 02:03:56 crc kubenswrapper[4775]: > Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.901379 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d2d45342-c4de-48db-8f6b-23f26103497b" (UID: "d2d45342-c4de-48db-8f6b-23f26103497b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.903984 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74e0a1d-125d-43f7-9ba3-cea70453fc01-config-data" (OuterVolumeSpecName: "config-data") pod "c74e0a1d-125d-43f7-9ba3-cea70453fc01" (UID: "c74e0a1d-125d-43f7-9ba3-cea70453fc01"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.906788 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d2d45342-c4de-48db-8f6b-23f26103497b" (UID: "d2d45342-c4de-48db-8f6b-23f26103497b"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.921138 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c74e0a1d-125d-43f7-9ba3-cea70453fc01-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c74e0a1d-125d-43f7-9ba3-cea70453fc01" (UID: "c74e0a1d-125d-43f7-9ba3-cea70453fc01"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.921839 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/406dd24f-d800-44be-b33d-019c532c2feb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "406dd24f-d800-44be-b33d-019c532c2feb" (UID: "406dd24f-d800-44be-b33d-019c532c2feb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.926633 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-config-data" (OuterVolumeSpecName: "config-data") pod "5e322679-b95d-4328-b03b-0be7f0e82ac3" (UID: "5e322679-b95d-4328-b03b-0be7f0e82ac3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.939254 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2d45342-c4de-48db-8f6b-23f26103497b" (UID: "d2d45342-c4de-48db-8f6b-23f26103497b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.943096 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/406dd24f-d800-44be-b33d-019c532c2feb-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "406dd24f-d800-44be-b33d-019c532c2feb" (UID: "406dd24f-d800-44be-b33d-019c532c2feb"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.943273 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.943297 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.943306 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e322679-b95d-4328-b03b-0be7f0e82ac3-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.943315 4775 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/406dd24f-d800-44be-b33d-019c532c2feb-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.943324 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c74e0a1d-125d-43f7-9ba3-cea70453fc01-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.943332 4775 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.943340 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406dd24f-d800-44be-b33d-019c532c2feb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.943348 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c74e0a1d-125d-43f7-9ba3-cea70453fc01-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:56 crc kubenswrapper[4775]: E1002 02:03:56.943895 4775 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Oct 02 02:03:56 crc kubenswrapper[4775]: E1002 02:03:56.943938 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data podName:6638baa9-b367-40d1-8111-673dca6434ad nodeName:}" failed. No retries permitted until 2025-10-02 02:04:04.943925291 +0000 UTC m=+1382.110669331 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data") pod "rabbitmq-server-0" (UID: "6638baa9-b367-40d1-8111-673dca6434ad") : configmap "rabbitmq-config-data" not found Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.944338 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-config-data" (OuterVolumeSpecName: "config-data") pod "d2d45342-c4de-48db-8f6b-23f26103497b" (UID: "d2d45342-c4de-48db-8f6b-23f26103497b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.994912 4775 scope.go:117] "RemoveContainer" containerID="e38c74d33b3ad7b8f747e53560c787d296c45bf29b3227467260f620932b62a9" Oct 02 02:03:56 crc kubenswrapper[4775]: E1002 02:03:56.995273 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e38c74d33b3ad7b8f747e53560c787d296c45bf29b3227467260f620932b62a9\": container with ID starting with e38c74d33b3ad7b8f747e53560c787d296c45bf29b3227467260f620932b62a9 not found: ID does not exist" containerID="e38c74d33b3ad7b8f747e53560c787d296c45bf29b3227467260f620932b62a9" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.995304 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e38c74d33b3ad7b8f747e53560c787d296c45bf29b3227467260f620932b62a9"} err="failed to get container status \"e38c74d33b3ad7b8f747e53560c787d296c45bf29b3227467260f620932b62a9\": rpc error: code = NotFound desc = could not find container \"e38c74d33b3ad7b8f747e53560c787d296c45bf29b3227467260f620932b62a9\": container with ID starting with e38c74d33b3ad7b8f747e53560c787d296c45bf29b3227467260f620932b62a9 not found: ID does not exist" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.995331 4775 scope.go:117] "RemoveContainer" containerID="04d6bdf03305dd966a2ca95b6a3220b41a5add02a16becda008422ee10720237" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.995789 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7da01b78-a318-448e-8a4c-7ad3d9e833c0/ovn-northd/0.log" Oct 02 02:03:56 crc kubenswrapper[4775]: I1002 02:03:56.995921 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 02:03:57 crc kubenswrapper[4775]: E1002 02:03:56.996941 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04d6bdf03305dd966a2ca95b6a3220b41a5add02a16becda008422ee10720237\": container with ID starting with 04d6bdf03305dd966a2ca95b6a3220b41a5add02a16becda008422ee10720237 not found: ID does not exist" containerID="04d6bdf03305dd966a2ca95b6a3220b41a5add02a16becda008422ee10720237" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:56.996988 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04d6bdf03305dd966a2ca95b6a3220b41a5add02a16becda008422ee10720237"} err="failed to get container status \"04d6bdf03305dd966a2ca95b6a3220b41a5add02a16becda008422ee10720237\": rpc error: code = NotFound desc = could not find container \"04d6bdf03305dd966a2ca95b6a3220b41a5add02a16becda008422ee10720237\": container with ID starting with 04d6bdf03305dd966a2ca95b6a3220b41a5add02a16becda008422ee10720237 not found: ID does not exist" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:56.997014 4775 scope.go:117] "RemoveContainer" containerID="db8059deef67f936d22e07a37161aed4eda44bdc81845a2918822f9946fd9c00" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.019982 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-777987c7d4-d5vzz"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.020134 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-777987c7d4-d5vzz"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.020208 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.032088 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.039927 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.043561 4775 scope.go:117] "RemoveContainer" containerID="35fba6a1f824b8a5a30f3f0d8b52922737c0bdb3b40b6987fcc54b88ccfc67ab" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.045656 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2d45342-c4de-48db-8f6b-23f26103497b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.053058 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.057460 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.061361 4775 scope.go:117] "RemoveContainer" containerID="db8059deef67f936d22e07a37161aed4eda44bdc81845a2918822f9946fd9c00" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.064107 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: E1002 02:03:57.064897 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db8059deef67f936d22e07a37161aed4eda44bdc81845a2918822f9946fd9c00\": container with ID starting with db8059deef67f936d22e07a37161aed4eda44bdc81845a2918822f9946fd9c00 not found: ID does not exist" containerID="db8059deef67f936d22e07a37161aed4eda44bdc81845a2918822f9946fd9c00" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.064926 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db8059deef67f936d22e07a37161aed4eda44bdc81845a2918822f9946fd9c00"} err="failed to get container status \"db8059deef67f936d22e07a37161aed4eda44bdc81845a2918822f9946fd9c00\": rpc error: code = NotFound desc = could not find container \"db8059deef67f936d22e07a37161aed4eda44bdc81845a2918822f9946fd9c00\": container with ID starting with db8059deef67f936d22e07a37161aed4eda44bdc81845a2918822f9946fd9c00 not found: ID does not exist" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.064945 4775 scope.go:117] "RemoveContainer" containerID="35fba6a1f824b8a5a30f3f0d8b52922737c0bdb3b40b6987fcc54b88ccfc67ab" Oct 02 02:03:57 crc kubenswrapper[4775]: E1002 02:03:57.065241 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35fba6a1f824b8a5a30f3f0d8b52922737c0bdb3b40b6987fcc54b88ccfc67ab\": container with ID starting with 35fba6a1f824b8a5a30f3f0d8b52922737c0bdb3b40b6987fcc54b88ccfc67ab not found: ID does not exist" containerID="35fba6a1f824b8a5a30f3f0d8b52922737c0bdb3b40b6987fcc54b88ccfc67ab" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.065285 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35fba6a1f824b8a5a30f3f0d8b52922737c0bdb3b40b6987fcc54b88ccfc67ab"} err="failed to get container status \"35fba6a1f824b8a5a30f3f0d8b52922737c0bdb3b40b6987fcc54b88ccfc67ab\": rpc error: code = NotFound desc = could not find container \"35fba6a1f824b8a5a30f3f0d8b52922737c0bdb3b40b6987fcc54b88ccfc67ab\": container with ID starting with 35fba6a1f824b8a5a30f3f0d8b52922737c0bdb3b40b6987fcc54b88ccfc67ab not found: ID does not exist" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.065313 4775 scope.go:117] "RemoveContainer" containerID="9d068bf0ca5c469b57dcad75c4e6ebff62f9badd694815d6349f36cd138a20a8" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.081164 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-79594d69d8-x75fk"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.084567 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-79594d69d8-x75fk"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.090370 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.096247 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.112416 4775 scope.go:117] "RemoveContainer" containerID="303fb7d6dfef981bea20389d78f64a2cf286846122e5c5d3878e6b65fef9edb7" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.113597 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.119345 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.146934 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da01b78-a318-448e-8a4c-7ad3d9e833c0-config\") pod \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.147008 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7da01b78-a318-448e-8a4c-7ad3d9e833c0-scripts\") pod \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.147096 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-ovn-northd-tls-certs\") pod \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.147156 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4n2g\" (UniqueName: \"kubernetes.io/projected/7da01b78-a318-448e-8a4c-7ad3d9e833c0-kube-api-access-q4n2g\") pod \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.147188 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-combined-ca-bundle\") pod \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.147339 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7da01b78-a318-448e-8a4c-7ad3d9e833c0-ovn-rundir\") pod \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.147390 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-metrics-certs-tls-certs\") pod \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\" (UID: \"7da01b78-a318-448e-8a4c-7ad3d9e833c0\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.150389 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da01b78-a318-448e-8a4c-7ad3d9e833c0-config" (OuterVolumeSpecName: "config") pod "7da01b78-a318-448e-8a4c-7ad3d9e833c0" (UID: "7da01b78-a318-448e-8a4c-7ad3d9e833c0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.150760 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7da01b78-a318-448e-8a4c-7ad3d9e833c0-scripts" (OuterVolumeSpecName: "scripts") pod "7da01b78-a318-448e-8a4c-7ad3d9e833c0" (UID: "7da01b78-a318-448e-8a4c-7ad3d9e833c0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.152213 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7da01b78-a318-448e-8a4c-7ad3d9e833c0-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "7da01b78-a318-448e-8a4c-7ad3d9e833c0" (UID: "7da01b78-a318-448e-8a4c-7ad3d9e833c0"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.155848 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7da01b78-a318-448e-8a4c-7ad3d9e833c0-kube-api-access-q4n2g" (OuterVolumeSpecName: "kube-api-access-q4n2g") pod "7da01b78-a318-448e-8a4c-7ad3d9e833c0" (UID: "7da01b78-a318-448e-8a4c-7ad3d9e833c0"). InnerVolumeSpecName "kube-api-access-q4n2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.182930 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.186869 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.204755 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.208251 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.214699 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7da01b78-a318-448e-8a4c-7ad3d9e833c0" (UID: "7da01b78-a318-448e-8a4c-7ad3d9e833c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.216134 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.220163 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.235716 4775 scope.go:117] "RemoveContainer" containerID="9d068bf0ca5c469b57dcad75c4e6ebff62f9badd694815d6349f36cd138a20a8" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.235715 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "7da01b78-a318-448e-8a4c-7ad3d9e833c0" (UID: "7da01b78-a318-448e-8a4c-7ad3d9e833c0"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: E1002 02:03:57.236148 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d068bf0ca5c469b57dcad75c4e6ebff62f9badd694815d6349f36cd138a20a8\": container with ID starting with 9d068bf0ca5c469b57dcad75c4e6ebff62f9badd694815d6349f36cd138a20a8 not found: ID does not exist" containerID="9d068bf0ca5c469b57dcad75c4e6ebff62f9badd694815d6349f36cd138a20a8" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.236181 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d068bf0ca5c469b57dcad75c4e6ebff62f9badd694815d6349f36cd138a20a8"} err="failed to get container status \"9d068bf0ca5c469b57dcad75c4e6ebff62f9badd694815d6349f36cd138a20a8\": rpc error: code = NotFound desc = could not find container \"9d068bf0ca5c469b57dcad75c4e6ebff62f9badd694815d6349f36cd138a20a8\": container with ID starting with 9d068bf0ca5c469b57dcad75c4e6ebff62f9badd694815d6349f36cd138a20a8 not found: ID does not exist" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.236206 4775 scope.go:117] "RemoveContainer" containerID="303fb7d6dfef981bea20389d78f64a2cf286846122e5c5d3878e6b65fef9edb7" Oct 02 02:03:57 crc kubenswrapper[4775]: E1002 02:03:57.237296 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"303fb7d6dfef981bea20389d78f64a2cf286846122e5c5d3878e6b65fef9edb7\": container with ID starting with 303fb7d6dfef981bea20389d78f64a2cf286846122e5c5d3878e6b65fef9edb7 not found: ID does not exist" containerID="303fb7d6dfef981bea20389d78f64a2cf286846122e5c5d3878e6b65fef9edb7" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.237335 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"303fb7d6dfef981bea20389d78f64a2cf286846122e5c5d3878e6b65fef9edb7"} err="failed to get container status \"303fb7d6dfef981bea20389d78f64a2cf286846122e5c5d3878e6b65fef9edb7\": rpc error: code = NotFound desc = could not find container \"303fb7d6dfef981bea20389d78f64a2cf286846122e5c5d3878e6b65fef9edb7\": container with ID starting with 303fb7d6dfef981bea20389d78f64a2cf286846122e5c5d3878e6b65fef9edb7 not found: ID does not exist" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.237370 4775 scope.go:117] "RemoveContainer" containerID="63252743cedda5425e3430e58790235c337b97aaa0683e299068d0092eca5e7f" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.247668 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "7da01b78-a318-448e-8a4c-7ad3d9e833c0" (UID: "7da01b78-a318-448e-8a4c-7ad3d9e833c0"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.250686 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7da01b78-a318-448e-8a4c-7ad3d9e833c0-ovn-rundir\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.250708 4775 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.250718 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da01b78-a318-448e-8a4c-7ad3d9e833c0-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.250728 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7da01b78-a318-448e-8a4c-7ad3d9e833c0-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.250736 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.250744 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4n2g\" (UniqueName: \"kubernetes.io/projected/7da01b78-a318-448e-8a4c-7ad3d9e833c0-kube-api-access-q4n2g\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.250753 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7da01b78-a318-448e-8a4c-7ad3d9e833c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.257510 4775 scope.go:117] "RemoveContainer" containerID="12f66def0a878a3dc74caa0f75419c15f378c889154267260d12b5003e5ebee3" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.273448 4775 scope.go:117] "RemoveContainer" containerID="63252743cedda5425e3430e58790235c337b97aaa0683e299068d0092eca5e7f" Oct 02 02:03:57 crc kubenswrapper[4775]: E1002 02:03:57.273711 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63252743cedda5425e3430e58790235c337b97aaa0683e299068d0092eca5e7f\": container with ID starting with 63252743cedda5425e3430e58790235c337b97aaa0683e299068d0092eca5e7f not found: ID does not exist" containerID="63252743cedda5425e3430e58790235c337b97aaa0683e299068d0092eca5e7f" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.273740 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63252743cedda5425e3430e58790235c337b97aaa0683e299068d0092eca5e7f"} err="failed to get container status \"63252743cedda5425e3430e58790235c337b97aaa0683e299068d0092eca5e7f\": rpc error: code = NotFound desc = could not find container \"63252743cedda5425e3430e58790235c337b97aaa0683e299068d0092eca5e7f\": container with ID starting with 63252743cedda5425e3430e58790235c337b97aaa0683e299068d0092eca5e7f not found: ID does not exist" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.273762 4775 scope.go:117] "RemoveContainer" containerID="12f66def0a878a3dc74caa0f75419c15f378c889154267260d12b5003e5ebee3" Oct 02 02:03:57 crc kubenswrapper[4775]: E1002 02:03:57.273925 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12f66def0a878a3dc74caa0f75419c15f378c889154267260d12b5003e5ebee3\": container with ID starting with 12f66def0a878a3dc74caa0f75419c15f378c889154267260d12b5003e5ebee3 not found: ID does not exist" containerID="12f66def0a878a3dc74caa0f75419c15f378c889154267260d12b5003e5ebee3" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.273944 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12f66def0a878a3dc74caa0f75419c15f378c889154267260d12b5003e5ebee3"} err="failed to get container status \"12f66def0a878a3dc74caa0f75419c15f378c889154267260d12b5003e5ebee3\": rpc error: code = NotFound desc = could not find container \"12f66def0a878a3dc74caa0f75419c15f378c889154267260d12b5003e5ebee3\": container with ID starting with 12f66def0a878a3dc74caa0f75419c15f378c889154267260d12b5003e5ebee3 not found: ID does not exist" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.273967 4775 scope.go:117] "RemoveContainer" containerID="062b09ca95edb1ddddcb2de55757014ea328d800d380273e8fb5c4e3bb488c11" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.293910 4775 scope.go:117] "RemoveContainer" containerID="33e28213f9c089646224401e310e6c70d1efdff65261c89fe379e544e09ae98f" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.323082 4775 scope.go:117] "RemoveContainer" containerID="0d717b8fb0cfdfcf5e5314de128b353f94395b4da93ebdbec0e1f666b4b81244" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.371057 4775 scope.go:117] "RemoveContainer" containerID="ddc807916bf2152781e0b263b81d2ba2f8ff95a514f51a7d7918774142370f47" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.389457 4775 scope.go:117] "RemoveContainer" containerID="9f162e9f65c919c062e4c98961fb915d6d73a9e2d76b98f4844b90766369c972" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.413912 4775 scope.go:117] "RemoveContainer" containerID="2cf05392cd7bff8e546c2fbad0965375d71482567f75af41133d05802bad7f02" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.443181 4775 scope.go:117] "RemoveContainer" containerID="1c6c4af35ca368385ff971101cf668bbedef2ac9915012172cc65ff5519d556e" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.479938 4775 scope.go:117] "RemoveContainer" containerID="0a11b7f88b7e2c013d1797c32a7ec78c81f775682c0323ea8b58d4f35f6e04b7" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.495436 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.496594 4775 scope.go:117] "RemoveContainer" containerID="4feb9154e0329bc25829a25b496b1b2573ea599dcffdc17eb82bbc2b54f58aa4" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.550237 4775 scope.go:117] "RemoveContainer" containerID="8df9de2a0e96e291b49efa835c2476332824829ee66a90f08c5fa720d9e412f8" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.657197 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"002eb223-af34-4c80-9570-894add9b0e3c\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.657271 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-galera-tls-certs\") pod \"002eb223-af34-4c80-9570-894add9b0e3c\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.657302 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-combined-ca-bundle\") pod \"002eb223-af34-4c80-9570-894add9b0e3c\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.657332 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzcxp\" (UniqueName: \"kubernetes.io/projected/002eb223-af34-4c80-9570-894add9b0e3c-kube-api-access-vzcxp\") pod \"002eb223-af34-4c80-9570-894add9b0e3c\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.657358 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-kolla-config\") pod \"002eb223-af34-4c80-9570-894add9b0e3c\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.657392 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/002eb223-af34-4c80-9570-894add9b0e3c-config-data-generated\") pod \"002eb223-af34-4c80-9570-894add9b0e3c\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.657441 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-operator-scripts\") pod \"002eb223-af34-4c80-9570-894add9b0e3c\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.657476 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-secrets\") pod \"002eb223-af34-4c80-9570-894add9b0e3c\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.657518 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-config-data-default\") pod \"002eb223-af34-4c80-9570-894add9b0e3c\" (UID: \"002eb223-af34-4c80-9570-894add9b0e3c\") " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.659233 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/002eb223-af34-4c80-9570-894add9b0e3c-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "002eb223-af34-4c80-9570-894add9b0e3c" (UID: "002eb223-af34-4c80-9570-894add9b0e3c"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.659630 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "002eb223-af34-4c80-9570-894add9b0e3c" (UID: "002eb223-af34-4c80-9570-894add9b0e3c"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.659628 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "002eb223-af34-4c80-9570-894add9b0e3c" (UID: "002eb223-af34-4c80-9570-894add9b0e3c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.659722 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "002eb223-af34-4c80-9570-894add9b0e3c" (UID: "002eb223-af34-4c80-9570-894add9b0e3c"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.668093 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-secrets" (OuterVolumeSpecName: "secrets") pod "002eb223-af34-4c80-9570-894add9b0e3c" (UID: "002eb223-af34-4c80-9570-894add9b0e3c"). InnerVolumeSpecName "secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.668297 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/002eb223-af34-4c80-9570-894add9b0e3c-kube-api-access-vzcxp" (OuterVolumeSpecName: "kube-api-access-vzcxp") pod "002eb223-af34-4c80-9570-894add9b0e3c" (UID: "002eb223-af34-4c80-9570-894add9b0e3c"). InnerVolumeSpecName "kube-api-access-vzcxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.683034 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "mysql-db") pod "002eb223-af34-4c80-9570-894add9b0e3c" (UID: "002eb223-af34-4c80-9570-894add9b0e3c"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.693271 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "002eb223-af34-4c80-9570-894add9b0e3c" (UID: "002eb223-af34-4c80-9570-894add9b0e3c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.715290 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "002eb223-af34-4c80-9570-894add9b0e3c" (UID: "002eb223-af34-4c80-9570-894add9b0e3c"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.759699 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-config-data-default\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.759769 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.759790 4775 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.759811 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.759829 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzcxp\" (UniqueName: \"kubernetes.io/projected/002eb223-af34-4c80-9570-894add9b0e3c-kube-api-access-vzcxp\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.759848 4775 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-kolla-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.759864 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/002eb223-af34-4c80-9570-894add9b0e3c-config-data-generated\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.759880 4775 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/002eb223-af34-4c80-9570-894add9b0e3c-operator-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.759896 4775 reconciler_common.go:293] "Volume detached for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/002eb223-af34-4c80-9570-894add9b0e3c-secrets\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: E1002 02:03:57.760610 4775 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Oct 02 02:03:57 crc kubenswrapper[4775]: E1002 02:03:57.760841 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data podName:fb0a7ddd-1558-4ad8-a8f4-e140306d19f9 nodeName:}" failed. No retries permitted until 2025-10-02 02:04:05.760810079 +0000 UTC m=+1382.927554159 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data") pod "rabbitmq-cell1-server-0" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9") : configmap "rabbitmq-cell1-config-data" not found Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.784768 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e5e6844-45f6-4fd9-ac4f-3df2086164eb" path="/var/lib/kubelet/pods/1e5e6844-45f6-4fd9-ac4f-3df2086164eb/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.785750 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24ef8940-eb43-459c-9cfe-854df57cfe74" path="/var/lib/kubelet/pods/24ef8940-eb43-459c-9cfe-854df57cfe74/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.786888 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="406dd24f-d800-44be-b33d-019c532c2feb" path="/var/lib/kubelet/pods/406dd24f-d800-44be-b33d-019c532c2feb/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.788042 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46453bb2-6af0-4dd4-83ba-b14bf3f96310" path="/var/lib/kubelet/pods/46453bb2-6af0-4dd4-83ba-b14bf3f96310/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.789866 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56813183-3dfc-430c-aaac-bda27d91340b" path="/var/lib/kubelet/pods/56813183-3dfc-430c-aaac-bda27d91340b/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.791064 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5bca47d0-79b7-4992-867b-d5e05e02cdc8" path="/var/lib/kubelet/pods/5bca47d0-79b7-4992-867b-d5e05e02cdc8/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.793391 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e322679-b95d-4328-b03b-0be7f0e82ac3" path="/var/lib/kubelet/pods/5e322679-b95d-4328-b03b-0be7f0e82ac3/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.794869 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" path="/var/lib/kubelet/pods/8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.797138 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d6bb525-8c94-461f-84c8-9e466b789706" path="/var/lib/kubelet/pods/8d6bb525-8c94-461f-84c8-9e466b789706/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.799243 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f286b47-509d-479d-bad5-0bbec930558a" path="/var/lib/kubelet/pods/8f286b47-509d-479d-bad5-0bbec930558a/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.800503 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" path="/var/lib/kubelet/pods/8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.800732 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.806475 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5c60336-6e25-4c68-98e6-52801f34578b" path="/var/lib/kubelet/pods/a5c60336-6e25-4c68-98e6-52801f34578b/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.807489 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af16c5f4-fcd0-4450-a8b7-789ff104e477" path="/var/lib/kubelet/pods/af16c5f4-fcd0-4450-a8b7-789ff104e477/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.808545 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c74e0a1d-125d-43f7-9ba3-cea70453fc01" path="/var/lib/kubelet/pods/c74e0a1d-125d-43f7-9ba3-cea70453fc01/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.814503 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a" path="/var/lib/kubelet/pods/ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.816681 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdea921f-d632-4463-a011-2e6f23f70546" path="/var/lib/kubelet/pods/cdea921f-d632-4463-a011-2e6f23f70546/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.818668 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" path="/var/lib/kubelet/pods/d2d45342-c4de-48db-8f6b-23f26103497b/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.821626 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3732523-d2de-45c9-ac69-770197f413f3" path="/var/lib/kubelet/pods/e3732523-d2de-45c9-ac69-770197f413f3/volumes" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.843703 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_7da01b78-a318-448e-8a4c-7ad3d9e833c0/ovn-northd/0.log" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.843840 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.843801 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"7da01b78-a318-448e-8a4c-7ad3d9e833c0","Type":"ContainerDied","Data":"480c57752e934fae2adadd95f35d2b6dba80fd0e3972c13c154118141eeb3a0b"} Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.843989 4775 scope.go:117] "RemoveContainer" containerID="5b7338db5965862d0736735b7d766b4ed845f3e4cf8b4b81acd87174941ecfa6" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.861190 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.864766 4775 generic.go:334] "Generic (PLEG): container finished" podID="002eb223-af34-4c80-9570-894add9b0e3c" containerID="8569575c44b11cca45a36d7a16be245143dd1895ca868b19ec6f5372e6eb3a4d" exitCode=0 Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.864827 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"002eb223-af34-4c80-9570-894add9b0e3c","Type":"ContainerDied","Data":"8569575c44b11cca45a36d7a16be245143dd1895ca868b19ec6f5372e6eb3a4d"} Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.864859 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"002eb223-af34-4c80-9570-894add9b0e3c","Type":"ContainerDied","Data":"6e8fd5394954be2252799b84e7920c64d1173fc31dfcd3a7f3e1daaa735ff5bc"} Oct 02 02:03:57 crc kubenswrapper[4775]: I1002 02:03:57.864938 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 02:03:58 crc kubenswrapper[4775]: E1002 02:03:58.098352 4775 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Oct 02 02:03:58 crc kubenswrapper[4775]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-02T02:03:50Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 02 02:03:58 crc kubenswrapper[4775]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Oct 02 02:03:58 crc kubenswrapper[4775]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-jdbs4" message=< Oct 02 02:03:58 crc kubenswrapper[4775]: Exiting ovn-controller (1) [FAILED] Oct 02 02:03:58 crc kubenswrapper[4775]: Killing ovn-controller (1) [ OK ] Oct 02 02:03:58 crc kubenswrapper[4775]: Killing ovn-controller (1) with SIGKILL [ OK ] Oct 02 02:03:58 crc kubenswrapper[4775]: 2025-10-02T02:03:50Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 02 02:03:58 crc kubenswrapper[4775]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Oct 02 02:03:58 crc kubenswrapper[4775]: > Oct 02 02:03:58 crc kubenswrapper[4775]: E1002 02:03:58.098407 4775 kuberuntime_container.go:691] "PreStop hook failed" err=< Oct 02 02:03:58 crc kubenswrapper[4775]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-10-02T02:03:50Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Oct 02 02:03:58 crc kubenswrapper[4775]: /etc/init.d/functions: line 589: 400 Alarm clock "$@" Oct 02 02:03:58 crc kubenswrapper[4775]: > pod="openstack/ovn-controller-jdbs4" podUID="6efbe4b3-953d-4927-95c6-9b92708f51eb" containerName="ovn-controller" containerID="cri-o://324549bfccf9b231fa568c7775d9ff418271c352066eb77651c4318e12b8d121" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.098457 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-jdbs4" podUID="6efbe4b3-953d-4927-95c6-9b92708f51eb" containerName="ovn-controller" containerID="cri-o://324549bfccf9b231fa568c7775d9ff418271c352066eb77651c4318e12b8d121" gracePeriod=22 Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.199247 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.211862 4775 scope.go:117] "RemoveContainer" containerID="39f92c334c8a479d69a76f6ee477dc9e38b2f2fb13b731fddbffda7d7dac3511" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.213063 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.217939 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.222465 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.235943 4775 scope.go:117] "RemoveContainer" containerID="8569575c44b11cca45a36d7a16be245143dd1895ca868b19ec6f5372e6eb3a4d" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.260865 4775 scope.go:117] "RemoveContainer" containerID="2563efa50bd9d9570bf76e9b0e57b1c4fa37d48c0fae0b8b0e1744feefe65559" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.314041 4775 scope.go:117] "RemoveContainer" containerID="8569575c44b11cca45a36d7a16be245143dd1895ca868b19ec6f5372e6eb3a4d" Oct 02 02:03:58 crc kubenswrapper[4775]: E1002 02:03:58.314777 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8569575c44b11cca45a36d7a16be245143dd1895ca868b19ec6f5372e6eb3a4d\": container with ID starting with 8569575c44b11cca45a36d7a16be245143dd1895ca868b19ec6f5372e6eb3a4d not found: ID does not exist" containerID="8569575c44b11cca45a36d7a16be245143dd1895ca868b19ec6f5372e6eb3a4d" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.314826 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8569575c44b11cca45a36d7a16be245143dd1895ca868b19ec6f5372e6eb3a4d"} err="failed to get container status \"8569575c44b11cca45a36d7a16be245143dd1895ca868b19ec6f5372e6eb3a4d\": rpc error: code = NotFound desc = could not find container \"8569575c44b11cca45a36d7a16be245143dd1895ca868b19ec6f5372e6eb3a4d\": container with ID starting with 8569575c44b11cca45a36d7a16be245143dd1895ca868b19ec6f5372e6eb3a4d not found: ID does not exist" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.314858 4775 scope.go:117] "RemoveContainer" containerID="2563efa50bd9d9570bf76e9b0e57b1c4fa37d48c0fae0b8b0e1744feefe65559" Oct 02 02:03:58 crc kubenswrapper[4775]: E1002 02:03:58.315425 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2563efa50bd9d9570bf76e9b0e57b1c4fa37d48c0fae0b8b0e1744feefe65559\": container with ID starting with 2563efa50bd9d9570bf76e9b0e57b1c4fa37d48c0fae0b8b0e1744feefe65559 not found: ID does not exist" containerID="2563efa50bd9d9570bf76e9b0e57b1c4fa37d48c0fae0b8b0e1744feefe65559" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.315512 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2563efa50bd9d9570bf76e9b0e57b1c4fa37d48c0fae0b8b0e1744feefe65559"} err="failed to get container status \"2563efa50bd9d9570bf76e9b0e57b1c4fa37d48c0fae0b8b0e1744feefe65559\": rpc error: code = NotFound desc = could not find container \"2563efa50bd9d9570bf76e9b0e57b1c4fa37d48c0fae0b8b0e1744feefe65559\": container with ID starting with 2563efa50bd9d9570bf76e9b0e57b1c4fa37d48c0fae0b8b0e1744feefe65559 not found: ID does not exist" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.424625 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.545937 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jdbs4_6efbe4b3-953d-4927-95c6-9b92708f51eb/ovn-controller/0.log" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.546023 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jdbs4" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.559166 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.569705 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-564487c87-zppvw" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.574335 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-tls\") pod \"6638baa9-b367-40d1-8111-673dca6434ad\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.574375 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-confd\") pod \"6638baa9-b367-40d1-8111-673dca6434ad\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.574415 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-plugins-conf\") pod \"6638baa9-b367-40d1-8111-673dca6434ad\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.574433 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-server-conf\") pod \"6638baa9-b367-40d1-8111-673dca6434ad\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.574462 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkgpt\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-kube-api-access-bkgpt\") pod \"6638baa9-b367-40d1-8111-673dca6434ad\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.574492 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data\") pod \"6638baa9-b367-40d1-8111-673dca6434ad\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.574527 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-erlang-cookie\") pod \"6638baa9-b367-40d1-8111-673dca6434ad\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.574543 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-plugins\") pod \"6638baa9-b367-40d1-8111-673dca6434ad\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.574580 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6638baa9-b367-40d1-8111-673dca6434ad-erlang-cookie-secret\") pod \"6638baa9-b367-40d1-8111-673dca6434ad\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.574608 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"6638baa9-b367-40d1-8111-673dca6434ad\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.574650 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6638baa9-b367-40d1-8111-673dca6434ad-pod-info\") pod \"6638baa9-b367-40d1-8111-673dca6434ad\" (UID: \"6638baa9-b367-40d1-8111-673dca6434ad\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.575426 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "6638baa9-b367-40d1-8111-673dca6434ad" (UID: "6638baa9-b367-40d1-8111-673dca6434ad"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.575986 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "6638baa9-b367-40d1-8111-673dca6434ad" (UID: "6638baa9-b367-40d1-8111-673dca6434ad"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.577031 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "6638baa9-b367-40d1-8111-673dca6434ad" (UID: "6638baa9-b367-40d1-8111-673dca6434ad"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.580281 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/6638baa9-b367-40d1-8111-673dca6434ad-pod-info" (OuterVolumeSpecName: "pod-info") pod "6638baa9-b367-40d1-8111-673dca6434ad" (UID: "6638baa9-b367-40d1-8111-673dca6434ad"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.580713 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-kube-api-access-bkgpt" (OuterVolumeSpecName: "kube-api-access-bkgpt") pod "6638baa9-b367-40d1-8111-673dca6434ad" (UID: "6638baa9-b367-40d1-8111-673dca6434ad"). InnerVolumeSpecName "kube-api-access-bkgpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.581531 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "6638baa9-b367-40d1-8111-673dca6434ad" (UID: "6638baa9-b367-40d1-8111-673dca6434ad"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.582149 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6638baa9-b367-40d1-8111-673dca6434ad-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "6638baa9-b367-40d1-8111-673dca6434ad" (UID: "6638baa9-b367-40d1-8111-673dca6434ad"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.582286 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "6638baa9-b367-40d1-8111-673dca6434ad" (UID: "6638baa9-b367-40d1-8111-673dca6434ad"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.600635 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data" (OuterVolumeSpecName: "config-data") pod "6638baa9-b367-40d1-8111-673dca6434ad" (UID: "6638baa9-b367-40d1-8111-673dca6434ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.627427 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-server-conf" (OuterVolumeSpecName: "server-conf") pod "6638baa9-b367-40d1-8111-673dca6434ad" (UID: "6638baa9-b367-40d1-8111-673dca6434ad"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.665117 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "6638baa9-b367-40d1-8111-673dca6434ad" (UID: "6638baa9-b367-40d1-8111-673dca6434ad"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675516 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-public-tls-certs\") pod \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675562 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-confd\") pod \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675588 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675610 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-erlang-cookie-secret\") pod \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675634 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwbt6\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-kube-api-access-wwbt6\") pod \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675651 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-server-conf\") pod \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675670 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-tls\") pod \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675692 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-plugins-conf\") pod \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675708 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-credential-keys\") pod \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675733 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-plugins\") pod \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675757 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-pod-info\") pod \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675777 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-log-ovn\") pod \"6efbe4b3-953d-4927-95c6-9b92708f51eb\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675798 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-fernet-keys\") pod \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675813 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-internal-tls-certs\") pod \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675843 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efbe4b3-953d-4927-95c6-9b92708f51eb-combined-ca-bundle\") pod \"6efbe4b3-953d-4927-95c6-9b92708f51eb\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675872 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6efbe4b3-953d-4927-95c6-9b92708f51eb-ovn-controller-tls-certs\") pod \"6efbe4b3-953d-4927-95c6-9b92708f51eb\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675888 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6efbe4b3-953d-4927-95c6-9b92708f51eb-scripts\") pod \"6efbe4b3-953d-4927-95c6-9b92708f51eb\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675907 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfxrf\" (UniqueName: \"kubernetes.io/projected/6efbe4b3-953d-4927-95c6-9b92708f51eb-kube-api-access-vfxrf\") pod \"6efbe4b3-953d-4927-95c6-9b92708f51eb\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675924 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-scripts\") pod \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675963 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-config-data\") pod \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675978 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-run-ovn\") pod \"6efbe4b3-953d-4927-95c6-9b92708f51eb\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.675994 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-combined-ca-bundle\") pod \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676013 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lpv5\" (UniqueName: \"kubernetes.io/projected/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-kube-api-access-7lpv5\") pod \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\" (UID: \"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676033 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data\") pod \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676048 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-erlang-cookie\") pod \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\" (UID: \"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676068 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-run\") pod \"6efbe4b3-953d-4927-95c6-9b92708f51eb\" (UID: \"6efbe4b3-953d-4927-95c6-9b92708f51eb\") " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676351 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676369 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676378 4775 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676385 4775 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676394 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkgpt\" (UniqueName: \"kubernetes.io/projected/6638baa9-b367-40d1-8111-673dca6434ad-kube-api-access-bkgpt\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676404 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6638baa9-b367-40d1-8111-673dca6434ad-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676413 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676421 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6638baa9-b367-40d1-8111-673dca6434ad-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676429 4775 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6638baa9-b367-40d1-8111-673dca6434ad-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676447 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.676457 4775 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6638baa9-b367-40d1-8111-673dca6434ad-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.678015 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6efbe4b3-953d-4927-95c6-9b92708f51eb-scripts" (OuterVolumeSpecName: "scripts") pod "6efbe4b3-953d-4927-95c6-9b92708f51eb" (UID: "6efbe4b3-953d-4927-95c6-9b92708f51eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.678836 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-scripts" (OuterVolumeSpecName: "scripts") pod "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e" (UID: "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.679834 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.680895 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-run" (OuterVolumeSpecName: "var-run") pod "6efbe4b3-953d-4927-95c6-9b92708f51eb" (UID: "6efbe4b3-953d-4927-95c6-9b92708f51eb"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.681124 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6efbe4b3-953d-4927-95c6-9b92708f51eb" (UID: "6efbe4b3-953d-4927-95c6-9b92708f51eb"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.683992 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6efbe4b3-953d-4927-95c6-9b92708f51eb" (UID: "6efbe4b3-953d-4927-95c6-9b92708f51eb"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.689435 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.689758 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.692129 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-pod-info" (OuterVolumeSpecName: "pod-info") pod "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.694055 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-kube-api-access-7lpv5" (OuterVolumeSpecName: "kube-api-access-7lpv5") pod "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e" (UID: "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e"). InnerVolumeSpecName "kube-api-access-7lpv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.695626 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e" (UID: "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.695730 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6efbe4b3-953d-4927-95c6-9b92708f51eb-kube-api-access-vfxrf" (OuterVolumeSpecName: "kube-api-access-vfxrf") pod "6efbe4b3-953d-4927-95c6-9b92708f51eb" (UID: "6efbe4b3-953d-4927-95c6-9b92708f51eb"). InnerVolumeSpecName "kube-api-access-vfxrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.696088 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: E1002 02:03:58.698535 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.709939 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e" (UID: "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.711106 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: E1002 02:03:58.717370 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 02:03:58 crc kubenswrapper[4775]: E1002 02:03:58.718743 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 02:03:58 crc kubenswrapper[4775]: E1002 02:03:58.718774 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="a64f8ba3-65b9-489f-ace1-757f3ecce6b4" containerName="nova-scheduler-scheduler" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.719998 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.720363 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-kube-api-access-wwbt6" (OuterVolumeSpecName: "kube-api-access-wwbt6") pod "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9"). InnerVolumeSpecName "kube-api-access-wwbt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.721430 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.725319 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e" (UID: "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.727046 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-config-data" (OuterVolumeSpecName: "config-data") pod "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e" (UID: "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.741293 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data" (OuterVolumeSpecName: "config-data") pod "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.760847 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6efbe4b3-953d-4927-95c6-9b92708f51eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6efbe4b3-953d-4927-95c6-9b92708f51eb" (UID: "6efbe4b3-953d-4927-95c6-9b92708f51eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.762525 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-server-conf" (OuterVolumeSpecName: "server-conf") pod "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.764225 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e" (UID: "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.766061 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e" (UID: "1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777409 4775 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777435 4775 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777444 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777452 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efbe4b3-953d-4927-95c6-9b92708f51eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777460 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777469 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6efbe4b3-953d-4927-95c6-9b92708f51eb-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777477 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfxrf\" (UniqueName: \"kubernetes.io/projected/6efbe4b3-953d-4927-95c6-9b92708f51eb-kube-api-access-vfxrf\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777486 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777557 4775 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777566 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777574 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777582 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lpv5\" (UniqueName: \"kubernetes.io/projected/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-kube-api-access-7lpv5\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777590 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777598 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777606 4775 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6efbe4b3-953d-4927-95c6-9b92708f51eb-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777613 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777640 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777648 4775 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777656 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwbt6\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-kube-api-access-wwbt6\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777664 4775 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777672 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777679 4775 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777687 4775 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777694 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.777702 4775 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.782898 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6efbe4b3-953d-4927-95c6-9b92708f51eb-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "6efbe4b3-953d-4927-95c6-9b92708f51eb" (UID: "6efbe4b3-953d-4927-95c6-9b92708f51eb"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.790673 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.817333 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" (UID: "fb0a7ddd-1558-4ad8-a8f4-e140306d19f9"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.880309 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/6efbe4b3-953d-4927-95c6-9b92708f51eb-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.880351 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.880390 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.886742 4775 generic.go:334] "Generic (PLEG): container finished" podID="fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" containerID="c4f574fe362f1c98b06c39ae135acf138f8c156d889aa930a36b192af34be971" exitCode=0 Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.886995 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9","Type":"ContainerDied","Data":"c4f574fe362f1c98b06c39ae135acf138f8c156d889aa930a36b192af34be971"} Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.887128 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"fb0a7ddd-1558-4ad8-a8f4-e140306d19f9","Type":"ContainerDied","Data":"1afd2c34b0364a61dfeee78253b8f225d6fa44e8781067eff8c919f7b3f28494"} Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.887217 4775 scope.go:117] "RemoveContainer" containerID="c4f574fe362f1c98b06c39ae135acf138f8c156d889aa930a36b192af34be971" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.887442 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.898752 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jdbs4_6efbe4b3-953d-4927-95c6-9b92708f51eb/ovn-controller/0.log" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.898807 4775 generic.go:334] "Generic (PLEG): container finished" podID="6efbe4b3-953d-4927-95c6-9b92708f51eb" containerID="324549bfccf9b231fa568c7775d9ff418271c352066eb77651c4318e12b8d121" exitCode=137 Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.898863 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jdbs4" event={"ID":"6efbe4b3-953d-4927-95c6-9b92708f51eb","Type":"ContainerDied","Data":"324549bfccf9b231fa568c7775d9ff418271c352066eb77651c4318e12b8d121"} Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.898890 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jdbs4" event={"ID":"6efbe4b3-953d-4927-95c6-9b92708f51eb","Type":"ContainerDied","Data":"ce846bbdc94bc41919dce048b75479d3b6f39d5af12599c009c079ae055ad4bc"} Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.898942 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jdbs4" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.902176 4775 generic.go:334] "Generic (PLEG): container finished" podID="6638baa9-b367-40d1-8111-673dca6434ad" containerID="836e86acc5faf2a3e9656e8d8d2d493dd928f3335bd576fc6d293e8ae72cc25c" exitCode=0 Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.902235 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6638baa9-b367-40d1-8111-673dca6434ad","Type":"ContainerDied","Data":"836e86acc5faf2a3e9656e8d8d2d493dd928f3335bd576fc6d293e8ae72cc25c"} Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.902265 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"6638baa9-b367-40d1-8111-673dca6434ad","Type":"ContainerDied","Data":"dd1198b6383e7906624aea451ad61826c0a3208f8155d62582ed5f31e692daaa"} Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.902331 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.906617 4775 generic.go:334] "Generic (PLEG): container finished" podID="1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e" containerID="7ad48d14db57c9f5d0c7e7d3fcaf64adb99c760e739a396090a74eb51c2a9919" exitCode=0 Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.906892 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-564487c87-zppvw" event={"ID":"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e","Type":"ContainerDied","Data":"7ad48d14db57c9f5d0c7e7d3fcaf64adb99c760e739a396090a74eb51c2a9919"} Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.907099 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-564487c87-zppvw" event={"ID":"1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e","Type":"ContainerDied","Data":"804847e082ad32e4c521651e8fb68070dd2ed6848268db247f5369cd1293a592"} Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.907419 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-564487c87-zppvw" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.930147 4775 scope.go:117] "RemoveContainer" containerID="ab0021968685a2c8011cf5e84c4a1e4f33be77976a9063ea49b60d6569a853a3" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.943057 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.953070 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.958777 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.970091 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.980985 4775 scope.go:117] "RemoveContainer" containerID="c4f574fe362f1c98b06c39ae135acf138f8c156d889aa930a36b192af34be971" Oct 02 02:03:58 crc kubenswrapper[4775]: E1002 02:03:58.981887 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4f574fe362f1c98b06c39ae135acf138f8c156d889aa930a36b192af34be971\": container with ID starting with c4f574fe362f1c98b06c39ae135acf138f8c156d889aa930a36b192af34be971 not found: ID does not exist" containerID="c4f574fe362f1c98b06c39ae135acf138f8c156d889aa930a36b192af34be971" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.981919 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4f574fe362f1c98b06c39ae135acf138f8c156d889aa930a36b192af34be971"} err="failed to get container status \"c4f574fe362f1c98b06c39ae135acf138f8c156d889aa930a36b192af34be971\": rpc error: code = NotFound desc = could not find container \"c4f574fe362f1c98b06c39ae135acf138f8c156d889aa930a36b192af34be971\": container with ID starting with c4f574fe362f1c98b06c39ae135acf138f8c156d889aa930a36b192af34be971 not found: ID does not exist" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.981943 4775 scope.go:117] "RemoveContainer" containerID="ab0021968685a2c8011cf5e84c4a1e4f33be77976a9063ea49b60d6569a853a3" Oct 02 02:03:58 crc kubenswrapper[4775]: E1002 02:03:58.982357 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab0021968685a2c8011cf5e84c4a1e4f33be77976a9063ea49b60d6569a853a3\": container with ID starting with ab0021968685a2c8011cf5e84c4a1e4f33be77976a9063ea49b60d6569a853a3 not found: ID does not exist" containerID="ab0021968685a2c8011cf5e84c4a1e4f33be77976a9063ea49b60d6569a853a3" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.982387 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab0021968685a2c8011cf5e84c4a1e4f33be77976a9063ea49b60d6569a853a3"} err="failed to get container status \"ab0021968685a2c8011cf5e84c4a1e4f33be77976a9063ea49b60d6569a853a3\": rpc error: code = NotFound desc = could not find container \"ab0021968685a2c8011cf5e84c4a1e4f33be77976a9063ea49b60d6569a853a3\": container with ID starting with ab0021968685a2c8011cf5e84c4a1e4f33be77976a9063ea49b60d6569a853a3 not found: ID does not exist" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.982405 4775 scope.go:117] "RemoveContainer" containerID="324549bfccf9b231fa568c7775d9ff418271c352066eb77651c4318e12b8d121" Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.989327 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jdbs4"] Oct 02 02:03:58 crc kubenswrapper[4775]: I1002 02:03:58.998904 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jdbs4"] Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.004374 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-564487c87-zppvw"] Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.004495 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="a5c60336-6e25-4c68-98e6-52801f34578b" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.197:6080/vnc_lite.html\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.005394 4775 scope.go:117] "RemoveContainer" containerID="324549bfccf9b231fa568c7775d9ff418271c352066eb77651c4318e12b8d121" Oct 02 02:03:59 crc kubenswrapper[4775]: E1002 02:03:59.005889 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"324549bfccf9b231fa568c7775d9ff418271c352066eb77651c4318e12b8d121\": container with ID starting with 324549bfccf9b231fa568c7775d9ff418271c352066eb77651c4318e12b8d121 not found: ID does not exist" containerID="324549bfccf9b231fa568c7775d9ff418271c352066eb77651c4318e12b8d121" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.005948 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"324549bfccf9b231fa568c7775d9ff418271c352066eb77651c4318e12b8d121"} err="failed to get container status \"324549bfccf9b231fa568c7775d9ff418271c352066eb77651c4318e12b8d121\": rpc error: code = NotFound desc = could not find container \"324549bfccf9b231fa568c7775d9ff418271c352066eb77651c4318e12b8d121\": container with ID starting with 324549bfccf9b231fa568c7775d9ff418271c352066eb77651c4318e12b8d121 not found: ID does not exist" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.006013 4775 scope.go:117] "RemoveContainer" containerID="836e86acc5faf2a3e9656e8d8d2d493dd928f3335bd576fc6d293e8ae72cc25c" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.009855 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-564487c87-zppvw"] Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.029401 4775 scope.go:117] "RemoveContainer" containerID="21c7d213dfbd2ef3e9ab592c212ff69020698f6ba49a9ebb52c79a43aa64f473" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.056665 4775 scope.go:117] "RemoveContainer" containerID="836e86acc5faf2a3e9656e8d8d2d493dd928f3335bd576fc6d293e8ae72cc25c" Oct 02 02:03:59 crc kubenswrapper[4775]: E1002 02:03:59.057185 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"836e86acc5faf2a3e9656e8d8d2d493dd928f3335bd576fc6d293e8ae72cc25c\": container with ID starting with 836e86acc5faf2a3e9656e8d8d2d493dd928f3335bd576fc6d293e8ae72cc25c not found: ID does not exist" containerID="836e86acc5faf2a3e9656e8d8d2d493dd928f3335bd576fc6d293e8ae72cc25c" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.057342 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"836e86acc5faf2a3e9656e8d8d2d493dd928f3335bd576fc6d293e8ae72cc25c"} err="failed to get container status \"836e86acc5faf2a3e9656e8d8d2d493dd928f3335bd576fc6d293e8ae72cc25c\": rpc error: code = NotFound desc = could not find container \"836e86acc5faf2a3e9656e8d8d2d493dd928f3335bd576fc6d293e8ae72cc25c\": container with ID starting with 836e86acc5faf2a3e9656e8d8d2d493dd928f3335bd576fc6d293e8ae72cc25c not found: ID does not exist" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.057444 4775 scope.go:117] "RemoveContainer" containerID="21c7d213dfbd2ef3e9ab592c212ff69020698f6ba49a9ebb52c79a43aa64f473" Oct 02 02:03:59 crc kubenswrapper[4775]: E1002 02:03:59.058006 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21c7d213dfbd2ef3e9ab592c212ff69020698f6ba49a9ebb52c79a43aa64f473\": container with ID starting with 21c7d213dfbd2ef3e9ab592c212ff69020698f6ba49a9ebb52c79a43aa64f473 not found: ID does not exist" containerID="21c7d213dfbd2ef3e9ab592c212ff69020698f6ba49a9ebb52c79a43aa64f473" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.058042 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21c7d213dfbd2ef3e9ab592c212ff69020698f6ba49a9ebb52c79a43aa64f473"} err="failed to get container status \"21c7d213dfbd2ef3e9ab592c212ff69020698f6ba49a9ebb52c79a43aa64f473\": rpc error: code = NotFound desc = could not find container \"21c7d213dfbd2ef3e9ab592c212ff69020698f6ba49a9ebb52c79a43aa64f473\": container with ID starting with 21c7d213dfbd2ef3e9ab592c212ff69020698f6ba49a9ebb52c79a43aa64f473 not found: ID does not exist" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.058066 4775 scope.go:117] "RemoveContainer" containerID="7ad48d14db57c9f5d0c7e7d3fcaf64adb99c760e739a396090a74eb51c2a9919" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.085200 4775 scope.go:117] "RemoveContainer" containerID="7ad48d14db57c9f5d0c7e7d3fcaf64adb99c760e739a396090a74eb51c2a9919" Oct 02 02:03:59 crc kubenswrapper[4775]: E1002 02:03:59.085702 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ad48d14db57c9f5d0c7e7d3fcaf64adb99c760e739a396090a74eb51c2a9919\": container with ID starting with 7ad48d14db57c9f5d0c7e7d3fcaf64adb99c760e739a396090a74eb51c2a9919 not found: ID does not exist" containerID="7ad48d14db57c9f5d0c7e7d3fcaf64adb99c760e739a396090a74eb51c2a9919" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.085830 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ad48d14db57c9f5d0c7e7d3fcaf64adb99c760e739a396090a74eb51c2a9919"} err="failed to get container status \"7ad48d14db57c9f5d0c7e7d3fcaf64adb99c760e739a396090a74eb51c2a9919\": rpc error: code = NotFound desc = could not find container \"7ad48d14db57c9f5d0c7e7d3fcaf64adb99c760e739a396090a74eb51c2a9919\": container with ID starting with 7ad48d14db57c9f5d0c7e7d3fcaf64adb99c760e739a396090a74eb51c2a9919 not found: ID does not exist" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.501135 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-585fd6468c-8gx5n" podUID="446bd539-6bf4-42b1-ac19-40c8c80b2d45" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.169:8080/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.502933 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-585fd6468c-8gx5n" podUID="446bd539-6bf4-42b1-ac19-40c8c80b2d45" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.169:8080/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.574615 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.692034 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-combined-ca-bundle\") pod \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\" (UID: \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\") " Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.692155 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp9x5\" (UniqueName: \"kubernetes.io/projected/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-kube-api-access-tp9x5\") pod \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\" (UID: \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\") " Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.692232 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-config-data\") pod \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\" (UID: \"a64f8ba3-65b9-489f-ace1-757f3ecce6b4\") " Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.710216 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-kube-api-access-tp9x5" (OuterVolumeSpecName: "kube-api-access-tp9x5") pod "a64f8ba3-65b9-489f-ace1-757f3ecce6b4" (UID: "a64f8ba3-65b9-489f-ace1-757f3ecce6b4"). InnerVolumeSpecName "kube-api-access-tp9x5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.721134 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a64f8ba3-65b9-489f-ace1-757f3ecce6b4" (UID: "a64f8ba3-65b9-489f-ace1-757f3ecce6b4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.724210 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-config-data" (OuterVolumeSpecName: "config-data") pod "a64f8ba3-65b9-489f-ace1-757f3ecce6b4" (UID: "a64f8ba3-65b9-489f-ace1-757f3ecce6b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.783516 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="002eb223-af34-4c80-9570-894add9b0e3c" path="/var/lib/kubelet/pods/002eb223-af34-4c80-9570-894add9b0e3c/volumes" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.796072 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.796117 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp9x5\" (UniqueName: \"kubernetes.io/projected/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-kube-api-access-tp9x5\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.796129 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a64f8ba3-65b9-489f-ace1-757f3ecce6b4-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.798178 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e" path="/var/lib/kubelet/pods/1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e/volumes" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.799558 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6638baa9-b367-40d1-8111-673dca6434ad" path="/var/lib/kubelet/pods/6638baa9-b367-40d1-8111-673dca6434ad/volumes" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.801247 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6efbe4b3-953d-4927-95c6-9b92708f51eb" path="/var/lib/kubelet/pods/6efbe4b3-953d-4927-95c6-9b92708f51eb/volumes" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.803252 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7da01b78-a318-448e-8a4c-7ad3d9e833c0" path="/var/lib/kubelet/pods/7da01b78-a318-448e-8a4c-7ad3d9e833c0/volumes" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.804445 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" path="/var/lib/kubelet/pods/fb0a7ddd-1558-4ad8-a8f4-e140306d19f9/volumes" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.925285 4775 generic.go:334] "Generic (PLEG): container finished" podID="a64f8ba3-65b9-489f-ace1-757f3ecce6b4" containerID="cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa" exitCode=0 Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.925384 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a64f8ba3-65b9-489f-ace1-757f3ecce6b4","Type":"ContainerDied","Data":"cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa"} Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.925422 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"a64f8ba3-65b9-489f-ace1-757f3ecce6b4","Type":"ContainerDied","Data":"e3bfc96e50f98ba49539fa2c9cb90f53ad1f2a6bf089d7ba9643db31fd98274d"} Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.925450 4775 scope.go:117] "RemoveContainer" containerID="cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.925611 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.960339 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.966700 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.968165 4775 scope.go:117] "RemoveContainer" containerID="cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa" Oct 02 02:03:59 crc kubenswrapper[4775]: E1002 02:03:59.968635 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa\": container with ID starting with cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa not found: ID does not exist" containerID="cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa" Oct 02 02:03:59 crc kubenswrapper[4775]: I1002 02:03:59.968690 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa"} err="failed to get container status \"cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa\": rpc error: code = NotFound desc = could not find container \"cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa\": container with ID starting with cd6503264221202318d9726fca6b4f7e17cf352313434793a3a5d993d5d2c7fa not found: ID does not exist" Oct 02 02:04:00 crc kubenswrapper[4775]: E1002 02:04:00.190277 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:00 crc kubenswrapper[4775]: E1002 02:04:00.190859 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:00 crc kubenswrapper[4775]: E1002 02:04:00.191206 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:00 crc kubenswrapper[4775]: E1002 02:04:00.191241 4775 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovsdb-server" Oct 02 02:04:00 crc kubenswrapper[4775]: E1002 02:04:00.192079 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:00 crc kubenswrapper[4775]: E1002 02:04:00.193778 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:00 crc kubenswrapper[4775]: E1002 02:04:00.196011 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:00 crc kubenswrapper[4775]: E1002 02:04:00.196162 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovs-vswitchd" Oct 02 02:04:01 crc kubenswrapper[4775]: I1002 02:04:01.778347 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a64f8ba3-65b9-489f-ace1-757f3ecce6b4" path="/var/lib/kubelet/pods/a64f8ba3-65b9-489f-ace1-757f3ecce6b4/volumes" Oct 02 02:04:05 crc kubenswrapper[4775]: E1002 02:04:05.190009 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:05 crc kubenswrapper[4775]: E1002 02:04:05.190792 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:05 crc kubenswrapper[4775]: E1002 02:04:05.191925 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:05 crc kubenswrapper[4775]: E1002 02:04:05.192217 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:05 crc kubenswrapper[4775]: E1002 02:04:05.192295 4775 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovsdb-server" Oct 02 02:04:05 crc kubenswrapper[4775]: E1002 02:04:05.194691 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:05 crc kubenswrapper[4775]: E1002 02:04:05.197403 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:05 crc kubenswrapper[4775]: E1002 02:04:05.197464 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovs-vswitchd" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.035891 4775 generic.go:334] "Generic (PLEG): container finished" podID="2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" containerID="8c9ca1dbcfc1426fde26f45e447e342f40b762f7d726a5c8194371f054ef7feb" exitCode=0 Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.036004 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-546f7b9c49-h6rcs" event={"ID":"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab","Type":"ContainerDied","Data":"8c9ca1dbcfc1426fde26f45e447e342f40b762f7d726a5c8194371f054ef7feb"} Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.411808 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.524804 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-internal-tls-certs\") pod \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.524886 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-combined-ca-bundle\") pod \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.524940 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-httpd-config\") pod \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.524983 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-config\") pod \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.525009 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-public-tls-certs\") pod \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.525059 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-ovndb-tls-certs\") pod \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.525122 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm5nn\" (UniqueName: \"kubernetes.io/projected/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-kube-api-access-lm5nn\") pod \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\" (UID: \"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab\") " Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.533055 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-kube-api-access-lm5nn" (OuterVolumeSpecName: "kube-api-access-lm5nn") pod "2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" (UID: "2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab"). InnerVolumeSpecName "kube-api-access-lm5nn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.545523 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" (UID: "2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.593933 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" (UID: "2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.597666 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" (UID: "2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.601706 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" (UID: "2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.604184 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-config" (OuterVolumeSpecName: "config") pod "2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" (UID: "2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.624739 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" (UID: "2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.626870 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.626928 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.627055 4775 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.627092 4775 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.627118 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm5nn\" (UniqueName: \"kubernetes.io/projected/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-kube-api-access-lm5nn\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.627142 4775 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:07 crc kubenswrapper[4775]: I1002 02:04:07.627165 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:08 crc kubenswrapper[4775]: I1002 02:04:08.047098 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-546f7b9c49-h6rcs" event={"ID":"2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab","Type":"ContainerDied","Data":"0e053b46f04d55b9e7591b085c54f9c8a069336ff3dd6b39e07e7d39e1ed955f"} Oct 02 02:04:08 crc kubenswrapper[4775]: I1002 02:04:08.047177 4775 scope.go:117] "RemoveContainer" containerID="85353a3b8a9a2c508bfbb1d8f15bdea36723aba6e9d1ddda8f826588f795b512" Oct 02 02:04:08 crc kubenswrapper[4775]: I1002 02:04:08.047183 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-546f7b9c49-h6rcs" Oct 02 02:04:08 crc kubenswrapper[4775]: I1002 02:04:08.073558 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-546f7b9c49-h6rcs"] Oct 02 02:04:08 crc kubenswrapper[4775]: I1002 02:04:08.079308 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-546f7b9c49-h6rcs"] Oct 02 02:04:08 crc kubenswrapper[4775]: I1002 02:04:08.083093 4775 scope.go:117] "RemoveContainer" containerID="8c9ca1dbcfc1426fde26f45e447e342f40b762f7d726a5c8194371f054ef7feb" Oct 02 02:04:09 crc kubenswrapper[4775]: I1002 02:04:09.781573 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" path="/var/lib/kubelet/pods/2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab/volumes" Oct 02 02:04:10 crc kubenswrapper[4775]: E1002 02:04:10.190033 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:10 crc kubenswrapper[4775]: E1002 02:04:10.190836 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:10 crc kubenswrapper[4775]: E1002 02:04:10.191525 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:10 crc kubenswrapper[4775]: E1002 02:04:10.191590 4775 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovsdb-server" Oct 02 02:04:10 crc kubenswrapper[4775]: E1002 02:04:10.192893 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:10 crc kubenswrapper[4775]: E1002 02:04:10.195265 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:10 crc kubenswrapper[4775]: E1002 02:04:10.197533 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:10 crc kubenswrapper[4775]: E1002 02:04:10.197591 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovs-vswitchd" Oct 02 02:04:15 crc kubenswrapper[4775]: E1002 02:04:15.189628 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:15 crc kubenswrapper[4775]: E1002 02:04:15.191476 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:15 crc kubenswrapper[4775]: E1002 02:04:15.191823 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:15 crc kubenswrapper[4775]: E1002 02:04:15.191888 4775 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovsdb-server" Oct 02 02:04:15 crc kubenswrapper[4775]: E1002 02:04:15.192182 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:15 crc kubenswrapper[4775]: E1002 02:04:15.194075 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:15 crc kubenswrapper[4775]: E1002 02:04:15.196279 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:15 crc kubenswrapper[4775]: E1002 02:04:15.196383 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovs-vswitchd" Oct 02 02:04:20 crc kubenswrapper[4775]: E1002 02:04:20.190073 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:20 crc kubenswrapper[4775]: E1002 02:04:20.191333 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:20 crc kubenswrapper[4775]: E1002 02:04:20.191769 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:20 crc kubenswrapper[4775]: E1002 02:04:20.191932 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Oct 02 02:04:20 crc kubenswrapper[4775]: E1002 02:04:20.192048 4775 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovsdb-server" Oct 02 02:04:20 crc kubenswrapper[4775]: E1002 02:04:20.193746 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:20 crc kubenswrapper[4775]: E1002 02:04:20.196052 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Oct 02 02:04:20 crc kubenswrapper[4775]: E1002 02:04:20.196176 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-xwnrv" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovs-vswitchd" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.189501 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xwnrv_2a54aa55-e3e2-4844-b175-e68442220b92/ovs-vswitchd/0.log" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.190723 4775 generic.go:334] "Generic (PLEG): container finished" podID="2a54aa55-e3e2-4844-b175-e68442220b92" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" exitCode=137 Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.190776 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xwnrv" event={"ID":"2a54aa55-e3e2-4844-b175-e68442220b92","Type":"ContainerDied","Data":"2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303"} Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.194171 4775 generic.go:334] "Generic (PLEG): container finished" podID="24e08fdf-b02d-4d6b-b451-8aa967ff04f6" containerID="86650058d32beec95536f3008ef303ebaebff3385b901bf98a315478321c1bce" exitCode=137 Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.194225 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"24e08fdf-b02d-4d6b-b451-8aa967ff04f6","Type":"ContainerDied","Data":"86650058d32beec95536f3008ef303ebaebff3385b901bf98a315478321c1bce"} Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.206147 4775 generic.go:334] "Generic (PLEG): container finished" podID="f8f38763-2981-4941-944d-b3fdc598be04" containerID="9a1e1ae6edc64baff01fb577c5fdff389a505836706e4e8e364c628542190596" exitCode=137 Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.206205 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"9a1e1ae6edc64baff01fb577c5fdff389a505836706e4e8e364c628542190596"} Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.324610 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xwnrv_2a54aa55-e3e2-4844-b175-e68442220b92/ovs-vswitchd/0.log" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.325973 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.345107 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.385562 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.447942 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-config-data-custom\") pod \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.447989 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-scripts\") pod \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.448017 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-run\") pod \"2a54aa55-e3e2-4844-b175-e68442220b92\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.448040 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-combined-ca-bundle\") pod \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.448063 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xs57f\" (UniqueName: \"kubernetes.io/projected/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-kube-api-access-xs57f\") pod \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.448084 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cb7bf\" (UniqueName: \"kubernetes.io/projected/2a54aa55-e3e2-4844-b175-e68442220b92-kube-api-access-cb7bf\") pod \"2a54aa55-e3e2-4844-b175-e68442220b92\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.448133 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-etc-ovs\") pod \"2a54aa55-e3e2-4844-b175-e68442220b92\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.448151 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-config-data\") pod \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.448165 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-etc-machine-id\") pod \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\" (UID: \"24e08fdf-b02d-4d6b-b451-8aa967ff04f6\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.448181 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a54aa55-e3e2-4844-b175-e68442220b92-scripts\") pod \"2a54aa55-e3e2-4844-b175-e68442220b92\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.448216 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-lib\") pod \"2a54aa55-e3e2-4844-b175-e68442220b92\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.448268 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-log\") pod \"2a54aa55-e3e2-4844-b175-e68442220b92\" (UID: \"2a54aa55-e3e2-4844-b175-e68442220b92\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.448379 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-log" (OuterVolumeSpecName: "var-log") pod "2a54aa55-e3e2-4844-b175-e68442220b92" (UID: "2a54aa55-e3e2-4844-b175-e68442220b92"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.449155 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-run" (OuterVolumeSpecName: "var-run") pod "2a54aa55-e3e2-4844-b175-e68442220b92" (UID: "2a54aa55-e3e2-4844-b175-e68442220b92"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.449559 4775 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-log\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.449589 4775 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.449728 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "24e08fdf-b02d-4d6b-b451-8aa967ff04f6" (UID: "24e08fdf-b02d-4d6b-b451-8aa967ff04f6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.449816 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "2a54aa55-e3e2-4844-b175-e68442220b92" (UID: "2a54aa55-e3e2-4844-b175-e68442220b92"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.450711 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a54aa55-e3e2-4844-b175-e68442220b92-scripts" (OuterVolumeSpecName: "scripts") pod "2a54aa55-e3e2-4844-b175-e68442220b92" (UID: "2a54aa55-e3e2-4844-b175-e68442220b92"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.450745 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-lib" (OuterVolumeSpecName: "var-lib") pod "2a54aa55-e3e2-4844-b175-e68442220b92" (UID: "2a54aa55-e3e2-4844-b175-e68442220b92"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.453702 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-scripts" (OuterVolumeSpecName: "scripts") pod "24e08fdf-b02d-4d6b-b451-8aa967ff04f6" (UID: "24e08fdf-b02d-4d6b-b451-8aa967ff04f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.455113 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "24e08fdf-b02d-4d6b-b451-8aa967ff04f6" (UID: "24e08fdf-b02d-4d6b-b451-8aa967ff04f6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.456815 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a54aa55-e3e2-4844-b175-e68442220b92-kube-api-access-cb7bf" (OuterVolumeSpecName: "kube-api-access-cb7bf") pod "2a54aa55-e3e2-4844-b175-e68442220b92" (UID: "2a54aa55-e3e2-4844-b175-e68442220b92"). InnerVolumeSpecName "kube-api-access-cb7bf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.457085 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-kube-api-access-xs57f" (OuterVolumeSpecName: "kube-api-access-xs57f") pod "24e08fdf-b02d-4d6b-b451-8aa967ff04f6" (UID: "24e08fdf-b02d-4d6b-b451-8aa967ff04f6"). InnerVolumeSpecName "kube-api-access-xs57f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.483352 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "24e08fdf-b02d-4d6b-b451-8aa967ff04f6" (UID: "24e08fdf-b02d-4d6b-b451-8aa967ff04f6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.525524 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-config-data" (OuterVolumeSpecName: "config-data") pod "24e08fdf-b02d-4d6b-b451-8aa967ff04f6" (UID: "24e08fdf-b02d-4d6b-b451-8aa967ff04f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.550648 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f8f38763-2981-4941-944d-b3fdc598be04-lock\") pod \"f8f38763-2981-4941-944d-b3fdc598be04\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.550703 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27k7d\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-kube-api-access-27k7d\") pod \"f8f38763-2981-4941-944d-b3fdc598be04\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.550794 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift\") pod \"f8f38763-2981-4941-944d-b3fdc598be04\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.550811 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"f8f38763-2981-4941-944d-b3fdc598be04\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.550840 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f8f38763-2981-4941-944d-b3fdc598be04-cache\") pod \"f8f38763-2981-4941-944d-b3fdc598be04\" (UID: \"f8f38763-2981-4941-944d-b3fdc598be04\") " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.551132 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.551148 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.551157 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.551165 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xs57f\" (UniqueName: \"kubernetes.io/projected/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-kube-api-access-xs57f\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.551175 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cb7bf\" (UniqueName: \"kubernetes.io/projected/2a54aa55-e3e2-4844-b175-e68442220b92-kube-api-access-cb7bf\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.551183 4775 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-etc-ovs\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.551193 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.551201 4775 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/24e08fdf-b02d-4d6b-b451-8aa967ff04f6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.551209 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a54aa55-e3e2-4844-b175-e68442220b92-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.551216 4775 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2a54aa55-e3e2-4844-b175-e68442220b92-var-lib\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.551678 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8f38763-2981-4941-944d-b3fdc598be04-cache" (OuterVolumeSpecName: "cache") pod "f8f38763-2981-4941-944d-b3fdc598be04" (UID: "f8f38763-2981-4941-944d-b3fdc598be04"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.552053 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8f38763-2981-4941-944d-b3fdc598be04-lock" (OuterVolumeSpecName: "lock") pod "f8f38763-2981-4941-944d-b3fdc598be04" (UID: "f8f38763-2981-4941-944d-b3fdc598be04"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.558203 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-kube-api-access-27k7d" (OuterVolumeSpecName: "kube-api-access-27k7d") pod "f8f38763-2981-4941-944d-b3fdc598be04" (UID: "f8f38763-2981-4941-944d-b3fdc598be04"). InnerVolumeSpecName "kube-api-access-27k7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.560775 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "swift") pod "f8f38763-2981-4941-944d-b3fdc598be04" (UID: "f8f38763-2981-4941-944d-b3fdc598be04"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.561517 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "f8f38763-2981-4941-944d-b3fdc598be04" (UID: "f8f38763-2981-4941-944d-b3fdc598be04"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.652680 4775 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.652756 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.652772 4775 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/f8f38763-2981-4941-944d-b3fdc598be04-cache\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.652783 4775 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/f8f38763-2981-4941-944d-b3fdc598be04-lock\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.652796 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27k7d\" (UniqueName: \"kubernetes.io/projected/f8f38763-2981-4941-944d-b3fdc598be04-kube-api-access-27k7d\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.674277 4775 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 02 02:04:21 crc kubenswrapper[4775]: I1002 02:04:21.754207 4775 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.226682 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"24e08fdf-b02d-4d6b-b451-8aa967ff04f6","Type":"ContainerDied","Data":"4c6de7d12a2ce34841501a709102fa7d8c0d30c95ba74db6eead2bb64802a87f"} Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.226772 4775 scope.go:117] "RemoveContainer" containerID="ecfa4d3c77b7cfaba81f2e5e014ad3a08c3704a721e8b2bcc3173012b42a0aff" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.226701 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.238347 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"f8f38763-2981-4941-944d-b3fdc598be04","Type":"ContainerDied","Data":"50c7ff7d8ff529121dc875f3de87c4eaed5a9d4222abbe043b5147a5e835b4b1"} Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.238897 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.254728 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xwnrv_2a54aa55-e3e2-4844-b175-e68442220b92/ovs-vswitchd/0.log" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.256310 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xwnrv" event={"ID":"2a54aa55-e3e2-4844-b175-e68442220b92","Type":"ContainerDied","Data":"82de9cf13ef9c6d4a716e8c4172b37431c0748f8bd4e24b9c52ea8a28d17c6be"} Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.256545 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xwnrv" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.261068 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.266708 4775 scope.go:117] "RemoveContainer" containerID="86650058d32beec95536f3008ef303ebaebff3385b901bf98a315478321c1bce" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.269003 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.293551 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.301020 4775 scope.go:117] "RemoveContainer" containerID="9a1e1ae6edc64baff01fb577c5fdff389a505836706e4e8e364c628542190596" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.303140 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.309502 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-xwnrv"] Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.315546 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-xwnrv"] Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.324052 4775 scope.go:117] "RemoveContainer" containerID="e43595485c4ef19d17323cd2f15cc8c6c01cfd72d63f9027002a311666221ea2" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.342356 4775 scope.go:117] "RemoveContainer" containerID="a9fa0f88f8c28ceafea5f7d19fe1fa35b3b7c2e9511b51b17451d5c799b2adc6" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.394164 4775 scope.go:117] "RemoveContainer" containerID="3b361b7fc7887886b0530a00fe1e6d2dd5a5e58cc96a870064fbcbeb95903b7b" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.418435 4775 scope.go:117] "RemoveContainer" containerID="6772443aac50d3396138519e7b7ffbb69baa9434eb913eb475d376b8df66e3ec" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.448902 4775 scope.go:117] "RemoveContainer" containerID="ad68cb16ec4f02e54fb4a47b88804b69e1f0edd04ff7e5ec3f86cdfe3bdfc45d" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.473407 4775 scope.go:117] "RemoveContainer" containerID="b14587fcd25cde3d5786c84529e1eac9820c2a094fc6d465d49b957d0ca1f117" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.504473 4775 scope.go:117] "RemoveContainer" containerID="783ca618d590525d6b1848438fe648b59eddb9e00181a39a7cabe0cd1d417d06" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.537181 4775 scope.go:117] "RemoveContainer" containerID="05475c086b8f31273f54dfe829e118fed85817c9d16d5585b851871fc39726bb" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.573760 4775 scope.go:117] "RemoveContainer" containerID="24cb4406a78e877a5465e016c64de5577c4354d83b07c019e2d33609646a1f9a" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.606759 4775 scope.go:117] "RemoveContainer" containerID="a994b8ddd7e1ce4fa5946caa21707a15ffb193c4b308890c750b2ed3745c88d6" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.630169 4775 scope.go:117] "RemoveContainer" containerID="35e88601383e15eea377e15e1d17c552b0acdc4bf1bb0f64c98b395f45c898cf" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.652730 4775 scope.go:117] "RemoveContainer" containerID="1ad3d27e30ff5a6104b653354bc8348baa64801d396c0e15ef454a96004528cf" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.680992 4775 scope.go:117] "RemoveContainer" containerID="93013dec09b53cd82ac1d950f89f6d21acde0be93880dc71f932f03315d26749" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.707206 4775 scope.go:117] "RemoveContainer" containerID="e1f6027f6175a22e8ea34c7faa9698a5ad269b7fc4b557121c200890ba72ee33" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.739801 4775 scope.go:117] "RemoveContainer" containerID="2a8351a253609f07866ab561162286029d54b19b0a2547d63508fd926e222303" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.769660 4775 scope.go:117] "RemoveContainer" containerID="19ef00568f02cb4f983db56c1c49d5800f16a2e3435ed494f7163a344d172f0c" Oct 02 02:04:22 crc kubenswrapper[4775]: I1002 02:04:22.797934 4775 scope.go:117] "RemoveContainer" containerID="dff1a870836ccb85517741d710815364b5071bb66fd4210ba110dfd834a3ced9" Oct 02 02:04:23 crc kubenswrapper[4775]: I1002 02:04:23.784771 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24e08fdf-b02d-4d6b-b451-8aa967ff04f6" path="/var/lib/kubelet/pods/24e08fdf-b02d-4d6b-b451-8aa967ff04f6/volumes" Oct 02 02:04:23 crc kubenswrapper[4775]: I1002 02:04:23.786560 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" path="/var/lib/kubelet/pods/2a54aa55-e3e2-4844-b175-e68442220b92/volumes" Oct 02 02:04:23 crc kubenswrapper[4775]: I1002 02:04:23.788208 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8f38763-2981-4941-944d-b3fdc598be04" path="/var/lib/kubelet/pods/f8f38763-2981-4941-944d-b3fdc598be04/volumes" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.853895 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rtxmd"] Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855328 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="002eb223-af34-4c80-9570-894add9b0e3c" containerName="mysql-bootstrap" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855359 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="002eb223-af34-4c80-9570-894add9b0e3c" containerName="mysql-bootstrap" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855390 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-expirer" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855404 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-expirer" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855423 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" containerName="neutron-api" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855437 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" containerName="neutron-api" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855461 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-updater" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855473 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-updater" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855490 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-auditor" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855502 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-auditor" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855527 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6638baa9-b367-40d1-8111-673dca6434ad" containerName="setup-container" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855539 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6638baa9-b367-40d1-8111-673dca6434ad" containerName="setup-container" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855553 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-reaper" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855565 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-reaper" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855584 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1364df9e-151f-40d0-9a5a-606a54021ef7" containerName="init" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855596 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1364df9e-151f-40d0-9a5a-606a54021ef7" containerName="init" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855610 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6638baa9-b367-40d1-8111-673dca6434ad" containerName="rabbitmq" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855621 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6638baa9-b367-40d1-8111-673dca6434ad" containerName="rabbitmq" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855677 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7685ff7-607b-416c-9382-2c00edecaa54" containerName="openstack-network-exporter" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855693 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7685ff7-607b-416c-9382-2c00edecaa54" containerName="openstack-network-exporter" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855715 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6efbe4b3-953d-4927-95c6-9b92708f51eb" containerName="ovn-controller" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855729 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6efbe4b3-953d-4927-95c6-9b92708f51eb" containerName="ovn-controller" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855748 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovs-vswitchd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855764 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovs-vswitchd" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855781 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="163a6d17-be4c-47fc-8835-e218e8a803f1" containerName="barbican-worker-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855801 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="163a6d17-be4c-47fc-8835-e218e8a803f1" containerName="barbican-worker-log" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855824 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f286b47-509d-479d-bad5-0bbec930558a" containerName="glance-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855839 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f286b47-509d-479d-bad5-0bbec930558a" containerName="glance-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855869 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46453bb2-6af0-4dd4-83ba-b14bf3f96310" containerName="nova-cell1-conductor-conductor" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855884 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="46453bb2-6af0-4dd4-83ba-b14bf3f96310" containerName="nova-cell1-conductor-conductor" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855908 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" containerName="openstack-network-exporter" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855925 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" containerName="openstack-network-exporter" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.855945 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="163a6d17-be4c-47fc-8835-e218e8a803f1" containerName="barbican-worker" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.855988 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="163a6d17-be4c-47fc-8835-e218e8a803f1" containerName="barbican-worker" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856011 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" containerName="glance-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856023 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" containerName="glance-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856041 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24e08fdf-b02d-4d6b-b451-8aa967ff04f6" containerName="probe" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856052 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="24e08fdf-b02d-4d6b-b451-8aa967ff04f6" containerName="probe" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856064 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af16c5f4-fcd0-4450-a8b7-789ff104e477" containerName="barbican-keystone-listener" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856076 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="af16c5f4-fcd0-4450-a8b7-789ff104e477" containerName="barbican-keystone-listener" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856091 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446bd539-6bf4-42b1-ac19-40c8c80b2d45" containerName="proxy-server" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856103 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="446bd539-6bf4-42b1-ac19-40c8c80b2d45" containerName="proxy-server" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856116 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="proxy-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856128 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="proxy-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856144 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56813183-3dfc-430c-aaac-bda27d91340b" containerName="placement-api" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856156 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="56813183-3dfc-430c-aaac-bda27d91340b" containerName="placement-api" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856176 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f286b47-509d-479d-bad5-0bbec930558a" containerName="glance-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856187 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f286b47-509d-479d-bad5-0bbec930558a" containerName="glance-log" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856209 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-auditor" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856221 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-auditor" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856234 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7da01b78-a318-448e-8a4c-7ad3d9e833c0" containerName="ovn-northd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856247 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7da01b78-a318-448e-8a4c-7ad3d9e833c0" containerName="ovn-northd" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856267 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9813c97-677a-42c2-b526-576a4c5f2968" containerName="openstack-network-exporter" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856279 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9813c97-677a-42c2-b526-576a4c5f2968" containerName="openstack-network-exporter" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856295 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e322679-b95d-4328-b03b-0be7f0e82ac3" containerName="nova-metadata-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856306 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e322679-b95d-4328-b03b-0be7f0e82ac3" containerName="nova-metadata-log" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856327 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-server" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856339 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-server" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856355 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e5e6844-45f6-4fd9-ac4f-3df2086164eb" containerName="mariadb-account-delete" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856366 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e5e6844-45f6-4fd9-ac4f-3df2086164eb" containerName="mariadb-account-delete" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856382 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="ceilometer-notification-agent" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856394 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="ceilometer-notification-agent" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856416 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24ef8940-eb43-459c-9cfe-854df57cfe74" containerName="kube-state-metrics" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856427 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="24ef8940-eb43-459c-9cfe-854df57cfe74" containerName="kube-state-metrics" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856441 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="rsync" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856455 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="rsync" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856469 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-auditor" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856481 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-auditor" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856494 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-replicator" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856506 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-replicator" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856525 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="406dd24f-d800-44be-b33d-019c532c2feb" containerName="memcached" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856537 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="406dd24f-d800-44be-b33d-019c532c2feb" containerName="memcached" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856548 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" containerName="neutron-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856560 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" containerName="neutron-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856580 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" containerName="nova-api-api" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856592 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" containerName="nova-api-api" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856614 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7da01b78-a318-448e-8a4c-7ad3d9e833c0" containerName="openstack-network-exporter" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856626 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7da01b78-a318-448e-8a4c-7ad3d9e833c0" containerName="openstack-network-exporter" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856640 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3adc5647-1145-4c44-9ecc-66d1d9a19023" containerName="galera" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856651 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3adc5647-1145-4c44-9ecc-66d1d9a19023" containerName="galera" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856667 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="swift-recon-cron" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856679 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="swift-recon-cron" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856696 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" containerName="nova-api-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856709 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" containerName="nova-api-log" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856730 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="002eb223-af34-4c80-9570-894add9b0e3c" containerName="galera" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856741 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="002eb223-af34-4c80-9570-894add9b0e3c" containerName="galera" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856761 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bca47d0-79b7-4992-867b-d5e05e02cdc8" containerName="cinder-api" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856773 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bca47d0-79b7-4992-867b-d5e05e02cdc8" containerName="cinder-api" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856792 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3732523-d2de-45c9-ac69-770197f413f3" containerName="mariadb-account-delete" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856803 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3732523-d2de-45c9-ac69-770197f413f3" containerName="mariadb-account-delete" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856818 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" containerName="rabbitmq" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856831 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" containerName="rabbitmq" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856846 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-updater" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856860 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-updater" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856879 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-replicator" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856891 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-replicator" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856907 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e" containerName="keystone-api" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856919 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e" containerName="keystone-api" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856933 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af16c5f4-fcd0-4450-a8b7-789ff104e477" containerName="barbican-keystone-listener-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.856945 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="af16c5f4-fcd0-4450-a8b7-789ff104e477" containerName="barbican-keystone-listener-log" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.856990 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c74e0a1d-125d-43f7-9ba3-cea70453fc01" containerName="nova-cell0-conductor-conductor" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857003 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c74e0a1d-125d-43f7-9ba3-cea70453fc01" containerName="nova-cell0-conductor-conductor" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857017 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c60336-6e25-4c68-98e6-52801f34578b" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857028 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c60336-6e25-4c68-98e6-52801f34578b" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857045 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a64f8ba3-65b9-489f-ace1-757f3ecce6b4" containerName="nova-scheduler-scheduler" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857057 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a64f8ba3-65b9-489f-ace1-757f3ecce6b4" containerName="nova-scheduler-scheduler" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857070 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6bb525-8c94-461f-84c8-9e466b789706" containerName="barbican-api-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857082 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6bb525-8c94-461f-84c8-9e466b789706" containerName="barbican-api-log" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857098 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cdea921f-d632-4463-a011-2e6f23f70546" containerName="mariadb-account-delete" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857114 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdea921f-d632-4463-a011-2e6f23f70546" containerName="mariadb-account-delete" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857134 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-server" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857149 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-server" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857168 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1364df9e-151f-40d0-9a5a-606a54021ef7" containerName="dnsmasq-dns" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857182 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1364df9e-151f-40d0-9a5a-606a54021ef7" containerName="dnsmasq-dns" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857213 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-server" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857229 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-server" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857252 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" containerName="glance-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857266 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" containerName="glance-log" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857282 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovsdb-server" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857297 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovsdb-server" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857318 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6bb525-8c94-461f-84c8-9e466b789706" containerName="barbican-api" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857333 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6bb525-8c94-461f-84c8-9e466b789706" containerName="barbican-api" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857353 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3adc5647-1145-4c44-9ecc-66d1d9a19023" containerName="mysql-bootstrap" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857367 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3adc5647-1145-4c44-9ecc-66d1d9a19023" containerName="mysql-bootstrap" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857396 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7685ff7-607b-416c-9382-2c00edecaa54" containerName="ovsdbserver-sb" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857412 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7685ff7-607b-416c-9382-2c00edecaa54" containerName="ovsdbserver-sb" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857435 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" containerName="setup-container" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857449 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" containerName="setup-container" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857476 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56813183-3dfc-430c-aaac-bda27d91340b" containerName="placement-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857492 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="56813183-3dfc-430c-aaac-bda27d91340b" containerName="placement-log" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857511 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="sg-core" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857526 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="sg-core" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857549 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" containerName="ovsdbserver-nb" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857564 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" containerName="ovsdbserver-nb" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857593 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="ceilometer-central-agent" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857614 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="ceilometer-central-agent" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857642 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bca47d0-79b7-4992-867b-d5e05e02cdc8" containerName="cinder-api-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857658 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bca47d0-79b7-4992-867b-d5e05e02cdc8" containerName="cinder-api-log" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857679 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="446bd539-6bf4-42b1-ac19-40c8c80b2d45" containerName="proxy-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857695 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="446bd539-6bf4-42b1-ac19-40c8c80b2d45" containerName="proxy-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857718 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovsdb-server-init" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857735 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovsdb-server-init" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857756 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-replicator" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857771 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-replicator" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857797 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24e08fdf-b02d-4d6b-b451-8aa967ff04f6" containerName="cinder-scheduler" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857813 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="24e08fdf-b02d-4d6b-b451-8aa967ff04f6" containerName="cinder-scheduler" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857845 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e322679-b95d-4328-b03b-0be7f0e82ac3" containerName="nova-metadata-metadata" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857861 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e322679-b95d-4328-b03b-0be7f0e82ac3" containerName="nova-metadata-metadata" Oct 02 02:04:34 crc kubenswrapper[4775]: E1002 02:04:34.857878 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a" containerName="mariadb-account-delete" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.857892 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a" containerName="mariadb-account-delete" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858228 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-replicator" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858257 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7da01b78-a318-448e-8a4c-7ad3d9e833c0" containerName="openstack-network-exporter" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858286 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" containerName="neutron-api" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858315 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="163a6d17-be4c-47fc-8835-e218e8a803f1" containerName="barbican-worker" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858338 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="24ef8940-eb43-459c-9cfe-854df57cfe74" containerName="kube-state-metrics" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858369 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f286b47-509d-479d-bad5-0bbec930558a" containerName="glance-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858386 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" containerName="ovsdbserver-nb" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858407 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5ce93aa-f0e3-48eb-b8a9-c9836edb3e92" containerName="openstack-network-exporter" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858434 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="56813183-3dfc-430c-aaac-bda27d91340b" containerName="placement-api" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858459 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="24e08fdf-b02d-4d6b-b451-8aa967ff04f6" containerName="cinder-scheduler" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858488 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="46453bb2-6af0-4dd4-83ba-b14bf3f96310" containerName="nova-cell1-conductor-conductor" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858518 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-server" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858543 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" containerName="nova-api-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858557 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" containerName="glance-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858569 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7da01b78-a318-448e-8a4c-7ad3d9e833c0" containerName="ovn-northd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858585 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="proxy-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858608 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e322679-b95d-4328-b03b-0be7f0e82ac3" containerName="nova-metadata-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858623 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e5e6844-45f6-4fd9-ac4f-3df2086164eb" containerName="mariadb-account-delete" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858640 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="rsync" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858652 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="cdea921f-d632-4463-a011-2e6f23f70546" containerName="mariadb-account-delete" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858670 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovsdb-server" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858685 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-auditor" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858704 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-auditor" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858721 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5c60336-6e25-4c68-98e6-52801f34578b" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858742 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="163a6d17-be4c-47fc-8835-e218e8a803f1" containerName="barbican-worker-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858757 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7685ff7-607b-416c-9382-2c00edecaa54" containerName="openstack-network-exporter" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858773 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="ceilometer-central-agent" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858796 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6efbe4b3-953d-4927-95c6-9b92708f51eb" containerName="ovn-controller" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858817 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-updater" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858830 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a6ac89d-a0f4-4b93-a9ff-6e81372f9c51" containerName="glance-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858841 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca2bcd71-30d6-4fa8-a2bf-b04f32d33b0a" containerName="mariadb-account-delete" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858862 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6638baa9-b367-40d1-8111-673dca6434ad" containerName="rabbitmq" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858873 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-expirer" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858888 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-updater" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858903 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1364df9e-151f-40d0-9a5a-606a54021ef7" containerName="dnsmasq-dns" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858921 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-replicator" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858933 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-auditor" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.858987 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="object-server" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859008 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a54aa55-e3e2-4844-b175-e68442220b92" containerName="ovs-vswitchd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859028 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="ceilometer-notification-agent" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859047 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d6bb525-8c94-461f-84c8-9e466b789706" containerName="barbican-api" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859066 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d1b18bb-c5b8-4ada-9eb5-1d59a27527ab" containerName="neutron-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859087 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="446bd539-6bf4-42b1-ac19-40c8c80b2d45" containerName="proxy-server" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859107 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bca47d0-79b7-4992-867b-d5e05e02cdc8" containerName="cinder-api-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859122 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3732523-d2de-45c9-ac69-770197f413f3" containerName="mariadb-account-delete" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859137 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c74e0a1d-125d-43f7-9ba3-cea70453fc01" containerName="nova-cell0-conductor-conductor" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859150 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="container-server" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859162 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="446bd539-6bf4-42b1-ac19-40c8c80b2d45" containerName="proxy-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859174 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e322679-b95d-4328-b03b-0be7f0e82ac3" containerName="nova-metadata-metadata" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859192 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2d45342-c4de-48db-8f6b-23f26103497b" containerName="sg-core" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859214 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="swift-recon-cron" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859230 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9813c97-677a-42c2-b526-576a4c5f2968" containerName="openstack-network-exporter" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859246 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-replicator" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859263 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e3fe0c6-6851-4416-9e5b-7a12fd8a6f2e" containerName="keystone-api" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859277 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fe5ed4c-2cc2-41eb-bce8-c102c87af7d0" containerName="nova-api-api" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859292 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3adc5647-1145-4c44-9ecc-66d1d9a19023" containerName="galera" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859315 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="406dd24f-d800-44be-b33d-019c532c2feb" containerName="memcached" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859327 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7685ff7-607b-416c-9382-2c00edecaa54" containerName="ovsdbserver-sb" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859344 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bca47d0-79b7-4992-867b-d5e05e02cdc8" containerName="cinder-api" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859357 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="24e08fdf-b02d-4d6b-b451-8aa967ff04f6" containerName="probe" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859374 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a64f8ba3-65b9-489f-ace1-757f3ecce6b4" containerName="nova-scheduler-scheduler" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859388 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb0a7ddd-1558-4ad8-a8f4-e140306d19f9" containerName="rabbitmq" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859401 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="af16c5f4-fcd0-4450-a8b7-789ff104e477" containerName="barbican-keystone-listener-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859418 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="002eb223-af34-4c80-9570-894add9b0e3c" containerName="galera" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859434 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8f38763-2981-4941-944d-b3fdc598be04" containerName="account-reaper" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859448 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d6bb525-8c94-461f-84c8-9e466b789706" containerName="barbican-api-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859463 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="56813183-3dfc-430c-aaac-bda27d91340b" containerName="placement-log" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859478 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="af16c5f4-fcd0-4450-a8b7-789ff104e477" containerName="barbican-keystone-listener" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.859492 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f286b47-509d-479d-bad5-0bbec930558a" containerName="glance-httpd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.865051 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.865887 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rtxmd"] Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.955341 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3414e6f7-99da-4d3f-870e-ebe575364a6a-catalog-content\") pod \"community-operators-rtxmd\" (UID: \"3414e6f7-99da-4d3f-870e-ebe575364a6a\") " pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.955578 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9ddb\" (UniqueName: \"kubernetes.io/projected/3414e6f7-99da-4d3f-870e-ebe575364a6a-kube-api-access-h9ddb\") pod \"community-operators-rtxmd\" (UID: \"3414e6f7-99da-4d3f-870e-ebe575364a6a\") " pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:34 crc kubenswrapper[4775]: I1002 02:04:34.955616 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3414e6f7-99da-4d3f-870e-ebe575364a6a-utilities\") pod \"community-operators-rtxmd\" (UID: \"3414e6f7-99da-4d3f-870e-ebe575364a6a\") " pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:35 crc kubenswrapper[4775]: I1002 02:04:35.056823 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3414e6f7-99da-4d3f-870e-ebe575364a6a-utilities\") pod \"community-operators-rtxmd\" (UID: \"3414e6f7-99da-4d3f-870e-ebe575364a6a\") " pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:35 crc kubenswrapper[4775]: I1002 02:04:35.056911 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9ddb\" (UniqueName: \"kubernetes.io/projected/3414e6f7-99da-4d3f-870e-ebe575364a6a-kube-api-access-h9ddb\") pod \"community-operators-rtxmd\" (UID: \"3414e6f7-99da-4d3f-870e-ebe575364a6a\") " pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:35 crc kubenswrapper[4775]: I1002 02:04:35.057372 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3414e6f7-99da-4d3f-870e-ebe575364a6a-catalog-content\") pod \"community-operators-rtxmd\" (UID: \"3414e6f7-99da-4d3f-870e-ebe575364a6a\") " pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:35 crc kubenswrapper[4775]: I1002 02:04:35.057664 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3414e6f7-99da-4d3f-870e-ebe575364a6a-utilities\") pod \"community-operators-rtxmd\" (UID: \"3414e6f7-99da-4d3f-870e-ebe575364a6a\") " pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:35 crc kubenswrapper[4775]: I1002 02:04:35.057823 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3414e6f7-99da-4d3f-870e-ebe575364a6a-catalog-content\") pod \"community-operators-rtxmd\" (UID: \"3414e6f7-99da-4d3f-870e-ebe575364a6a\") " pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:35 crc kubenswrapper[4775]: I1002 02:04:35.089451 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9ddb\" (UniqueName: \"kubernetes.io/projected/3414e6f7-99da-4d3f-870e-ebe575364a6a-kube-api-access-h9ddb\") pod \"community-operators-rtxmd\" (UID: \"3414e6f7-99da-4d3f-870e-ebe575364a6a\") " pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:35 crc kubenswrapper[4775]: I1002 02:04:35.227940 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:35 crc kubenswrapper[4775]: I1002 02:04:35.514160 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rtxmd"] Oct 02 02:04:35 crc kubenswrapper[4775]: W1002 02:04:35.516515 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3414e6f7_99da_4d3f_870e_ebe575364a6a.slice/crio-1705e34471e9e73e6163eddce95de307e463b52eb6a37290507ee8b9fe2d876d WatchSource:0}: Error finding container 1705e34471e9e73e6163eddce95de307e463b52eb6a37290507ee8b9fe2d876d: Status 404 returned error can't find the container with id 1705e34471e9e73e6163eddce95de307e463b52eb6a37290507ee8b9fe2d876d Oct 02 02:04:36 crc kubenswrapper[4775]: I1002 02:04:36.435138 4775 generic.go:334] "Generic (PLEG): container finished" podID="3414e6f7-99da-4d3f-870e-ebe575364a6a" containerID="5b4b4553a268cfe2ba5b841dbc7d7f9b46c200b626551096d4f8807108ff253e" exitCode=0 Oct 02 02:04:36 crc kubenswrapper[4775]: I1002 02:04:36.435249 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rtxmd" event={"ID":"3414e6f7-99da-4d3f-870e-ebe575364a6a","Type":"ContainerDied","Data":"5b4b4553a268cfe2ba5b841dbc7d7f9b46c200b626551096d4f8807108ff253e"} Oct 02 02:04:36 crc kubenswrapper[4775]: I1002 02:04:36.435487 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rtxmd" event={"ID":"3414e6f7-99da-4d3f-870e-ebe575364a6a","Type":"ContainerStarted","Data":"1705e34471e9e73e6163eddce95de307e463b52eb6a37290507ee8b9fe2d876d"} Oct 02 02:04:36 crc kubenswrapper[4775]: I1002 02:04:36.438421 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 02:04:38 crc kubenswrapper[4775]: I1002 02:04:38.460463 4775 generic.go:334] "Generic (PLEG): container finished" podID="3414e6f7-99da-4d3f-870e-ebe575364a6a" containerID="9f3bf924e7c9120dbb6524b0c297a365abf7f5a3e45a55598a198ce6b85db3f3" exitCode=0 Oct 02 02:04:38 crc kubenswrapper[4775]: I1002 02:04:38.460710 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rtxmd" event={"ID":"3414e6f7-99da-4d3f-870e-ebe575364a6a","Type":"ContainerDied","Data":"9f3bf924e7c9120dbb6524b0c297a365abf7f5a3e45a55598a198ce6b85db3f3"} Oct 02 02:04:40 crc kubenswrapper[4775]: I1002 02:04:40.500892 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rtxmd" event={"ID":"3414e6f7-99da-4d3f-870e-ebe575364a6a","Type":"ContainerStarted","Data":"49c8fe9392bad05284abe558b985bc551c7bd029e12d390cf231dfab8c392287"} Oct 02 02:04:40 crc kubenswrapper[4775]: I1002 02:04:40.535334 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rtxmd" podStartSLOduration=3.572215351 podStartE2EDuration="6.535304291s" podCreationTimestamp="2025-10-02 02:04:34 +0000 UTC" firstStartedPulling="2025-10-02 02:04:36.437819484 +0000 UTC m=+1413.604563554" lastFinishedPulling="2025-10-02 02:04:39.400908444 +0000 UTC m=+1416.567652494" observedRunningTime="2025-10-02 02:04:40.526484374 +0000 UTC m=+1417.693228424" watchObservedRunningTime="2025-10-02 02:04:40.535304291 +0000 UTC m=+1417.702048371" Oct 02 02:04:45 crc kubenswrapper[4775]: I1002 02:04:45.228204 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:45 crc kubenswrapper[4775]: I1002 02:04:45.228727 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:45 crc kubenswrapper[4775]: I1002 02:04:45.299347 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:45 crc kubenswrapper[4775]: I1002 02:04:45.607668 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:45 crc kubenswrapper[4775]: I1002 02:04:45.666173 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rtxmd"] Oct 02 02:04:47 crc kubenswrapper[4775]: I1002 02:04:47.581595 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rtxmd" podUID="3414e6f7-99da-4d3f-870e-ebe575364a6a" containerName="registry-server" containerID="cri-o://49c8fe9392bad05284abe558b985bc551c7bd029e12d390cf231dfab8c392287" gracePeriod=2 Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.053376 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.171669 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3414e6f7-99da-4d3f-870e-ebe575364a6a-catalog-content\") pod \"3414e6f7-99da-4d3f-870e-ebe575364a6a\" (UID: \"3414e6f7-99da-4d3f-870e-ebe575364a6a\") " Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.171838 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9ddb\" (UniqueName: \"kubernetes.io/projected/3414e6f7-99da-4d3f-870e-ebe575364a6a-kube-api-access-h9ddb\") pod \"3414e6f7-99da-4d3f-870e-ebe575364a6a\" (UID: \"3414e6f7-99da-4d3f-870e-ebe575364a6a\") " Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.173056 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3414e6f7-99da-4d3f-870e-ebe575364a6a-utilities\") pod \"3414e6f7-99da-4d3f-870e-ebe575364a6a\" (UID: \"3414e6f7-99da-4d3f-870e-ebe575364a6a\") " Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.174146 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3414e6f7-99da-4d3f-870e-ebe575364a6a-utilities" (OuterVolumeSpecName: "utilities") pod "3414e6f7-99da-4d3f-870e-ebe575364a6a" (UID: "3414e6f7-99da-4d3f-870e-ebe575364a6a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.180232 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3414e6f7-99da-4d3f-870e-ebe575364a6a-kube-api-access-h9ddb" (OuterVolumeSpecName: "kube-api-access-h9ddb") pod "3414e6f7-99da-4d3f-870e-ebe575364a6a" (UID: "3414e6f7-99da-4d3f-870e-ebe575364a6a"). InnerVolumeSpecName "kube-api-access-h9ddb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.231129 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3414e6f7-99da-4d3f-870e-ebe575364a6a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3414e6f7-99da-4d3f-870e-ebe575364a6a" (UID: "3414e6f7-99da-4d3f-870e-ebe575364a6a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.274189 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9ddb\" (UniqueName: \"kubernetes.io/projected/3414e6f7-99da-4d3f-870e-ebe575364a6a-kube-api-access-h9ddb\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.274241 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3414e6f7-99da-4d3f-870e-ebe575364a6a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.274278 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3414e6f7-99da-4d3f-870e-ebe575364a6a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.596049 4775 generic.go:334] "Generic (PLEG): container finished" podID="3414e6f7-99da-4d3f-870e-ebe575364a6a" containerID="49c8fe9392bad05284abe558b985bc551c7bd029e12d390cf231dfab8c392287" exitCode=0 Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.596124 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rtxmd" event={"ID":"3414e6f7-99da-4d3f-870e-ebe575364a6a","Type":"ContainerDied","Data":"49c8fe9392bad05284abe558b985bc551c7bd029e12d390cf231dfab8c392287"} Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.596158 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rtxmd" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.596186 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rtxmd" event={"ID":"3414e6f7-99da-4d3f-870e-ebe575364a6a","Type":"ContainerDied","Data":"1705e34471e9e73e6163eddce95de307e463b52eb6a37290507ee8b9fe2d876d"} Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.596239 4775 scope.go:117] "RemoveContainer" containerID="49c8fe9392bad05284abe558b985bc551c7bd029e12d390cf231dfab8c392287" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.651413 4775 scope.go:117] "RemoveContainer" containerID="9f3bf924e7c9120dbb6524b0c297a365abf7f5a3e45a55598a198ce6b85db3f3" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.655090 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rtxmd"] Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.666758 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rtxmd"] Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.684985 4775 scope.go:117] "RemoveContainer" containerID="5b4b4553a268cfe2ba5b841dbc7d7f9b46c200b626551096d4f8807108ff253e" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.726494 4775 scope.go:117] "RemoveContainer" containerID="49c8fe9392bad05284abe558b985bc551c7bd029e12d390cf231dfab8c392287" Oct 02 02:04:48 crc kubenswrapper[4775]: E1002 02:04:48.727242 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49c8fe9392bad05284abe558b985bc551c7bd029e12d390cf231dfab8c392287\": container with ID starting with 49c8fe9392bad05284abe558b985bc551c7bd029e12d390cf231dfab8c392287 not found: ID does not exist" containerID="49c8fe9392bad05284abe558b985bc551c7bd029e12d390cf231dfab8c392287" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.727310 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49c8fe9392bad05284abe558b985bc551c7bd029e12d390cf231dfab8c392287"} err="failed to get container status \"49c8fe9392bad05284abe558b985bc551c7bd029e12d390cf231dfab8c392287\": rpc error: code = NotFound desc = could not find container \"49c8fe9392bad05284abe558b985bc551c7bd029e12d390cf231dfab8c392287\": container with ID starting with 49c8fe9392bad05284abe558b985bc551c7bd029e12d390cf231dfab8c392287 not found: ID does not exist" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.727357 4775 scope.go:117] "RemoveContainer" containerID="9f3bf924e7c9120dbb6524b0c297a365abf7f5a3e45a55598a198ce6b85db3f3" Oct 02 02:04:48 crc kubenswrapper[4775]: E1002 02:04:48.728008 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f3bf924e7c9120dbb6524b0c297a365abf7f5a3e45a55598a198ce6b85db3f3\": container with ID starting with 9f3bf924e7c9120dbb6524b0c297a365abf7f5a3e45a55598a198ce6b85db3f3 not found: ID does not exist" containerID="9f3bf924e7c9120dbb6524b0c297a365abf7f5a3e45a55598a198ce6b85db3f3" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.728065 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f3bf924e7c9120dbb6524b0c297a365abf7f5a3e45a55598a198ce6b85db3f3"} err="failed to get container status \"9f3bf924e7c9120dbb6524b0c297a365abf7f5a3e45a55598a198ce6b85db3f3\": rpc error: code = NotFound desc = could not find container \"9f3bf924e7c9120dbb6524b0c297a365abf7f5a3e45a55598a198ce6b85db3f3\": container with ID starting with 9f3bf924e7c9120dbb6524b0c297a365abf7f5a3e45a55598a198ce6b85db3f3 not found: ID does not exist" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.728105 4775 scope.go:117] "RemoveContainer" containerID="5b4b4553a268cfe2ba5b841dbc7d7f9b46c200b626551096d4f8807108ff253e" Oct 02 02:04:48 crc kubenswrapper[4775]: E1002 02:04:48.728534 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b4b4553a268cfe2ba5b841dbc7d7f9b46c200b626551096d4f8807108ff253e\": container with ID starting with 5b4b4553a268cfe2ba5b841dbc7d7f9b46c200b626551096d4f8807108ff253e not found: ID does not exist" containerID="5b4b4553a268cfe2ba5b841dbc7d7f9b46c200b626551096d4f8807108ff253e" Oct 02 02:04:48 crc kubenswrapper[4775]: I1002 02:04:48.728580 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b4b4553a268cfe2ba5b841dbc7d7f9b46c200b626551096d4f8807108ff253e"} err="failed to get container status \"5b4b4553a268cfe2ba5b841dbc7d7f9b46c200b626551096d4f8807108ff253e\": rpc error: code = NotFound desc = could not find container \"5b4b4553a268cfe2ba5b841dbc7d7f9b46c200b626551096d4f8807108ff253e\": container with ID starting with 5b4b4553a268cfe2ba5b841dbc7d7f9b46c200b626551096d4f8807108ff253e not found: ID does not exist" Oct 02 02:04:49 crc kubenswrapper[4775]: I1002 02:04:49.783119 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3414e6f7-99da-4d3f-870e-ebe575364a6a" path="/var/lib/kubelet/pods/3414e6f7-99da-4d3f-870e-ebe575364a6a/volumes" Oct 02 02:04:50 crc kubenswrapper[4775]: I1002 02:04:50.957244 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wk5jf"] Oct 02 02:04:50 crc kubenswrapper[4775]: E1002 02:04:50.957672 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3414e6f7-99da-4d3f-870e-ebe575364a6a" containerName="registry-server" Oct 02 02:04:50 crc kubenswrapper[4775]: I1002 02:04:50.957695 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3414e6f7-99da-4d3f-870e-ebe575364a6a" containerName="registry-server" Oct 02 02:04:50 crc kubenswrapper[4775]: E1002 02:04:50.957743 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3414e6f7-99da-4d3f-870e-ebe575364a6a" containerName="extract-utilities" Oct 02 02:04:50 crc kubenswrapper[4775]: I1002 02:04:50.957759 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3414e6f7-99da-4d3f-870e-ebe575364a6a" containerName="extract-utilities" Oct 02 02:04:50 crc kubenswrapper[4775]: E1002 02:04:50.957788 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3414e6f7-99da-4d3f-870e-ebe575364a6a" containerName="extract-content" Oct 02 02:04:50 crc kubenswrapper[4775]: I1002 02:04:50.957801 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3414e6f7-99da-4d3f-870e-ebe575364a6a" containerName="extract-content" Oct 02 02:04:50 crc kubenswrapper[4775]: I1002 02:04:50.958092 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3414e6f7-99da-4d3f-870e-ebe575364a6a" containerName="registry-server" Oct 02 02:04:50 crc kubenswrapper[4775]: I1002 02:04:50.960280 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:04:50 crc kubenswrapper[4775]: I1002 02:04:50.982070 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wk5jf"] Oct 02 02:04:51 crc kubenswrapper[4775]: I1002 02:04:51.020654 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc94t\" (UniqueName: \"kubernetes.io/projected/1a402b48-b58a-43e0-93c5-44f9c805a28c-kube-api-access-gc94t\") pod \"redhat-operators-wk5jf\" (UID: \"1a402b48-b58a-43e0-93c5-44f9c805a28c\") " pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:04:51 crc kubenswrapper[4775]: I1002 02:04:51.020757 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a402b48-b58a-43e0-93c5-44f9c805a28c-utilities\") pod \"redhat-operators-wk5jf\" (UID: \"1a402b48-b58a-43e0-93c5-44f9c805a28c\") " pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:04:51 crc kubenswrapper[4775]: I1002 02:04:51.020800 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a402b48-b58a-43e0-93c5-44f9c805a28c-catalog-content\") pod \"redhat-operators-wk5jf\" (UID: \"1a402b48-b58a-43e0-93c5-44f9c805a28c\") " pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:04:51 crc kubenswrapper[4775]: I1002 02:04:51.122149 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a402b48-b58a-43e0-93c5-44f9c805a28c-utilities\") pod \"redhat-operators-wk5jf\" (UID: \"1a402b48-b58a-43e0-93c5-44f9c805a28c\") " pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:04:51 crc kubenswrapper[4775]: I1002 02:04:51.122288 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a402b48-b58a-43e0-93c5-44f9c805a28c-catalog-content\") pod \"redhat-operators-wk5jf\" (UID: \"1a402b48-b58a-43e0-93c5-44f9c805a28c\") " pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:04:51 crc kubenswrapper[4775]: I1002 02:04:51.122745 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a402b48-b58a-43e0-93c5-44f9c805a28c-utilities\") pod \"redhat-operators-wk5jf\" (UID: \"1a402b48-b58a-43e0-93c5-44f9c805a28c\") " pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:04:51 crc kubenswrapper[4775]: I1002 02:04:51.122872 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a402b48-b58a-43e0-93c5-44f9c805a28c-catalog-content\") pod \"redhat-operators-wk5jf\" (UID: \"1a402b48-b58a-43e0-93c5-44f9c805a28c\") " pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:04:51 crc kubenswrapper[4775]: I1002 02:04:51.123101 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc94t\" (UniqueName: \"kubernetes.io/projected/1a402b48-b58a-43e0-93c5-44f9c805a28c-kube-api-access-gc94t\") pod \"redhat-operators-wk5jf\" (UID: \"1a402b48-b58a-43e0-93c5-44f9c805a28c\") " pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:04:51 crc kubenswrapper[4775]: I1002 02:04:51.146328 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc94t\" (UniqueName: \"kubernetes.io/projected/1a402b48-b58a-43e0-93c5-44f9c805a28c-kube-api-access-gc94t\") pod \"redhat-operators-wk5jf\" (UID: \"1a402b48-b58a-43e0-93c5-44f9c805a28c\") " pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:04:51 crc kubenswrapper[4775]: I1002 02:04:51.286289 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:04:51 crc kubenswrapper[4775]: I1002 02:04:51.778221 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wk5jf"] Oct 02 02:04:52 crc kubenswrapper[4775]: I1002 02:04:52.663330 4775 generic.go:334] "Generic (PLEG): container finished" podID="1a402b48-b58a-43e0-93c5-44f9c805a28c" containerID="8f2aa898a288809818b11dee7e3e687f78a03d9f484420348bb5b54a6aebfa62" exitCode=0 Oct 02 02:04:52 crc kubenswrapper[4775]: I1002 02:04:52.663419 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk5jf" event={"ID":"1a402b48-b58a-43e0-93c5-44f9c805a28c","Type":"ContainerDied","Data":"8f2aa898a288809818b11dee7e3e687f78a03d9f484420348bb5b54a6aebfa62"} Oct 02 02:04:52 crc kubenswrapper[4775]: I1002 02:04:52.663810 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk5jf" event={"ID":"1a402b48-b58a-43e0-93c5-44f9c805a28c","Type":"ContainerStarted","Data":"14136b37e629aec7c496befd97dd804dd2bc8a9ad3f7397a58901ecfae2edefe"} Oct 02 02:04:54 crc kubenswrapper[4775]: I1002 02:04:54.686099 4775 generic.go:334] "Generic (PLEG): container finished" podID="1a402b48-b58a-43e0-93c5-44f9c805a28c" containerID="be2ddd03c2bbfae2439361796e840a28788cad04592500eee5caf0685a608dfb" exitCode=0 Oct 02 02:04:54 crc kubenswrapper[4775]: I1002 02:04:54.686221 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk5jf" event={"ID":"1a402b48-b58a-43e0-93c5-44f9c805a28c","Type":"ContainerDied","Data":"be2ddd03c2bbfae2439361796e840a28788cad04592500eee5caf0685a608dfb"} Oct 02 02:04:55 crc kubenswrapper[4775]: I1002 02:04:55.705110 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk5jf" event={"ID":"1a402b48-b58a-43e0-93c5-44f9c805a28c","Type":"ContainerStarted","Data":"164e18cf4e294eec95fd41662033d1ef0ba1dfe2350df83538b6eba49b35b7a6"} Oct 02 02:04:55 crc kubenswrapper[4775]: I1002 02:04:55.745225 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wk5jf" podStartSLOduration=3.048997045 podStartE2EDuration="5.745194858s" podCreationTimestamp="2025-10-02 02:04:50 +0000 UTC" firstStartedPulling="2025-10-02 02:04:52.665480741 +0000 UTC m=+1429.832224791" lastFinishedPulling="2025-10-02 02:04:55.361678534 +0000 UTC m=+1432.528422604" observedRunningTime="2025-10-02 02:04:55.735776646 +0000 UTC m=+1432.902520756" watchObservedRunningTime="2025-10-02 02:04:55.745194858 +0000 UTC m=+1432.911938938" Oct 02 02:05:01 crc kubenswrapper[4775]: I1002 02:05:01.287003 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:05:01 crc kubenswrapper[4775]: I1002 02:05:01.287705 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:05:02 crc kubenswrapper[4775]: I1002 02:05:02.366200 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wk5jf" podUID="1a402b48-b58a-43e0-93c5-44f9c805a28c" containerName="registry-server" probeResult="failure" output=< Oct 02 02:05:02 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 02:05:02 crc kubenswrapper[4775]: > Oct 02 02:05:05 crc kubenswrapper[4775]: I1002 02:05:05.617946 4775 scope.go:117] "RemoveContainer" containerID="7de9c008d839d22eacc4887e2eeff05b010de483778c5ec8d189d41a801d7a62" Oct 02 02:05:05 crc kubenswrapper[4775]: I1002 02:05:05.689877 4775 scope.go:117] "RemoveContainer" containerID="669d1080c1ef87bdc5aa946628176e085a6c1ef96db7059d070d3a1cdd097266" Oct 02 02:05:05 crc kubenswrapper[4775]: I1002 02:05:05.744394 4775 scope.go:117] "RemoveContainer" containerID="47cbd1a6f145ee46322b61858213b665fd7c5043d5451925bf26dd59dec44238" Oct 02 02:05:05 crc kubenswrapper[4775]: I1002 02:05:05.778547 4775 scope.go:117] "RemoveContainer" containerID="c0f5e7ac59fadc846940a5bf746aaf3da2e4edb2f52a02d0bbdc57d8847e6271" Oct 02 02:05:11 crc kubenswrapper[4775]: I1002 02:05:11.345865 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:05:11 crc kubenswrapper[4775]: I1002 02:05:11.430646 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:05:11 crc kubenswrapper[4775]: I1002 02:05:11.602748 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wk5jf"] Oct 02 02:05:12 crc kubenswrapper[4775]: I1002 02:05:12.910298 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wk5jf" podUID="1a402b48-b58a-43e0-93c5-44f9c805a28c" containerName="registry-server" containerID="cri-o://164e18cf4e294eec95fd41662033d1ef0ba1dfe2350df83538b6eba49b35b7a6" gracePeriod=2 Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.373084 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.402635 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc94t\" (UniqueName: \"kubernetes.io/projected/1a402b48-b58a-43e0-93c5-44f9c805a28c-kube-api-access-gc94t\") pod \"1a402b48-b58a-43e0-93c5-44f9c805a28c\" (UID: \"1a402b48-b58a-43e0-93c5-44f9c805a28c\") " Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.402729 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a402b48-b58a-43e0-93c5-44f9c805a28c-utilities\") pod \"1a402b48-b58a-43e0-93c5-44f9c805a28c\" (UID: \"1a402b48-b58a-43e0-93c5-44f9c805a28c\") " Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.402868 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a402b48-b58a-43e0-93c5-44f9c805a28c-catalog-content\") pod \"1a402b48-b58a-43e0-93c5-44f9c805a28c\" (UID: \"1a402b48-b58a-43e0-93c5-44f9c805a28c\") " Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.405619 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a402b48-b58a-43e0-93c5-44f9c805a28c-utilities" (OuterVolumeSpecName: "utilities") pod "1a402b48-b58a-43e0-93c5-44f9c805a28c" (UID: "1a402b48-b58a-43e0-93c5-44f9c805a28c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.418205 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a402b48-b58a-43e0-93c5-44f9c805a28c-kube-api-access-gc94t" (OuterVolumeSpecName: "kube-api-access-gc94t") pod "1a402b48-b58a-43e0-93c5-44f9c805a28c" (UID: "1a402b48-b58a-43e0-93c5-44f9c805a28c"). InnerVolumeSpecName "kube-api-access-gc94t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.503984 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a402b48-b58a-43e0-93c5-44f9c805a28c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.504255 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc94t\" (UniqueName: \"kubernetes.io/projected/1a402b48-b58a-43e0-93c5-44f9c805a28c-kube-api-access-gc94t\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.533575 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a402b48-b58a-43e0-93c5-44f9c805a28c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a402b48-b58a-43e0-93c5-44f9c805a28c" (UID: "1a402b48-b58a-43e0-93c5-44f9c805a28c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.605312 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a402b48-b58a-43e0-93c5-44f9c805a28c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.934704 4775 generic.go:334] "Generic (PLEG): container finished" podID="1a402b48-b58a-43e0-93c5-44f9c805a28c" containerID="164e18cf4e294eec95fd41662033d1ef0ba1dfe2350df83538b6eba49b35b7a6" exitCode=0 Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.934752 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk5jf" event={"ID":"1a402b48-b58a-43e0-93c5-44f9c805a28c","Type":"ContainerDied","Data":"164e18cf4e294eec95fd41662033d1ef0ba1dfe2350df83538b6eba49b35b7a6"} Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.934790 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wk5jf" event={"ID":"1a402b48-b58a-43e0-93c5-44f9c805a28c","Type":"ContainerDied","Data":"14136b37e629aec7c496befd97dd804dd2bc8a9ad3f7397a58901ecfae2edefe"} Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.934813 4775 scope.go:117] "RemoveContainer" containerID="164e18cf4e294eec95fd41662033d1ef0ba1dfe2350df83538b6eba49b35b7a6" Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.934845 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wk5jf" Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.965322 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wk5jf"] Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.971707 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wk5jf"] Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.974082 4775 scope.go:117] "RemoveContainer" containerID="be2ddd03c2bbfae2439361796e840a28788cad04592500eee5caf0685a608dfb" Oct 02 02:05:13 crc kubenswrapper[4775]: I1002 02:05:13.994912 4775 scope.go:117] "RemoveContainer" containerID="8f2aa898a288809818b11dee7e3e687f78a03d9f484420348bb5b54a6aebfa62" Oct 02 02:05:14 crc kubenswrapper[4775]: I1002 02:05:14.036404 4775 scope.go:117] "RemoveContainer" containerID="164e18cf4e294eec95fd41662033d1ef0ba1dfe2350df83538b6eba49b35b7a6" Oct 02 02:05:14 crc kubenswrapper[4775]: E1002 02:05:14.036896 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"164e18cf4e294eec95fd41662033d1ef0ba1dfe2350df83538b6eba49b35b7a6\": container with ID starting with 164e18cf4e294eec95fd41662033d1ef0ba1dfe2350df83538b6eba49b35b7a6 not found: ID does not exist" containerID="164e18cf4e294eec95fd41662033d1ef0ba1dfe2350df83538b6eba49b35b7a6" Oct 02 02:05:14 crc kubenswrapper[4775]: I1002 02:05:14.036975 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"164e18cf4e294eec95fd41662033d1ef0ba1dfe2350df83538b6eba49b35b7a6"} err="failed to get container status \"164e18cf4e294eec95fd41662033d1ef0ba1dfe2350df83538b6eba49b35b7a6\": rpc error: code = NotFound desc = could not find container \"164e18cf4e294eec95fd41662033d1ef0ba1dfe2350df83538b6eba49b35b7a6\": container with ID starting with 164e18cf4e294eec95fd41662033d1ef0ba1dfe2350df83538b6eba49b35b7a6 not found: ID does not exist" Oct 02 02:05:14 crc kubenswrapper[4775]: I1002 02:05:14.037014 4775 scope.go:117] "RemoveContainer" containerID="be2ddd03c2bbfae2439361796e840a28788cad04592500eee5caf0685a608dfb" Oct 02 02:05:14 crc kubenswrapper[4775]: E1002 02:05:14.037438 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be2ddd03c2bbfae2439361796e840a28788cad04592500eee5caf0685a608dfb\": container with ID starting with be2ddd03c2bbfae2439361796e840a28788cad04592500eee5caf0685a608dfb not found: ID does not exist" containerID="be2ddd03c2bbfae2439361796e840a28788cad04592500eee5caf0685a608dfb" Oct 02 02:05:14 crc kubenswrapper[4775]: I1002 02:05:14.037514 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be2ddd03c2bbfae2439361796e840a28788cad04592500eee5caf0685a608dfb"} err="failed to get container status \"be2ddd03c2bbfae2439361796e840a28788cad04592500eee5caf0685a608dfb\": rpc error: code = NotFound desc = could not find container \"be2ddd03c2bbfae2439361796e840a28788cad04592500eee5caf0685a608dfb\": container with ID starting with be2ddd03c2bbfae2439361796e840a28788cad04592500eee5caf0685a608dfb not found: ID does not exist" Oct 02 02:05:14 crc kubenswrapper[4775]: I1002 02:05:14.037562 4775 scope.go:117] "RemoveContainer" containerID="8f2aa898a288809818b11dee7e3e687f78a03d9f484420348bb5b54a6aebfa62" Oct 02 02:05:14 crc kubenswrapper[4775]: E1002 02:05:14.037896 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f2aa898a288809818b11dee7e3e687f78a03d9f484420348bb5b54a6aebfa62\": container with ID starting with 8f2aa898a288809818b11dee7e3e687f78a03d9f484420348bb5b54a6aebfa62 not found: ID does not exist" containerID="8f2aa898a288809818b11dee7e3e687f78a03d9f484420348bb5b54a6aebfa62" Oct 02 02:05:14 crc kubenswrapper[4775]: I1002 02:05:14.037925 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f2aa898a288809818b11dee7e3e687f78a03d9f484420348bb5b54a6aebfa62"} err="failed to get container status \"8f2aa898a288809818b11dee7e3e687f78a03d9f484420348bb5b54a6aebfa62\": rpc error: code = NotFound desc = could not find container \"8f2aa898a288809818b11dee7e3e687f78a03d9f484420348bb5b54a6aebfa62\": container with ID starting with 8f2aa898a288809818b11dee7e3e687f78a03d9f484420348bb5b54a6aebfa62 not found: ID does not exist" Oct 02 02:05:15 crc kubenswrapper[4775]: I1002 02:05:15.796281 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a402b48-b58a-43e0-93c5-44f9c805a28c" path="/var/lib/kubelet/pods/1a402b48-b58a-43e0-93c5-44f9c805a28c/volumes" Oct 02 02:05:37 crc kubenswrapper[4775]: I1002 02:05:37.237156 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:05:37 crc kubenswrapper[4775]: I1002 02:05:37.237831 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.170389 4775 scope.go:117] "RemoveContainer" containerID="97fb4e7966718769851726806782e899704e90130857cbf32f13bf170d9d8525" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.200183 4775 scope.go:117] "RemoveContainer" containerID="b8c221ce14f71e8617a931e96a1dc0043afe3b845f98c89cf4bbb25eb5600820" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.251314 4775 scope.go:117] "RemoveContainer" containerID="dd169edaae10c781fde2aef6637ad4843abc69b4d5b73927c6dc220be9781a04" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.283821 4775 scope.go:117] "RemoveContainer" containerID="4e8acae44e5a153d024f7a04cbd26852d78255397c1796bbd4fd45591adafa27" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.311128 4775 scope.go:117] "RemoveContainer" containerID="6896915cc67aadd994ad499da8c5082fb11be528fb9225adc20db0b4e4bee028" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.359072 4775 scope.go:117] "RemoveContainer" containerID="af48ca44cb3f8a0a6d273249e42c44fcc94dcbeba408fe59a65dad59c211b74b" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.390453 4775 scope.go:117] "RemoveContainer" containerID="2c422b8f2c034369225bbf9b72a631e551bac35a17c65cc2a633d7e5f7f6a986" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.416392 4775 scope.go:117] "RemoveContainer" containerID="fd0824af89285e773df2653c2d0e7d72035cffd31ec8d3fbdbb045018aca4264" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.440761 4775 scope.go:117] "RemoveContainer" containerID="63549f219f1ba666ecd985751af31b24d4b8335c07e064c09cf580d3c02e98c2" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.465853 4775 scope.go:117] "RemoveContainer" containerID="b4c3d90e94f7841e53ddb7eb14f21806ea980605924b570fe8fd49f97bbe5369" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.495229 4775 scope.go:117] "RemoveContainer" containerID="4af49828720104976bf7c6bb5fc2da2e9b2da748200c0332a9ce42070048ad17" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.516863 4775 scope.go:117] "RemoveContainer" containerID="8b262fc92c0f625440f236f2622f5d81981f9d0a4057e92dd3a337e627891bd7" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.558290 4775 scope.go:117] "RemoveContainer" containerID="5b6fefa0d5b41b1d53bf1ad854a62fb1e03570cf6981e37e33270368af706fc8" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.632336 4775 scope.go:117] "RemoveContainer" containerID="ea5cddb115e84b5b26c272b09e40f8bf1b5253c0f27dee2955ed905b7c9f3a4f" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.688139 4775 scope.go:117] "RemoveContainer" containerID="83c116e723946985568da8014ced2542488c38c32b369e695abb19f1ab5c6a58" Oct 02 02:06:06 crc kubenswrapper[4775]: I1002 02:06:06.705489 4775 scope.go:117] "RemoveContainer" containerID="fb50998c7136bd8340419cbe461f93a9ee8bc40a5577fb31338f8fea2f87aead" Oct 02 02:06:07 crc kubenswrapper[4775]: I1002 02:06:07.233695 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:06:07 crc kubenswrapper[4775]: I1002 02:06:07.233798 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.265109 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s5g8k"] Oct 02 02:06:20 crc kubenswrapper[4775]: E1002 02:06:20.266505 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a402b48-b58a-43e0-93c5-44f9c805a28c" containerName="extract-content" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.266540 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a402b48-b58a-43e0-93c5-44f9c805a28c" containerName="extract-content" Oct 02 02:06:20 crc kubenswrapper[4775]: E1002 02:06:20.266585 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a402b48-b58a-43e0-93c5-44f9c805a28c" containerName="registry-server" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.266605 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a402b48-b58a-43e0-93c5-44f9c805a28c" containerName="registry-server" Oct 02 02:06:20 crc kubenswrapper[4775]: E1002 02:06:20.266632 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a402b48-b58a-43e0-93c5-44f9c805a28c" containerName="extract-utilities" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.266649 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a402b48-b58a-43e0-93c5-44f9c805a28c" containerName="extract-utilities" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.267324 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a402b48-b58a-43e0-93c5-44f9c805a28c" containerName="registry-server" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.270364 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.280084 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s5g8k"] Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.381673 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f1f3b76-1e9c-4821-8684-8eae26effc30-catalog-content\") pod \"redhat-marketplace-s5g8k\" (UID: \"0f1f3b76-1e9c-4821-8684-8eae26effc30\") " pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.381901 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtxp8\" (UniqueName: \"kubernetes.io/projected/0f1f3b76-1e9c-4821-8684-8eae26effc30-kube-api-access-qtxp8\") pod \"redhat-marketplace-s5g8k\" (UID: \"0f1f3b76-1e9c-4821-8684-8eae26effc30\") " pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.382159 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f1f3b76-1e9c-4821-8684-8eae26effc30-utilities\") pod \"redhat-marketplace-s5g8k\" (UID: \"0f1f3b76-1e9c-4821-8684-8eae26effc30\") " pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.483590 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f1f3b76-1e9c-4821-8684-8eae26effc30-utilities\") pod \"redhat-marketplace-s5g8k\" (UID: \"0f1f3b76-1e9c-4821-8684-8eae26effc30\") " pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.483688 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f1f3b76-1e9c-4821-8684-8eae26effc30-catalog-content\") pod \"redhat-marketplace-s5g8k\" (UID: \"0f1f3b76-1e9c-4821-8684-8eae26effc30\") " pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.483749 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtxp8\" (UniqueName: \"kubernetes.io/projected/0f1f3b76-1e9c-4821-8684-8eae26effc30-kube-api-access-qtxp8\") pod \"redhat-marketplace-s5g8k\" (UID: \"0f1f3b76-1e9c-4821-8684-8eae26effc30\") " pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.484411 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f1f3b76-1e9c-4821-8684-8eae26effc30-utilities\") pod \"redhat-marketplace-s5g8k\" (UID: \"0f1f3b76-1e9c-4821-8684-8eae26effc30\") " pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.484918 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f1f3b76-1e9c-4821-8684-8eae26effc30-catalog-content\") pod \"redhat-marketplace-s5g8k\" (UID: \"0f1f3b76-1e9c-4821-8684-8eae26effc30\") " pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.523305 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtxp8\" (UniqueName: \"kubernetes.io/projected/0f1f3b76-1e9c-4821-8684-8eae26effc30-kube-api-access-qtxp8\") pod \"redhat-marketplace-s5g8k\" (UID: \"0f1f3b76-1e9c-4821-8684-8eae26effc30\") " pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:20 crc kubenswrapper[4775]: I1002 02:06:20.603898 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:21 crc kubenswrapper[4775]: I1002 02:06:21.092581 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s5g8k"] Oct 02 02:06:21 crc kubenswrapper[4775]: W1002 02:06:21.107578 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f1f3b76_1e9c_4821_8684_8eae26effc30.slice/crio-778e0d4016f4841e07c2996e99f4ee036f93e34f2d6bdc1d4168b8282479802d WatchSource:0}: Error finding container 778e0d4016f4841e07c2996e99f4ee036f93e34f2d6bdc1d4168b8282479802d: Status 404 returned error can't find the container with id 778e0d4016f4841e07c2996e99f4ee036f93e34f2d6bdc1d4168b8282479802d Oct 02 02:06:21 crc kubenswrapper[4775]: I1002 02:06:21.673698 4775 generic.go:334] "Generic (PLEG): container finished" podID="0f1f3b76-1e9c-4821-8684-8eae26effc30" containerID="8a00864b8fba2eec5bd91e89f0358929609c565b1585fdaaa4e82cbb46207210" exitCode=0 Oct 02 02:06:21 crc kubenswrapper[4775]: I1002 02:06:21.673811 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s5g8k" event={"ID":"0f1f3b76-1e9c-4821-8684-8eae26effc30","Type":"ContainerDied","Data":"8a00864b8fba2eec5bd91e89f0358929609c565b1585fdaaa4e82cbb46207210"} Oct 02 02:06:21 crc kubenswrapper[4775]: I1002 02:06:21.674181 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s5g8k" event={"ID":"0f1f3b76-1e9c-4821-8684-8eae26effc30","Type":"ContainerStarted","Data":"778e0d4016f4841e07c2996e99f4ee036f93e34f2d6bdc1d4168b8282479802d"} Oct 02 02:06:22 crc kubenswrapper[4775]: I1002 02:06:22.686111 4775 generic.go:334] "Generic (PLEG): container finished" podID="0f1f3b76-1e9c-4821-8684-8eae26effc30" containerID="5d0e7531196eb07a2dfdcc0f0b5a7d84448517c69e73e9042d6f1b7c6d1ea269" exitCode=0 Oct 02 02:06:22 crc kubenswrapper[4775]: I1002 02:06:22.686175 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s5g8k" event={"ID":"0f1f3b76-1e9c-4821-8684-8eae26effc30","Type":"ContainerDied","Data":"5d0e7531196eb07a2dfdcc0f0b5a7d84448517c69e73e9042d6f1b7c6d1ea269"} Oct 02 02:06:23 crc kubenswrapper[4775]: I1002 02:06:23.700321 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s5g8k" event={"ID":"0f1f3b76-1e9c-4821-8684-8eae26effc30","Type":"ContainerStarted","Data":"e215c5e3ccd5c1fb4a0437ded4823e389ae395047f2f6053839052679a5681d2"} Oct 02 02:06:23 crc kubenswrapper[4775]: I1002 02:06:23.731773 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s5g8k" podStartSLOduration=2.288690684 podStartE2EDuration="3.731727125s" podCreationTimestamp="2025-10-02 02:06:20 +0000 UTC" firstStartedPulling="2025-10-02 02:06:21.676703756 +0000 UTC m=+1518.843447836" lastFinishedPulling="2025-10-02 02:06:23.119740207 +0000 UTC m=+1520.286484277" observedRunningTime="2025-10-02 02:06:23.727549778 +0000 UTC m=+1520.894293858" watchObservedRunningTime="2025-10-02 02:06:23.731727125 +0000 UTC m=+1520.898471175" Oct 02 02:06:30 crc kubenswrapper[4775]: I1002 02:06:30.604381 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:30 crc kubenswrapper[4775]: I1002 02:06:30.605294 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:30 crc kubenswrapper[4775]: I1002 02:06:30.664624 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:30 crc kubenswrapper[4775]: I1002 02:06:30.833583 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:30 crc kubenswrapper[4775]: I1002 02:06:30.910896 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s5g8k"] Oct 02 02:06:32 crc kubenswrapper[4775]: I1002 02:06:32.787169 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-s5g8k" podUID="0f1f3b76-1e9c-4821-8684-8eae26effc30" containerName="registry-server" containerID="cri-o://e215c5e3ccd5c1fb4a0437ded4823e389ae395047f2f6053839052679a5681d2" gracePeriod=2 Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.308647 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.415938 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f1f3b76-1e9c-4821-8684-8eae26effc30-catalog-content\") pod \"0f1f3b76-1e9c-4821-8684-8eae26effc30\" (UID: \"0f1f3b76-1e9c-4821-8684-8eae26effc30\") " Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.416074 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f1f3b76-1e9c-4821-8684-8eae26effc30-utilities\") pod \"0f1f3b76-1e9c-4821-8684-8eae26effc30\" (UID: \"0f1f3b76-1e9c-4821-8684-8eae26effc30\") " Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.416116 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qtxp8\" (UniqueName: \"kubernetes.io/projected/0f1f3b76-1e9c-4821-8684-8eae26effc30-kube-api-access-qtxp8\") pod \"0f1f3b76-1e9c-4821-8684-8eae26effc30\" (UID: \"0f1f3b76-1e9c-4821-8684-8eae26effc30\") " Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.418637 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f1f3b76-1e9c-4821-8684-8eae26effc30-utilities" (OuterVolumeSpecName: "utilities") pod "0f1f3b76-1e9c-4821-8684-8eae26effc30" (UID: "0f1f3b76-1e9c-4821-8684-8eae26effc30"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.425814 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f1f3b76-1e9c-4821-8684-8eae26effc30-kube-api-access-qtxp8" (OuterVolumeSpecName: "kube-api-access-qtxp8") pod "0f1f3b76-1e9c-4821-8684-8eae26effc30" (UID: "0f1f3b76-1e9c-4821-8684-8eae26effc30"). InnerVolumeSpecName "kube-api-access-qtxp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.444773 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f1f3b76-1e9c-4821-8684-8eae26effc30-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f1f3b76-1e9c-4821-8684-8eae26effc30" (UID: "0f1f3b76-1e9c-4821-8684-8eae26effc30"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.518551 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f1f3b76-1e9c-4821-8684-8eae26effc30-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.518601 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qtxp8\" (UniqueName: \"kubernetes.io/projected/0f1f3b76-1e9c-4821-8684-8eae26effc30-kube-api-access-qtxp8\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.518623 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f1f3b76-1e9c-4821-8684-8eae26effc30-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.800117 4775 generic.go:334] "Generic (PLEG): container finished" podID="0f1f3b76-1e9c-4821-8684-8eae26effc30" containerID="e215c5e3ccd5c1fb4a0437ded4823e389ae395047f2f6053839052679a5681d2" exitCode=0 Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.800196 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s5g8k" event={"ID":"0f1f3b76-1e9c-4821-8684-8eae26effc30","Type":"ContainerDied","Data":"e215c5e3ccd5c1fb4a0437ded4823e389ae395047f2f6053839052679a5681d2"} Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.800240 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s5g8k" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.800262 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s5g8k" event={"ID":"0f1f3b76-1e9c-4821-8684-8eae26effc30","Type":"ContainerDied","Data":"778e0d4016f4841e07c2996e99f4ee036f93e34f2d6bdc1d4168b8282479802d"} Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.800304 4775 scope.go:117] "RemoveContainer" containerID="e215c5e3ccd5c1fb4a0437ded4823e389ae395047f2f6053839052679a5681d2" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.851039 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s5g8k"] Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.858408 4775 scope.go:117] "RemoveContainer" containerID="5d0e7531196eb07a2dfdcc0f0b5a7d84448517c69e73e9042d6f1b7c6d1ea269" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.859740 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-s5g8k"] Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.888667 4775 scope.go:117] "RemoveContainer" containerID="8a00864b8fba2eec5bd91e89f0358929609c565b1585fdaaa4e82cbb46207210" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.928670 4775 scope.go:117] "RemoveContainer" containerID="e215c5e3ccd5c1fb4a0437ded4823e389ae395047f2f6053839052679a5681d2" Oct 02 02:06:33 crc kubenswrapper[4775]: E1002 02:06:33.929323 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e215c5e3ccd5c1fb4a0437ded4823e389ae395047f2f6053839052679a5681d2\": container with ID starting with e215c5e3ccd5c1fb4a0437ded4823e389ae395047f2f6053839052679a5681d2 not found: ID does not exist" containerID="e215c5e3ccd5c1fb4a0437ded4823e389ae395047f2f6053839052679a5681d2" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.929394 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e215c5e3ccd5c1fb4a0437ded4823e389ae395047f2f6053839052679a5681d2"} err="failed to get container status \"e215c5e3ccd5c1fb4a0437ded4823e389ae395047f2f6053839052679a5681d2\": rpc error: code = NotFound desc = could not find container \"e215c5e3ccd5c1fb4a0437ded4823e389ae395047f2f6053839052679a5681d2\": container with ID starting with e215c5e3ccd5c1fb4a0437ded4823e389ae395047f2f6053839052679a5681d2 not found: ID does not exist" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.929439 4775 scope.go:117] "RemoveContainer" containerID="5d0e7531196eb07a2dfdcc0f0b5a7d84448517c69e73e9042d6f1b7c6d1ea269" Oct 02 02:06:33 crc kubenswrapper[4775]: E1002 02:06:33.929819 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d0e7531196eb07a2dfdcc0f0b5a7d84448517c69e73e9042d6f1b7c6d1ea269\": container with ID starting with 5d0e7531196eb07a2dfdcc0f0b5a7d84448517c69e73e9042d6f1b7c6d1ea269 not found: ID does not exist" containerID="5d0e7531196eb07a2dfdcc0f0b5a7d84448517c69e73e9042d6f1b7c6d1ea269" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.930006 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d0e7531196eb07a2dfdcc0f0b5a7d84448517c69e73e9042d6f1b7c6d1ea269"} err="failed to get container status \"5d0e7531196eb07a2dfdcc0f0b5a7d84448517c69e73e9042d6f1b7c6d1ea269\": rpc error: code = NotFound desc = could not find container \"5d0e7531196eb07a2dfdcc0f0b5a7d84448517c69e73e9042d6f1b7c6d1ea269\": container with ID starting with 5d0e7531196eb07a2dfdcc0f0b5a7d84448517c69e73e9042d6f1b7c6d1ea269 not found: ID does not exist" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.930102 4775 scope.go:117] "RemoveContainer" containerID="8a00864b8fba2eec5bd91e89f0358929609c565b1585fdaaa4e82cbb46207210" Oct 02 02:06:33 crc kubenswrapper[4775]: E1002 02:06:33.930624 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a00864b8fba2eec5bd91e89f0358929609c565b1585fdaaa4e82cbb46207210\": container with ID starting with 8a00864b8fba2eec5bd91e89f0358929609c565b1585fdaaa4e82cbb46207210 not found: ID does not exist" containerID="8a00864b8fba2eec5bd91e89f0358929609c565b1585fdaaa4e82cbb46207210" Oct 02 02:06:33 crc kubenswrapper[4775]: I1002 02:06:33.930786 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a00864b8fba2eec5bd91e89f0358929609c565b1585fdaaa4e82cbb46207210"} err="failed to get container status \"8a00864b8fba2eec5bd91e89f0358929609c565b1585fdaaa4e82cbb46207210\": rpc error: code = NotFound desc = could not find container \"8a00864b8fba2eec5bd91e89f0358929609c565b1585fdaaa4e82cbb46207210\": container with ID starting with 8a00864b8fba2eec5bd91e89f0358929609c565b1585fdaaa4e82cbb46207210 not found: ID does not exist" Oct 02 02:06:35 crc kubenswrapper[4775]: I1002 02:06:35.788478 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f1f3b76-1e9c-4821-8684-8eae26effc30" path="/var/lib/kubelet/pods/0f1f3b76-1e9c-4821-8684-8eae26effc30/volumes" Oct 02 02:06:37 crc kubenswrapper[4775]: I1002 02:06:37.233363 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:06:37 crc kubenswrapper[4775]: I1002 02:06:37.233883 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:06:37 crc kubenswrapper[4775]: I1002 02:06:37.234005 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 02:06:37 crc kubenswrapper[4775]: I1002 02:06:37.235185 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:06:37 crc kubenswrapper[4775]: I1002 02:06:37.235311 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" gracePeriod=600 Oct 02 02:06:37 crc kubenswrapper[4775]: I1002 02:06:37.844503 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" exitCode=0 Oct 02 02:06:37 crc kubenswrapper[4775]: I1002 02:06:37.844595 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99"} Oct 02 02:06:37 crc kubenswrapper[4775]: I1002 02:06:37.844713 4775 scope.go:117] "RemoveContainer" containerID="08d56b32ae1f08ce86ec95ee88e05d04f2a7f9064777b463fc71e568251c9583" Oct 02 02:06:37 crc kubenswrapper[4775]: E1002 02:06:37.871226 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:06:38 crc kubenswrapper[4775]: I1002 02:06:38.859875 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:06:38 crc kubenswrapper[4775]: E1002 02:06:38.860514 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:06:50 crc kubenswrapper[4775]: I1002 02:06:50.765349 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:06:50 crc kubenswrapper[4775]: E1002 02:06:50.766564 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:07:01 crc kubenswrapper[4775]: I1002 02:07:01.765609 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:07:01 crc kubenswrapper[4775]: E1002 02:07:01.766800 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:07:06 crc kubenswrapper[4775]: I1002 02:07:06.966924 4775 scope.go:117] "RemoveContainer" containerID="11fc048e233fc00c87c2d04d492ad8e05971c6d7aac57380af9bd95f239d9a3d" Oct 02 02:07:07 crc kubenswrapper[4775]: I1002 02:07:07.009919 4775 scope.go:117] "RemoveContainer" containerID="3aff4b0f97413fafd4a12944ee3201d623169d99e066f7bf1d2aa5a8977df17f" Oct 02 02:07:07 crc kubenswrapper[4775]: I1002 02:07:07.041322 4775 scope.go:117] "RemoveContainer" containerID="3fc5ef432fab421b40f1e30656d3ab99d83e495569dde9d28576177506951a8b" Oct 02 02:07:07 crc kubenswrapper[4775]: I1002 02:07:07.086764 4775 scope.go:117] "RemoveContainer" containerID="f03365d9159046e89cfea91c9f68ea0948f70fcf7620f4ce9ed01400f90dc48d" Oct 02 02:07:07 crc kubenswrapper[4775]: I1002 02:07:07.131111 4775 scope.go:117] "RemoveContainer" containerID="1cde1b09d6db770bfcaa3f911697783619cbc8a288399077a17545ecd5108f0f" Oct 02 02:07:07 crc kubenswrapper[4775]: I1002 02:07:07.156995 4775 scope.go:117] "RemoveContainer" containerID="b2061818493c7906d4c6165a2929e0f522ff078e37ced6c3df22b374b06afbf0" Oct 02 02:07:07 crc kubenswrapper[4775]: I1002 02:07:07.229789 4775 scope.go:117] "RemoveContainer" containerID="fb3313fc8fe1d8fb265ea0d218297b58a6c957b9693cc2db9285b5a3bdd5aa98" Oct 02 02:07:07 crc kubenswrapper[4775]: I1002 02:07:07.257006 4775 scope.go:117] "RemoveContainer" containerID="ae9bd041c2c057dee311d85415e1bfd222d0109f3f01cb2fcd2149362f273f6e" Oct 02 02:07:07 crc kubenswrapper[4775]: I1002 02:07:07.282971 4775 scope.go:117] "RemoveContainer" containerID="63635a39be96c75e16a0cd5c161326889020e66580536e869bd6d92c322a7bb6" Oct 02 02:07:07 crc kubenswrapper[4775]: I1002 02:07:07.311944 4775 scope.go:117] "RemoveContainer" containerID="10a348acd950eaad46c525129402d791195a91d90d80d01aa576064a8aeec468" Oct 02 02:07:16 crc kubenswrapper[4775]: I1002 02:07:16.766098 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:07:16 crc kubenswrapper[4775]: E1002 02:07:16.767411 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:07:27 crc kubenswrapper[4775]: I1002 02:07:27.765858 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:07:27 crc kubenswrapper[4775]: E1002 02:07:27.768277 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:07:39 crc kubenswrapper[4775]: I1002 02:07:39.765544 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:07:39 crc kubenswrapper[4775]: E1002 02:07:39.766301 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:07:50 crc kubenswrapper[4775]: I1002 02:07:50.765213 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:07:50 crc kubenswrapper[4775]: E1002 02:07:50.766219 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:08:01 crc kubenswrapper[4775]: I1002 02:08:01.765644 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:08:01 crc kubenswrapper[4775]: E1002 02:08:01.766458 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:08:07 crc kubenswrapper[4775]: I1002 02:08:07.512005 4775 scope.go:117] "RemoveContainer" containerID="424d08625ff4c08c37f7520793f0493961f7668b4af123a929558e4a08e9f894" Oct 02 02:08:07 crc kubenswrapper[4775]: I1002 02:08:07.553786 4775 scope.go:117] "RemoveContainer" containerID="4970d9325d20fcb9a2e4c4c76b0503d72a5a3a440e5f7bc4e1e03c493386fb84" Oct 02 02:08:07 crc kubenswrapper[4775]: I1002 02:08:07.606415 4775 scope.go:117] "RemoveContainer" containerID="220d9b9728d9af47f9f7f6b6a307303f3baa2a782a03e007f8e2182217bf116d" Oct 02 02:08:07 crc kubenswrapper[4775]: I1002 02:08:07.639036 4775 scope.go:117] "RemoveContainer" containerID="54e4763076c1e59bd43a8d1b7a00c8dc77ba9c288b69de634c971e266ee9e637" Oct 02 02:08:07 crc kubenswrapper[4775]: I1002 02:08:07.674619 4775 scope.go:117] "RemoveContainer" containerID="6a762875446f3a628ec2e350627ea55effedef200ef0ae26856c73ea9d59025d" Oct 02 02:08:07 crc kubenswrapper[4775]: I1002 02:08:07.727610 4775 scope.go:117] "RemoveContainer" containerID="cdf8eb01511f36ccb375b3270b0dcf26234a99df6e6ec88e6e129895d780779d" Oct 02 02:08:07 crc kubenswrapper[4775]: I1002 02:08:07.755919 4775 scope.go:117] "RemoveContainer" containerID="e93c0d6ee8dbff552b995e91001193f38793e530b0cc3fab57aace81d77b326e" Oct 02 02:08:07 crc kubenswrapper[4775]: I1002 02:08:07.785270 4775 scope.go:117] "RemoveContainer" containerID="e16c239ca1126dd791cb5053ad8e44fb7f55ca4a02d7a26f73e67a82f58c593d" Oct 02 02:08:07 crc kubenswrapper[4775]: I1002 02:08:07.844130 4775 scope.go:117] "RemoveContainer" containerID="ce4365b06788ec5e2d1ab7f03c97b339cfee51b36bf1b4f570178d210ed939dc" Oct 02 02:08:16 crc kubenswrapper[4775]: I1002 02:08:16.765842 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:08:16 crc kubenswrapper[4775]: E1002 02:08:16.766919 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:08:28 crc kubenswrapper[4775]: I1002 02:08:28.765156 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:08:28 crc kubenswrapper[4775]: E1002 02:08:28.765752 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:08:41 crc kubenswrapper[4775]: I1002 02:08:41.765869 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:08:41 crc kubenswrapper[4775]: E1002 02:08:41.766790 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:08:55 crc kubenswrapper[4775]: I1002 02:08:55.765889 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:08:55 crc kubenswrapper[4775]: E1002 02:08:55.767065 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:09:06 crc kubenswrapper[4775]: I1002 02:09:06.765118 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:09:06 crc kubenswrapper[4775]: E1002 02:09:06.765945 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:09:08 crc kubenswrapper[4775]: I1002 02:09:08.056797 4775 scope.go:117] "RemoveContainer" containerID="5ed104134ed32605e39b2c63eeab848fad590d9ebf90dc94c858b92bc55ff88b" Oct 02 02:09:08 crc kubenswrapper[4775]: I1002 02:09:08.087024 4775 scope.go:117] "RemoveContainer" containerID="b66fc6b4d3be69dcb9c2d949e6dfe8d6ab6ddb59f7b8ad17c8380c0b6034ccf6" Oct 02 02:09:08 crc kubenswrapper[4775]: I1002 02:09:08.119722 4775 scope.go:117] "RemoveContainer" containerID="95e4dfafbb440c4413dfe2954b28be460f01b7d55d95e1321662f71a709463a5" Oct 02 02:09:08 crc kubenswrapper[4775]: I1002 02:09:08.185396 4775 scope.go:117] "RemoveContainer" containerID="9bdbd014e9940be2ba8f1e074e43a05086b36eec14f5338e99629708c2efdca9" Oct 02 02:09:08 crc kubenswrapper[4775]: I1002 02:09:08.213985 4775 scope.go:117] "RemoveContainer" containerID="38c453bdb403040517a2812e8643acfa55f8124baa338e04c4028a69bd4cbbd1" Oct 02 02:09:08 crc kubenswrapper[4775]: I1002 02:09:08.242653 4775 scope.go:117] "RemoveContainer" containerID="de21c2812f8579db4805f66ccfa27b1e881d6fc1abd46304b474e22b41658e71" Oct 02 02:09:08 crc kubenswrapper[4775]: I1002 02:09:08.300347 4775 scope.go:117] "RemoveContainer" containerID="6551e9c5f62eb1393127c59a3b81e28867173c615ce4b3cc689c3b24c62e2044" Oct 02 02:09:13 crc kubenswrapper[4775]: I1002 02:09:13.839656 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zg74z"] Oct 02 02:09:13 crc kubenswrapper[4775]: E1002 02:09:13.841009 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1f3b76-1e9c-4821-8684-8eae26effc30" containerName="registry-server" Oct 02 02:09:13 crc kubenswrapper[4775]: I1002 02:09:13.841046 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1f3b76-1e9c-4821-8684-8eae26effc30" containerName="registry-server" Oct 02 02:09:13 crc kubenswrapper[4775]: E1002 02:09:13.841084 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1f3b76-1e9c-4821-8684-8eae26effc30" containerName="extract-content" Oct 02 02:09:13 crc kubenswrapper[4775]: I1002 02:09:13.841097 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1f3b76-1e9c-4821-8684-8eae26effc30" containerName="extract-content" Oct 02 02:09:13 crc kubenswrapper[4775]: E1002 02:09:13.841134 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f1f3b76-1e9c-4821-8684-8eae26effc30" containerName="extract-utilities" Oct 02 02:09:13 crc kubenswrapper[4775]: I1002 02:09:13.841146 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f1f3b76-1e9c-4821-8684-8eae26effc30" containerName="extract-utilities" Oct 02 02:09:13 crc kubenswrapper[4775]: I1002 02:09:13.841502 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f1f3b76-1e9c-4821-8684-8eae26effc30" containerName="registry-server" Oct 02 02:09:13 crc kubenswrapper[4775]: I1002 02:09:13.843463 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:13 crc kubenswrapper[4775]: I1002 02:09:13.871012 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zg74z"] Oct 02 02:09:13 crc kubenswrapper[4775]: I1002 02:09:13.906580 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-catalog-content\") pod \"certified-operators-zg74z\" (UID: \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\") " pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:13 crc kubenswrapper[4775]: I1002 02:09:13.906728 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8bsk\" (UniqueName: \"kubernetes.io/projected/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-kube-api-access-z8bsk\") pod \"certified-operators-zg74z\" (UID: \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\") " pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:13 crc kubenswrapper[4775]: I1002 02:09:13.906821 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-utilities\") pod \"certified-operators-zg74z\" (UID: \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\") " pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:14 crc kubenswrapper[4775]: I1002 02:09:14.008363 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8bsk\" (UniqueName: \"kubernetes.io/projected/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-kube-api-access-z8bsk\") pod \"certified-operators-zg74z\" (UID: \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\") " pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:14 crc kubenswrapper[4775]: I1002 02:09:14.008519 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-utilities\") pod \"certified-operators-zg74z\" (UID: \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\") " pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:14 crc kubenswrapper[4775]: I1002 02:09:14.008599 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-catalog-content\") pod \"certified-operators-zg74z\" (UID: \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\") " pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:14 crc kubenswrapper[4775]: I1002 02:09:14.009405 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-utilities\") pod \"certified-operators-zg74z\" (UID: \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\") " pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:14 crc kubenswrapper[4775]: I1002 02:09:14.009561 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-catalog-content\") pod \"certified-operators-zg74z\" (UID: \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\") " pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:14 crc kubenswrapper[4775]: I1002 02:09:14.037162 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8bsk\" (UniqueName: \"kubernetes.io/projected/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-kube-api-access-z8bsk\") pod \"certified-operators-zg74z\" (UID: \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\") " pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:14 crc kubenswrapper[4775]: I1002 02:09:14.192863 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:14 crc kubenswrapper[4775]: I1002 02:09:14.493219 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zg74z"] Oct 02 02:09:15 crc kubenswrapper[4775]: I1002 02:09:15.500024 4775 generic.go:334] "Generic (PLEG): container finished" podID="4fe64b05-e99c-4bab-9fb4-124e7abc02e1" containerID="10267e007f88ef0e8ba8d3b14e7ca06a74d846113d3bbcd6cfb748f33dc7db70" exitCode=0 Oct 02 02:09:15 crc kubenswrapper[4775]: I1002 02:09:15.500338 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg74z" event={"ID":"4fe64b05-e99c-4bab-9fb4-124e7abc02e1","Type":"ContainerDied","Data":"10267e007f88ef0e8ba8d3b14e7ca06a74d846113d3bbcd6cfb748f33dc7db70"} Oct 02 02:09:15 crc kubenswrapper[4775]: I1002 02:09:15.500599 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg74z" event={"ID":"4fe64b05-e99c-4bab-9fb4-124e7abc02e1","Type":"ContainerStarted","Data":"8c013f056132c96ae8f49bf4c2b15d25f49fde3a714a0041f5c818ea8cc41f07"} Oct 02 02:09:17 crc kubenswrapper[4775]: I1002 02:09:17.527565 4775 generic.go:334] "Generic (PLEG): container finished" podID="4fe64b05-e99c-4bab-9fb4-124e7abc02e1" containerID="da2871ed70dc2b1a69891d78539be9224e7f71b8a11acc209fe8d34199b2c57a" exitCode=0 Oct 02 02:09:17 crc kubenswrapper[4775]: I1002 02:09:17.527687 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg74z" event={"ID":"4fe64b05-e99c-4bab-9fb4-124e7abc02e1","Type":"ContainerDied","Data":"da2871ed70dc2b1a69891d78539be9224e7f71b8a11acc209fe8d34199b2c57a"} Oct 02 02:09:18 crc kubenswrapper[4775]: I1002 02:09:18.551641 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg74z" event={"ID":"4fe64b05-e99c-4bab-9fb4-124e7abc02e1","Type":"ContainerStarted","Data":"de1d290022d2c5a099bbf0b3d009d7d0798a736c4755806a8f2341f45fc485b3"} Oct 02 02:09:18 crc kubenswrapper[4775]: I1002 02:09:18.576791 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zg74z" podStartSLOduration=2.910402872 podStartE2EDuration="5.576763974s" podCreationTimestamp="2025-10-02 02:09:13 +0000 UTC" firstStartedPulling="2025-10-02 02:09:15.50380794 +0000 UTC m=+1692.670552010" lastFinishedPulling="2025-10-02 02:09:18.170169032 +0000 UTC m=+1695.336913112" observedRunningTime="2025-10-02 02:09:18.57660083 +0000 UTC m=+1695.743344900" watchObservedRunningTime="2025-10-02 02:09:18.576763974 +0000 UTC m=+1695.743508054" Oct 02 02:09:21 crc kubenswrapper[4775]: I1002 02:09:21.766264 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:09:21 crc kubenswrapper[4775]: E1002 02:09:21.767243 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:09:24 crc kubenswrapper[4775]: I1002 02:09:24.193548 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:24 crc kubenswrapper[4775]: I1002 02:09:24.193917 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:24 crc kubenswrapper[4775]: I1002 02:09:24.273379 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:24 crc kubenswrapper[4775]: I1002 02:09:24.720634 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:24 crc kubenswrapper[4775]: I1002 02:09:24.778156 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zg74z"] Oct 02 02:09:26 crc kubenswrapper[4775]: I1002 02:09:26.666096 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zg74z" podUID="4fe64b05-e99c-4bab-9fb4-124e7abc02e1" containerName="registry-server" containerID="cri-o://de1d290022d2c5a099bbf0b3d009d7d0798a736c4755806a8f2341f45fc485b3" gracePeriod=2 Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.191639 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.250662 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-catalog-content\") pod \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\" (UID: \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\") " Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.250831 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-utilities\") pod \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\" (UID: \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\") " Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.250882 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8bsk\" (UniqueName: \"kubernetes.io/projected/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-kube-api-access-z8bsk\") pod \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\" (UID: \"4fe64b05-e99c-4bab-9fb4-124e7abc02e1\") " Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.252183 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-utilities" (OuterVolumeSpecName: "utilities") pod "4fe64b05-e99c-4bab-9fb4-124e7abc02e1" (UID: "4fe64b05-e99c-4bab-9fb4-124e7abc02e1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.259237 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-kube-api-access-z8bsk" (OuterVolumeSpecName: "kube-api-access-z8bsk") pod "4fe64b05-e99c-4bab-9fb4-124e7abc02e1" (UID: "4fe64b05-e99c-4bab-9fb4-124e7abc02e1"). InnerVolumeSpecName "kube-api-access-z8bsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.302588 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4fe64b05-e99c-4bab-9fb4-124e7abc02e1" (UID: "4fe64b05-e99c-4bab-9fb4-124e7abc02e1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.353503 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.353556 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.353577 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8bsk\" (UniqueName: \"kubernetes.io/projected/4fe64b05-e99c-4bab-9fb4-124e7abc02e1-kube-api-access-z8bsk\") on node \"crc\" DevicePath \"\"" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.679303 4775 generic.go:334] "Generic (PLEG): container finished" podID="4fe64b05-e99c-4bab-9fb4-124e7abc02e1" containerID="de1d290022d2c5a099bbf0b3d009d7d0798a736c4755806a8f2341f45fc485b3" exitCode=0 Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.679368 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg74z" event={"ID":"4fe64b05-e99c-4bab-9fb4-124e7abc02e1","Type":"ContainerDied","Data":"de1d290022d2c5a099bbf0b3d009d7d0798a736c4755806a8f2341f45fc485b3"} Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.679681 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zg74z" event={"ID":"4fe64b05-e99c-4bab-9fb4-124e7abc02e1","Type":"ContainerDied","Data":"8c013f056132c96ae8f49bf4c2b15d25f49fde3a714a0041f5c818ea8cc41f07"} Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.679714 4775 scope.go:117] "RemoveContainer" containerID="de1d290022d2c5a099bbf0b3d009d7d0798a736c4755806a8f2341f45fc485b3" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.679398 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zg74z" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.724392 4775 scope.go:117] "RemoveContainer" containerID="da2871ed70dc2b1a69891d78539be9224e7f71b8a11acc209fe8d34199b2c57a" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.745343 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zg74z"] Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.754630 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zg74z"] Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.761227 4775 scope.go:117] "RemoveContainer" containerID="10267e007f88ef0e8ba8d3b14e7ca06a74d846113d3bbcd6cfb748f33dc7db70" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.779757 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fe64b05-e99c-4bab-9fb4-124e7abc02e1" path="/var/lib/kubelet/pods/4fe64b05-e99c-4bab-9fb4-124e7abc02e1/volumes" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.799233 4775 scope.go:117] "RemoveContainer" containerID="de1d290022d2c5a099bbf0b3d009d7d0798a736c4755806a8f2341f45fc485b3" Oct 02 02:09:27 crc kubenswrapper[4775]: E1002 02:09:27.799773 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de1d290022d2c5a099bbf0b3d009d7d0798a736c4755806a8f2341f45fc485b3\": container with ID starting with de1d290022d2c5a099bbf0b3d009d7d0798a736c4755806a8f2341f45fc485b3 not found: ID does not exist" containerID="de1d290022d2c5a099bbf0b3d009d7d0798a736c4755806a8f2341f45fc485b3" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.799820 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de1d290022d2c5a099bbf0b3d009d7d0798a736c4755806a8f2341f45fc485b3"} err="failed to get container status \"de1d290022d2c5a099bbf0b3d009d7d0798a736c4755806a8f2341f45fc485b3\": rpc error: code = NotFound desc = could not find container \"de1d290022d2c5a099bbf0b3d009d7d0798a736c4755806a8f2341f45fc485b3\": container with ID starting with de1d290022d2c5a099bbf0b3d009d7d0798a736c4755806a8f2341f45fc485b3 not found: ID does not exist" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.799854 4775 scope.go:117] "RemoveContainer" containerID="da2871ed70dc2b1a69891d78539be9224e7f71b8a11acc209fe8d34199b2c57a" Oct 02 02:09:27 crc kubenswrapper[4775]: E1002 02:09:27.800223 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da2871ed70dc2b1a69891d78539be9224e7f71b8a11acc209fe8d34199b2c57a\": container with ID starting with da2871ed70dc2b1a69891d78539be9224e7f71b8a11acc209fe8d34199b2c57a not found: ID does not exist" containerID="da2871ed70dc2b1a69891d78539be9224e7f71b8a11acc209fe8d34199b2c57a" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.800293 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da2871ed70dc2b1a69891d78539be9224e7f71b8a11acc209fe8d34199b2c57a"} err="failed to get container status \"da2871ed70dc2b1a69891d78539be9224e7f71b8a11acc209fe8d34199b2c57a\": rpc error: code = NotFound desc = could not find container \"da2871ed70dc2b1a69891d78539be9224e7f71b8a11acc209fe8d34199b2c57a\": container with ID starting with da2871ed70dc2b1a69891d78539be9224e7f71b8a11acc209fe8d34199b2c57a not found: ID does not exist" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.800327 4775 scope.go:117] "RemoveContainer" containerID="10267e007f88ef0e8ba8d3b14e7ca06a74d846113d3bbcd6cfb748f33dc7db70" Oct 02 02:09:27 crc kubenswrapper[4775]: E1002 02:09:27.800751 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10267e007f88ef0e8ba8d3b14e7ca06a74d846113d3bbcd6cfb748f33dc7db70\": container with ID starting with 10267e007f88ef0e8ba8d3b14e7ca06a74d846113d3bbcd6cfb748f33dc7db70 not found: ID does not exist" containerID="10267e007f88ef0e8ba8d3b14e7ca06a74d846113d3bbcd6cfb748f33dc7db70" Oct 02 02:09:27 crc kubenswrapper[4775]: I1002 02:09:27.800784 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10267e007f88ef0e8ba8d3b14e7ca06a74d846113d3bbcd6cfb748f33dc7db70"} err="failed to get container status \"10267e007f88ef0e8ba8d3b14e7ca06a74d846113d3bbcd6cfb748f33dc7db70\": rpc error: code = NotFound desc = could not find container \"10267e007f88ef0e8ba8d3b14e7ca06a74d846113d3bbcd6cfb748f33dc7db70\": container with ID starting with 10267e007f88ef0e8ba8d3b14e7ca06a74d846113d3bbcd6cfb748f33dc7db70 not found: ID does not exist" Oct 02 02:09:33 crc kubenswrapper[4775]: I1002 02:09:33.775444 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:09:33 crc kubenswrapper[4775]: E1002 02:09:33.776564 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:09:45 crc kubenswrapper[4775]: I1002 02:09:45.765342 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:09:45 crc kubenswrapper[4775]: E1002 02:09:45.766325 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:09:59 crc kubenswrapper[4775]: I1002 02:09:59.765845 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:09:59 crc kubenswrapper[4775]: E1002 02:09:59.766807 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:10:08 crc kubenswrapper[4775]: I1002 02:10:08.452229 4775 scope.go:117] "RemoveContainer" containerID="a430bd7d375548fc7aa2abbd9b702aea48844495f11cfa9ca7aff1e02c52b8b9" Oct 02 02:10:08 crc kubenswrapper[4775]: I1002 02:10:08.489239 4775 scope.go:117] "RemoveContainer" containerID="ddabebd6b3c747fa7b6ca77959080b27a169f7f5ace3d87005ed768fd6e84492" Oct 02 02:10:08 crc kubenswrapper[4775]: I1002 02:10:08.544064 4775 scope.go:117] "RemoveContainer" containerID="cda556d95d519500729d0be04b15c0114590b6b3efb514dfac80d925c08696b9" Oct 02 02:10:08 crc kubenswrapper[4775]: I1002 02:10:08.563557 4775 scope.go:117] "RemoveContainer" containerID="4112b6fd527ff9fca7a2399175d5a6860d1e392cc729ed9cbc4ea16d1712bb78" Oct 02 02:10:14 crc kubenswrapper[4775]: I1002 02:10:14.765107 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:10:14 crc kubenswrapper[4775]: E1002 02:10:14.766874 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:10:26 crc kubenswrapper[4775]: I1002 02:10:26.766190 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:10:26 crc kubenswrapper[4775]: E1002 02:10:26.766933 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:10:38 crc kubenswrapper[4775]: I1002 02:10:38.765801 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:10:38 crc kubenswrapper[4775]: E1002 02:10:38.766985 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:10:52 crc kubenswrapper[4775]: I1002 02:10:52.765901 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:10:52 crc kubenswrapper[4775]: E1002 02:10:52.767106 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:11:05 crc kubenswrapper[4775]: I1002 02:11:05.766125 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:11:05 crc kubenswrapper[4775]: E1002 02:11:05.767135 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:11:18 crc kubenswrapper[4775]: I1002 02:11:18.765735 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:11:18 crc kubenswrapper[4775]: E1002 02:11:18.767048 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:11:31 crc kubenswrapper[4775]: I1002 02:11:31.765735 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:11:31 crc kubenswrapper[4775]: E1002 02:11:31.766541 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:11:43 crc kubenswrapper[4775]: I1002 02:11:43.772798 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:11:44 crc kubenswrapper[4775]: I1002 02:11:44.943522 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"d47c09b02159d79a7c305b57f9ace268d1c9df5d8b7592ff15343a10d7d9bf60"} Oct 02 02:14:07 crc kubenswrapper[4775]: I1002 02:14:07.233203 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:14:07 crc kubenswrapper[4775]: I1002 02:14:07.233890 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:14:37 crc kubenswrapper[4775]: I1002 02:14:37.234342 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:14:37 crc kubenswrapper[4775]: I1002 02:14:37.235165 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.169787 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8"] Oct 02 02:15:00 crc kubenswrapper[4775]: E1002 02:15:00.170590 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fe64b05-e99c-4bab-9fb4-124e7abc02e1" containerName="registry-server" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.170604 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fe64b05-e99c-4bab-9fb4-124e7abc02e1" containerName="registry-server" Oct 02 02:15:00 crc kubenswrapper[4775]: E1002 02:15:00.170626 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fe64b05-e99c-4bab-9fb4-124e7abc02e1" containerName="extract-content" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.170634 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fe64b05-e99c-4bab-9fb4-124e7abc02e1" containerName="extract-content" Oct 02 02:15:00 crc kubenswrapper[4775]: E1002 02:15:00.170647 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fe64b05-e99c-4bab-9fb4-124e7abc02e1" containerName="extract-utilities" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.170655 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fe64b05-e99c-4bab-9fb4-124e7abc02e1" containerName="extract-utilities" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.170860 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fe64b05-e99c-4bab-9fb4-124e7abc02e1" containerName="registry-server" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.171417 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.174299 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.177188 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.188694 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0add0bd8-0890-44ee-baf8-0ff973140d3d-config-volume\") pod \"collect-profiles-29322855-nxwc8\" (UID: \"0add0bd8-0890-44ee-baf8-0ff973140d3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.188888 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttkfq\" (UniqueName: \"kubernetes.io/projected/0add0bd8-0890-44ee-baf8-0ff973140d3d-kube-api-access-ttkfq\") pod \"collect-profiles-29322855-nxwc8\" (UID: \"0add0bd8-0890-44ee-baf8-0ff973140d3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.188925 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8"] Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.189160 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0add0bd8-0890-44ee-baf8-0ff973140d3d-secret-volume\") pod \"collect-profiles-29322855-nxwc8\" (UID: \"0add0bd8-0890-44ee-baf8-0ff973140d3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.290077 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0add0bd8-0890-44ee-baf8-0ff973140d3d-secret-volume\") pod \"collect-profiles-29322855-nxwc8\" (UID: \"0add0bd8-0890-44ee-baf8-0ff973140d3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.290167 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0add0bd8-0890-44ee-baf8-0ff973140d3d-config-volume\") pod \"collect-profiles-29322855-nxwc8\" (UID: \"0add0bd8-0890-44ee-baf8-0ff973140d3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.290194 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttkfq\" (UniqueName: \"kubernetes.io/projected/0add0bd8-0890-44ee-baf8-0ff973140d3d-kube-api-access-ttkfq\") pod \"collect-profiles-29322855-nxwc8\" (UID: \"0add0bd8-0890-44ee-baf8-0ff973140d3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.291922 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0add0bd8-0890-44ee-baf8-0ff973140d3d-config-volume\") pod \"collect-profiles-29322855-nxwc8\" (UID: \"0add0bd8-0890-44ee-baf8-0ff973140d3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.314389 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0add0bd8-0890-44ee-baf8-0ff973140d3d-secret-volume\") pod \"collect-profiles-29322855-nxwc8\" (UID: \"0add0bd8-0890-44ee-baf8-0ff973140d3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.318661 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttkfq\" (UniqueName: \"kubernetes.io/projected/0add0bd8-0890-44ee-baf8-0ff973140d3d-kube-api-access-ttkfq\") pod \"collect-profiles-29322855-nxwc8\" (UID: \"0add0bd8-0890-44ee-baf8-0ff973140d3d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.494277 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" Oct 02 02:15:00 crc kubenswrapper[4775]: I1002 02:15:00.971526 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8"] Oct 02 02:15:01 crc kubenswrapper[4775]: I1002 02:15:01.877174 4775 generic.go:334] "Generic (PLEG): container finished" podID="0add0bd8-0890-44ee-baf8-0ff973140d3d" containerID="a84d3b997d8bde10f020c0eb5e84422a2906a19876e9708c1eeebb6b847d644c" exitCode=0 Oct 02 02:15:01 crc kubenswrapper[4775]: I1002 02:15:01.877280 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" event={"ID":"0add0bd8-0890-44ee-baf8-0ff973140d3d","Type":"ContainerDied","Data":"a84d3b997d8bde10f020c0eb5e84422a2906a19876e9708c1eeebb6b847d644c"} Oct 02 02:15:01 crc kubenswrapper[4775]: I1002 02:15:01.877509 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" event={"ID":"0add0bd8-0890-44ee-baf8-0ff973140d3d","Type":"ContainerStarted","Data":"d27c3b7e2828e1a90fa267ac9fd4f6d78b277192f90a748495d701c0360a93f0"} Oct 02 02:15:03 crc kubenswrapper[4775]: I1002 02:15:03.258853 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" Oct 02 02:15:03 crc kubenswrapper[4775]: I1002 02:15:03.342422 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0add0bd8-0890-44ee-baf8-0ff973140d3d-config-volume\") pod \"0add0bd8-0890-44ee-baf8-0ff973140d3d\" (UID: \"0add0bd8-0890-44ee-baf8-0ff973140d3d\") " Oct 02 02:15:03 crc kubenswrapper[4775]: I1002 02:15:03.342564 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0add0bd8-0890-44ee-baf8-0ff973140d3d-secret-volume\") pod \"0add0bd8-0890-44ee-baf8-0ff973140d3d\" (UID: \"0add0bd8-0890-44ee-baf8-0ff973140d3d\") " Oct 02 02:15:03 crc kubenswrapper[4775]: I1002 02:15:03.342640 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttkfq\" (UniqueName: \"kubernetes.io/projected/0add0bd8-0890-44ee-baf8-0ff973140d3d-kube-api-access-ttkfq\") pod \"0add0bd8-0890-44ee-baf8-0ff973140d3d\" (UID: \"0add0bd8-0890-44ee-baf8-0ff973140d3d\") " Oct 02 02:15:03 crc kubenswrapper[4775]: I1002 02:15:03.343565 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0add0bd8-0890-44ee-baf8-0ff973140d3d-config-volume" (OuterVolumeSpecName: "config-volume") pod "0add0bd8-0890-44ee-baf8-0ff973140d3d" (UID: "0add0bd8-0890-44ee-baf8-0ff973140d3d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:15:03 crc kubenswrapper[4775]: I1002 02:15:03.349565 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0add0bd8-0890-44ee-baf8-0ff973140d3d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0add0bd8-0890-44ee-baf8-0ff973140d3d" (UID: "0add0bd8-0890-44ee-baf8-0ff973140d3d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:15:03 crc kubenswrapper[4775]: I1002 02:15:03.351172 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0add0bd8-0890-44ee-baf8-0ff973140d3d-kube-api-access-ttkfq" (OuterVolumeSpecName: "kube-api-access-ttkfq") pod "0add0bd8-0890-44ee-baf8-0ff973140d3d" (UID: "0add0bd8-0890-44ee-baf8-0ff973140d3d"). InnerVolumeSpecName "kube-api-access-ttkfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:15:03 crc kubenswrapper[4775]: I1002 02:15:03.445359 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0add0bd8-0890-44ee-baf8-0ff973140d3d-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:03 crc kubenswrapper[4775]: I1002 02:15:03.445417 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttkfq\" (UniqueName: \"kubernetes.io/projected/0add0bd8-0890-44ee-baf8-0ff973140d3d-kube-api-access-ttkfq\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:03 crc kubenswrapper[4775]: I1002 02:15:03.445437 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0add0bd8-0890-44ee-baf8-0ff973140d3d-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:03 crc kubenswrapper[4775]: I1002 02:15:03.898453 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" event={"ID":"0add0bd8-0890-44ee-baf8-0ff973140d3d","Type":"ContainerDied","Data":"d27c3b7e2828e1a90fa267ac9fd4f6d78b277192f90a748495d701c0360a93f0"} Oct 02 02:15:03 crc kubenswrapper[4775]: I1002 02:15:03.898512 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d27c3b7e2828e1a90fa267ac9fd4f6d78b277192f90a748495d701c0360a93f0" Oct 02 02:15:03 crc kubenswrapper[4775]: I1002 02:15:03.898537 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8" Oct 02 02:15:04 crc kubenswrapper[4775]: I1002 02:15:04.362846 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj"] Oct 02 02:15:04 crc kubenswrapper[4775]: I1002 02:15:04.369234 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322810-q7gnj"] Oct 02 02:15:05 crc kubenswrapper[4775]: I1002 02:15:05.781872 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9ff5f53-63d9-47b2-b1ae-703e32c35ab1" path="/var/lib/kubelet/pods/b9ff5f53-63d9-47b2-b1ae-703e32c35ab1/volumes" Oct 02 02:15:07 crc kubenswrapper[4775]: I1002 02:15:07.236777 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:15:07 crc kubenswrapper[4775]: I1002 02:15:07.237355 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:15:07 crc kubenswrapper[4775]: I1002 02:15:07.237426 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 02:15:07 crc kubenswrapper[4775]: I1002 02:15:07.238666 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d47c09b02159d79a7c305b57f9ace268d1c9df5d8b7592ff15343a10d7d9bf60"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:15:07 crc kubenswrapper[4775]: I1002 02:15:07.238758 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://d47c09b02159d79a7c305b57f9ace268d1c9df5d8b7592ff15343a10d7d9bf60" gracePeriod=600 Oct 02 02:15:07 crc kubenswrapper[4775]: I1002 02:15:07.935872 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="d47c09b02159d79a7c305b57f9ace268d1c9df5d8b7592ff15343a10d7d9bf60" exitCode=0 Oct 02 02:15:07 crc kubenswrapper[4775]: I1002 02:15:07.935995 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"d47c09b02159d79a7c305b57f9ace268d1c9df5d8b7592ff15343a10d7d9bf60"} Oct 02 02:15:07 crc kubenswrapper[4775]: I1002 02:15:07.936241 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317"} Oct 02 02:15:07 crc kubenswrapper[4775]: I1002 02:15:07.936277 4775 scope.go:117] "RemoveContainer" containerID="92178009cc317d78436c605c22e47e833ccb032740e090e8f0e12728abb68b99" Oct 02 02:15:08 crc kubenswrapper[4775]: I1002 02:15:08.783484 4775 scope.go:117] "RemoveContainer" containerID="38e75b293fb6affc2f95a7680ce1755481bcef32a319098bd8b55afb6cb27312" Oct 02 02:15:12 crc kubenswrapper[4775]: I1002 02:15:12.792625 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6njs7"] Oct 02 02:15:12 crc kubenswrapper[4775]: E1002 02:15:12.793898 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0add0bd8-0890-44ee-baf8-0ff973140d3d" containerName="collect-profiles" Oct 02 02:15:12 crc kubenswrapper[4775]: I1002 02:15:12.793923 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0add0bd8-0890-44ee-baf8-0ff973140d3d" containerName="collect-profiles" Oct 02 02:15:12 crc kubenswrapper[4775]: I1002 02:15:12.794259 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0add0bd8-0890-44ee-baf8-0ff973140d3d" containerName="collect-profiles" Oct 02 02:15:12 crc kubenswrapper[4775]: I1002 02:15:12.798195 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:12 crc kubenswrapper[4775]: I1002 02:15:12.810982 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6njs7"] Oct 02 02:15:12 crc kubenswrapper[4775]: I1002 02:15:12.891601 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce40618-3663-46ed-b9b0-9ff3d4a12363-utilities\") pod \"redhat-operators-6njs7\" (UID: \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\") " pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:12 crc kubenswrapper[4775]: I1002 02:15:12.892097 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5zsp\" (UniqueName: \"kubernetes.io/projected/7ce40618-3663-46ed-b9b0-9ff3d4a12363-kube-api-access-h5zsp\") pod \"redhat-operators-6njs7\" (UID: \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\") " pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:12 crc kubenswrapper[4775]: I1002 02:15:12.892440 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce40618-3663-46ed-b9b0-9ff3d4a12363-catalog-content\") pod \"redhat-operators-6njs7\" (UID: \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\") " pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:12 crc kubenswrapper[4775]: I1002 02:15:12.993801 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce40618-3663-46ed-b9b0-9ff3d4a12363-catalog-content\") pod \"redhat-operators-6njs7\" (UID: \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\") " pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:12 crc kubenswrapper[4775]: I1002 02:15:12.993904 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce40618-3663-46ed-b9b0-9ff3d4a12363-utilities\") pod \"redhat-operators-6njs7\" (UID: \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\") " pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:12 crc kubenswrapper[4775]: I1002 02:15:12.993942 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5zsp\" (UniqueName: \"kubernetes.io/projected/7ce40618-3663-46ed-b9b0-9ff3d4a12363-kube-api-access-h5zsp\") pod \"redhat-operators-6njs7\" (UID: \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\") " pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:12 crc kubenswrapper[4775]: I1002 02:15:12.994494 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce40618-3663-46ed-b9b0-9ff3d4a12363-catalog-content\") pod \"redhat-operators-6njs7\" (UID: \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\") " pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:12 crc kubenswrapper[4775]: I1002 02:15:12.994662 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce40618-3663-46ed-b9b0-9ff3d4a12363-utilities\") pod \"redhat-operators-6njs7\" (UID: \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\") " pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:13 crc kubenswrapper[4775]: I1002 02:15:13.018713 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5zsp\" (UniqueName: \"kubernetes.io/projected/7ce40618-3663-46ed-b9b0-9ff3d4a12363-kube-api-access-h5zsp\") pod \"redhat-operators-6njs7\" (UID: \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\") " pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:13 crc kubenswrapper[4775]: I1002 02:15:13.134287 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:13 crc kubenswrapper[4775]: I1002 02:15:13.560355 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6njs7"] Oct 02 02:15:13 crc kubenswrapper[4775]: I1002 02:15:13.997248 4775 generic.go:334] "Generic (PLEG): container finished" podID="7ce40618-3663-46ed-b9b0-9ff3d4a12363" containerID="3e6400a5230ed2cc7eb22feb6dfa0dfe764bcc2bcec2fa578162cac0de62b55f" exitCode=0 Oct 02 02:15:13 crc kubenswrapper[4775]: I1002 02:15:13.997301 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6njs7" event={"ID":"7ce40618-3663-46ed-b9b0-9ff3d4a12363","Type":"ContainerDied","Data":"3e6400a5230ed2cc7eb22feb6dfa0dfe764bcc2bcec2fa578162cac0de62b55f"} Oct 02 02:15:13 crc kubenswrapper[4775]: I1002 02:15:13.997451 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6njs7" event={"ID":"7ce40618-3663-46ed-b9b0-9ff3d4a12363","Type":"ContainerStarted","Data":"d9ae264e3f048dc78b7943ffaec58e9254846d13709d248ec8500a4ed64bedea"} Oct 02 02:15:14 crc kubenswrapper[4775]: I1002 02:15:14.000136 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 02:15:16 crc kubenswrapper[4775]: I1002 02:15:16.017848 4775 generic.go:334] "Generic (PLEG): container finished" podID="7ce40618-3663-46ed-b9b0-9ff3d4a12363" containerID="6f104b4a0015d28c205a6ff3688feaf5d767ae12b63c51bddb7230a2683ce8a8" exitCode=0 Oct 02 02:15:16 crc kubenswrapper[4775]: I1002 02:15:16.017913 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6njs7" event={"ID":"7ce40618-3663-46ed-b9b0-9ff3d4a12363","Type":"ContainerDied","Data":"6f104b4a0015d28c205a6ff3688feaf5d767ae12b63c51bddb7230a2683ce8a8"} Oct 02 02:15:17 crc kubenswrapper[4775]: I1002 02:15:17.035156 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6njs7" event={"ID":"7ce40618-3663-46ed-b9b0-9ff3d4a12363","Type":"ContainerStarted","Data":"fb4dfa0358e4513e5b63456ebedca493f7150beabfd6e77882ecf331202526f5"} Oct 02 02:15:17 crc kubenswrapper[4775]: I1002 02:15:17.069493 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6njs7" podStartSLOduration=2.617170874 podStartE2EDuration="5.069459506s" podCreationTimestamp="2025-10-02 02:15:12 +0000 UTC" firstStartedPulling="2025-10-02 02:15:13.999820383 +0000 UTC m=+2051.166564433" lastFinishedPulling="2025-10-02 02:15:16.452108995 +0000 UTC m=+2053.618853065" observedRunningTime="2025-10-02 02:15:17.061752038 +0000 UTC m=+2054.228496128" watchObservedRunningTime="2025-10-02 02:15:17.069459506 +0000 UTC m=+2054.236203576" Oct 02 02:15:23 crc kubenswrapper[4775]: I1002 02:15:23.134741 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:23 crc kubenswrapper[4775]: I1002 02:15:23.135643 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:23 crc kubenswrapper[4775]: I1002 02:15:23.215220 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:24 crc kubenswrapper[4775]: I1002 02:15:24.174594 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:24 crc kubenswrapper[4775]: I1002 02:15:24.233690 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6njs7"] Oct 02 02:15:26 crc kubenswrapper[4775]: I1002 02:15:26.133987 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6njs7" podUID="7ce40618-3663-46ed-b9b0-9ff3d4a12363" containerName="registry-server" containerID="cri-o://fb4dfa0358e4513e5b63456ebedca493f7150beabfd6e77882ecf331202526f5" gracePeriod=2 Oct 02 02:15:28 crc kubenswrapper[4775]: I1002 02:15:28.166904 4775 generic.go:334] "Generic (PLEG): container finished" podID="7ce40618-3663-46ed-b9b0-9ff3d4a12363" containerID="fb4dfa0358e4513e5b63456ebedca493f7150beabfd6e77882ecf331202526f5" exitCode=0 Oct 02 02:15:28 crc kubenswrapper[4775]: I1002 02:15:28.167037 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6njs7" event={"ID":"7ce40618-3663-46ed-b9b0-9ff3d4a12363","Type":"ContainerDied","Data":"fb4dfa0358e4513e5b63456ebedca493f7150beabfd6e77882ecf331202526f5"} Oct 02 02:15:28 crc kubenswrapper[4775]: I1002 02:15:28.474707 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:28 crc kubenswrapper[4775]: I1002 02:15:28.650633 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce40618-3663-46ed-b9b0-9ff3d4a12363-utilities\") pod \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\" (UID: \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\") " Oct 02 02:15:28 crc kubenswrapper[4775]: I1002 02:15:28.651074 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5zsp\" (UniqueName: \"kubernetes.io/projected/7ce40618-3663-46ed-b9b0-9ff3d4a12363-kube-api-access-h5zsp\") pod \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\" (UID: \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\") " Oct 02 02:15:28 crc kubenswrapper[4775]: I1002 02:15:28.651125 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce40618-3663-46ed-b9b0-9ff3d4a12363-catalog-content\") pod \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\" (UID: \"7ce40618-3663-46ed-b9b0-9ff3d4a12363\") " Oct 02 02:15:28 crc kubenswrapper[4775]: I1002 02:15:28.652265 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ce40618-3663-46ed-b9b0-9ff3d4a12363-utilities" (OuterVolumeSpecName: "utilities") pod "7ce40618-3663-46ed-b9b0-9ff3d4a12363" (UID: "7ce40618-3663-46ed-b9b0-9ff3d4a12363"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:15:28 crc kubenswrapper[4775]: I1002 02:15:28.660244 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ce40618-3663-46ed-b9b0-9ff3d4a12363-kube-api-access-h5zsp" (OuterVolumeSpecName: "kube-api-access-h5zsp") pod "7ce40618-3663-46ed-b9b0-9ff3d4a12363" (UID: "7ce40618-3663-46ed-b9b0-9ff3d4a12363"). InnerVolumeSpecName "kube-api-access-h5zsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:15:28 crc kubenswrapper[4775]: I1002 02:15:28.738174 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ce40618-3663-46ed-b9b0-9ff3d4a12363-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ce40618-3663-46ed-b9b0-9ff3d4a12363" (UID: "7ce40618-3663-46ed-b9b0-9ff3d4a12363"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:15:28 crc kubenswrapper[4775]: I1002 02:15:28.752404 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5zsp\" (UniqueName: \"kubernetes.io/projected/7ce40618-3663-46ed-b9b0-9ff3d4a12363-kube-api-access-h5zsp\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:28 crc kubenswrapper[4775]: I1002 02:15:28.752451 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce40618-3663-46ed-b9b0-9ff3d4a12363-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:28 crc kubenswrapper[4775]: I1002 02:15:28.752474 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce40618-3663-46ed-b9b0-9ff3d4a12363-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:29 crc kubenswrapper[4775]: I1002 02:15:29.182306 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6njs7" event={"ID":"7ce40618-3663-46ed-b9b0-9ff3d4a12363","Type":"ContainerDied","Data":"d9ae264e3f048dc78b7943ffaec58e9254846d13709d248ec8500a4ed64bedea"} Oct 02 02:15:29 crc kubenswrapper[4775]: I1002 02:15:29.182382 4775 scope.go:117] "RemoveContainer" containerID="fb4dfa0358e4513e5b63456ebedca493f7150beabfd6e77882ecf331202526f5" Oct 02 02:15:29 crc kubenswrapper[4775]: I1002 02:15:29.182569 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6njs7" Oct 02 02:15:29 crc kubenswrapper[4775]: I1002 02:15:29.220261 4775 scope.go:117] "RemoveContainer" containerID="6f104b4a0015d28c205a6ff3688feaf5d767ae12b63c51bddb7230a2683ce8a8" Oct 02 02:15:29 crc kubenswrapper[4775]: I1002 02:15:29.232283 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6njs7"] Oct 02 02:15:29 crc kubenswrapper[4775]: I1002 02:15:29.236436 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6njs7"] Oct 02 02:15:29 crc kubenswrapper[4775]: I1002 02:15:29.273426 4775 scope.go:117] "RemoveContainer" containerID="3e6400a5230ed2cc7eb22feb6dfa0dfe764bcc2bcec2fa578162cac0de62b55f" Oct 02 02:15:29 crc kubenswrapper[4775]: I1002 02:15:29.780998 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ce40618-3663-46ed-b9b0-9ff3d4a12363" path="/var/lib/kubelet/pods/7ce40618-3663-46ed-b9b0-9ff3d4a12363/volumes" Oct 02 02:15:45 crc kubenswrapper[4775]: I1002 02:15:45.976432 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7g4r4"] Oct 02 02:15:45 crc kubenswrapper[4775]: E1002 02:15:45.979015 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce40618-3663-46ed-b9b0-9ff3d4a12363" containerName="extract-content" Oct 02 02:15:45 crc kubenswrapper[4775]: I1002 02:15:45.979044 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce40618-3663-46ed-b9b0-9ff3d4a12363" containerName="extract-content" Oct 02 02:15:45 crc kubenswrapper[4775]: E1002 02:15:45.979072 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce40618-3663-46ed-b9b0-9ff3d4a12363" containerName="registry-server" Oct 02 02:15:45 crc kubenswrapper[4775]: I1002 02:15:45.979087 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce40618-3663-46ed-b9b0-9ff3d4a12363" containerName="registry-server" Oct 02 02:15:45 crc kubenswrapper[4775]: E1002 02:15:45.979108 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce40618-3663-46ed-b9b0-9ff3d4a12363" containerName="extract-utilities" Oct 02 02:15:45 crc kubenswrapper[4775]: I1002 02:15:45.979121 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce40618-3663-46ed-b9b0-9ff3d4a12363" containerName="extract-utilities" Oct 02 02:15:45 crc kubenswrapper[4775]: I1002 02:15:45.979420 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce40618-3663-46ed-b9b0-9ff3d4a12363" containerName="registry-server" Oct 02 02:15:45 crc kubenswrapper[4775]: I1002 02:15:45.981651 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:45 crc kubenswrapper[4775]: I1002 02:15:45.997012 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7g4r4"] Oct 02 02:15:46 crc kubenswrapper[4775]: I1002 02:15:46.026930 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11366f75-c06f-4201-8f34-22ff54619310-catalog-content\") pod \"community-operators-7g4r4\" (UID: \"11366f75-c06f-4201-8f34-22ff54619310\") " pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:46 crc kubenswrapper[4775]: I1002 02:15:46.027041 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11366f75-c06f-4201-8f34-22ff54619310-utilities\") pod \"community-operators-7g4r4\" (UID: \"11366f75-c06f-4201-8f34-22ff54619310\") " pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:46 crc kubenswrapper[4775]: I1002 02:15:46.027084 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8bh7\" (UniqueName: \"kubernetes.io/projected/11366f75-c06f-4201-8f34-22ff54619310-kube-api-access-x8bh7\") pod \"community-operators-7g4r4\" (UID: \"11366f75-c06f-4201-8f34-22ff54619310\") " pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:46 crc kubenswrapper[4775]: I1002 02:15:46.128267 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11366f75-c06f-4201-8f34-22ff54619310-catalog-content\") pod \"community-operators-7g4r4\" (UID: \"11366f75-c06f-4201-8f34-22ff54619310\") " pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:46 crc kubenswrapper[4775]: I1002 02:15:46.128890 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11366f75-c06f-4201-8f34-22ff54619310-utilities\") pod \"community-operators-7g4r4\" (UID: \"11366f75-c06f-4201-8f34-22ff54619310\") " pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:46 crc kubenswrapper[4775]: I1002 02:15:46.128978 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8bh7\" (UniqueName: \"kubernetes.io/projected/11366f75-c06f-4201-8f34-22ff54619310-kube-api-access-x8bh7\") pod \"community-operators-7g4r4\" (UID: \"11366f75-c06f-4201-8f34-22ff54619310\") " pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:46 crc kubenswrapper[4775]: I1002 02:15:46.128816 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11366f75-c06f-4201-8f34-22ff54619310-catalog-content\") pod \"community-operators-7g4r4\" (UID: \"11366f75-c06f-4201-8f34-22ff54619310\") " pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:46 crc kubenswrapper[4775]: I1002 02:15:46.129561 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11366f75-c06f-4201-8f34-22ff54619310-utilities\") pod \"community-operators-7g4r4\" (UID: \"11366f75-c06f-4201-8f34-22ff54619310\") " pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:46 crc kubenswrapper[4775]: I1002 02:15:46.148031 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8bh7\" (UniqueName: \"kubernetes.io/projected/11366f75-c06f-4201-8f34-22ff54619310-kube-api-access-x8bh7\") pod \"community-operators-7g4r4\" (UID: \"11366f75-c06f-4201-8f34-22ff54619310\") " pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:46 crc kubenswrapper[4775]: I1002 02:15:46.322095 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:46 crc kubenswrapper[4775]: I1002 02:15:46.697780 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7g4r4"] Oct 02 02:15:47 crc kubenswrapper[4775]: I1002 02:15:47.356296 4775 generic.go:334] "Generic (PLEG): container finished" podID="11366f75-c06f-4201-8f34-22ff54619310" containerID="602b5a86000fb157f7a88b8ab29ec923e8709951baddbd705845d06e5343639f" exitCode=0 Oct 02 02:15:47 crc kubenswrapper[4775]: I1002 02:15:47.356384 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7g4r4" event={"ID":"11366f75-c06f-4201-8f34-22ff54619310","Type":"ContainerDied","Data":"602b5a86000fb157f7a88b8ab29ec923e8709951baddbd705845d06e5343639f"} Oct 02 02:15:47 crc kubenswrapper[4775]: I1002 02:15:47.356625 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7g4r4" event={"ID":"11366f75-c06f-4201-8f34-22ff54619310","Type":"ContainerStarted","Data":"bc3ad2698b52df131fac28eab6d7f5f4dfd0b9abf6d0b383172674b681589f58"} Oct 02 02:15:48 crc kubenswrapper[4775]: I1002 02:15:48.386535 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7g4r4" event={"ID":"11366f75-c06f-4201-8f34-22ff54619310","Type":"ContainerStarted","Data":"1c7ff21cb0ebf9ee1fa393e4be765dd38c187a5163be8dec962dc6a4b130cd6c"} Oct 02 02:15:49 crc kubenswrapper[4775]: I1002 02:15:49.397123 4775 generic.go:334] "Generic (PLEG): container finished" podID="11366f75-c06f-4201-8f34-22ff54619310" containerID="1c7ff21cb0ebf9ee1fa393e4be765dd38c187a5163be8dec962dc6a4b130cd6c" exitCode=0 Oct 02 02:15:49 crc kubenswrapper[4775]: I1002 02:15:49.397185 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7g4r4" event={"ID":"11366f75-c06f-4201-8f34-22ff54619310","Type":"ContainerDied","Data":"1c7ff21cb0ebf9ee1fa393e4be765dd38c187a5163be8dec962dc6a4b130cd6c"} Oct 02 02:15:50 crc kubenswrapper[4775]: I1002 02:15:50.409349 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7g4r4" event={"ID":"11366f75-c06f-4201-8f34-22ff54619310","Type":"ContainerStarted","Data":"c3aa4fb32babd1bc61c9a9a42e0f605be20b6c827384656d2dcd6c42c82e756b"} Oct 02 02:15:50 crc kubenswrapper[4775]: I1002 02:15:50.438676 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7g4r4" podStartSLOduration=2.995713186 podStartE2EDuration="5.438661218s" podCreationTimestamp="2025-10-02 02:15:45 +0000 UTC" firstStartedPulling="2025-10-02 02:15:47.358561477 +0000 UTC m=+2084.525305517" lastFinishedPulling="2025-10-02 02:15:49.801509469 +0000 UTC m=+2086.968253549" observedRunningTime="2025-10-02 02:15:50.436974085 +0000 UTC m=+2087.603718145" watchObservedRunningTime="2025-10-02 02:15:50.438661218 +0000 UTC m=+2087.605405258" Oct 02 02:15:56 crc kubenswrapper[4775]: I1002 02:15:56.323438 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:56 crc kubenswrapper[4775]: I1002 02:15:56.323857 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:56 crc kubenswrapper[4775]: I1002 02:15:56.388453 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:56 crc kubenswrapper[4775]: I1002 02:15:56.519839 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:56 crc kubenswrapper[4775]: I1002 02:15:56.648769 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7g4r4"] Oct 02 02:15:58 crc kubenswrapper[4775]: I1002 02:15:58.480348 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7g4r4" podUID="11366f75-c06f-4201-8f34-22ff54619310" containerName="registry-server" containerID="cri-o://c3aa4fb32babd1bc61c9a9a42e0f605be20b6c827384656d2dcd6c42c82e756b" gracePeriod=2 Oct 02 02:15:59 crc kubenswrapper[4775]: I1002 02:15:59.493713 4775 generic.go:334] "Generic (PLEG): container finished" podID="11366f75-c06f-4201-8f34-22ff54619310" containerID="c3aa4fb32babd1bc61c9a9a42e0f605be20b6c827384656d2dcd6c42c82e756b" exitCode=0 Oct 02 02:15:59 crc kubenswrapper[4775]: I1002 02:15:59.493779 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7g4r4" event={"ID":"11366f75-c06f-4201-8f34-22ff54619310","Type":"ContainerDied","Data":"c3aa4fb32babd1bc61c9a9a42e0f605be20b6c827384656d2dcd6c42c82e756b"} Oct 02 02:15:59 crc kubenswrapper[4775]: I1002 02:15:59.591207 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:15:59 crc kubenswrapper[4775]: I1002 02:15:59.753798 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11366f75-c06f-4201-8f34-22ff54619310-catalog-content\") pod \"11366f75-c06f-4201-8f34-22ff54619310\" (UID: \"11366f75-c06f-4201-8f34-22ff54619310\") " Oct 02 02:15:59 crc kubenswrapper[4775]: I1002 02:15:59.753868 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11366f75-c06f-4201-8f34-22ff54619310-utilities\") pod \"11366f75-c06f-4201-8f34-22ff54619310\" (UID: \"11366f75-c06f-4201-8f34-22ff54619310\") " Oct 02 02:15:59 crc kubenswrapper[4775]: I1002 02:15:59.753994 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8bh7\" (UniqueName: \"kubernetes.io/projected/11366f75-c06f-4201-8f34-22ff54619310-kube-api-access-x8bh7\") pod \"11366f75-c06f-4201-8f34-22ff54619310\" (UID: \"11366f75-c06f-4201-8f34-22ff54619310\") " Oct 02 02:15:59 crc kubenswrapper[4775]: I1002 02:15:59.754823 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11366f75-c06f-4201-8f34-22ff54619310-utilities" (OuterVolumeSpecName: "utilities") pod "11366f75-c06f-4201-8f34-22ff54619310" (UID: "11366f75-c06f-4201-8f34-22ff54619310"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:15:59 crc kubenswrapper[4775]: I1002 02:15:59.766050 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11366f75-c06f-4201-8f34-22ff54619310-kube-api-access-x8bh7" (OuterVolumeSpecName: "kube-api-access-x8bh7") pod "11366f75-c06f-4201-8f34-22ff54619310" (UID: "11366f75-c06f-4201-8f34-22ff54619310"). InnerVolumeSpecName "kube-api-access-x8bh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:15:59 crc kubenswrapper[4775]: I1002 02:15:59.801099 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11366f75-c06f-4201-8f34-22ff54619310-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "11366f75-c06f-4201-8f34-22ff54619310" (UID: "11366f75-c06f-4201-8f34-22ff54619310"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:15:59 crc kubenswrapper[4775]: I1002 02:15:59.855564 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/11366f75-c06f-4201-8f34-22ff54619310-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:59 crc kubenswrapper[4775]: I1002 02:15:59.855610 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/11366f75-c06f-4201-8f34-22ff54619310-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:15:59 crc kubenswrapper[4775]: I1002 02:15:59.855625 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8bh7\" (UniqueName: \"kubernetes.io/projected/11366f75-c06f-4201-8f34-22ff54619310-kube-api-access-x8bh7\") on node \"crc\" DevicePath \"\"" Oct 02 02:16:00 crc kubenswrapper[4775]: I1002 02:16:00.509839 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7g4r4" event={"ID":"11366f75-c06f-4201-8f34-22ff54619310","Type":"ContainerDied","Data":"bc3ad2698b52df131fac28eab6d7f5f4dfd0b9abf6d0b383172674b681589f58"} Oct 02 02:16:00 crc kubenswrapper[4775]: I1002 02:16:00.509911 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7g4r4" Oct 02 02:16:00 crc kubenswrapper[4775]: I1002 02:16:00.510405 4775 scope.go:117] "RemoveContainer" containerID="c3aa4fb32babd1bc61c9a9a42e0f605be20b6c827384656d2dcd6c42c82e756b" Oct 02 02:16:00 crc kubenswrapper[4775]: I1002 02:16:00.538668 4775 scope.go:117] "RemoveContainer" containerID="1c7ff21cb0ebf9ee1fa393e4be765dd38c187a5163be8dec962dc6a4b130cd6c" Oct 02 02:16:00 crc kubenswrapper[4775]: I1002 02:16:00.554045 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7g4r4"] Oct 02 02:16:00 crc kubenswrapper[4775]: I1002 02:16:00.563296 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7g4r4"] Oct 02 02:16:00 crc kubenswrapper[4775]: I1002 02:16:00.575758 4775 scope.go:117] "RemoveContainer" containerID="602b5a86000fb157f7a88b8ab29ec923e8709951baddbd705845d06e5343639f" Oct 02 02:16:01 crc kubenswrapper[4775]: I1002 02:16:01.776776 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11366f75-c06f-4201-8f34-22ff54619310" path="/var/lib/kubelet/pods/11366f75-c06f-4201-8f34-22ff54619310/volumes" Oct 02 02:17:07 crc kubenswrapper[4775]: I1002 02:17:07.234080 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:17:07 crc kubenswrapper[4775]: I1002 02:17:07.237051 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.480723 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f5xvq"] Oct 02 02:17:34 crc kubenswrapper[4775]: E1002 02:17:34.481379 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11366f75-c06f-4201-8f34-22ff54619310" containerName="extract-content" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.481390 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="11366f75-c06f-4201-8f34-22ff54619310" containerName="extract-content" Oct 02 02:17:34 crc kubenswrapper[4775]: E1002 02:17:34.481404 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11366f75-c06f-4201-8f34-22ff54619310" containerName="extract-utilities" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.481410 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="11366f75-c06f-4201-8f34-22ff54619310" containerName="extract-utilities" Oct 02 02:17:34 crc kubenswrapper[4775]: E1002 02:17:34.481434 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11366f75-c06f-4201-8f34-22ff54619310" containerName="registry-server" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.481440 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="11366f75-c06f-4201-8f34-22ff54619310" containerName="registry-server" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.481557 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="11366f75-c06f-4201-8f34-22ff54619310" containerName="registry-server" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.483116 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.558682 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5xvq"] Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.576433 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6346fe5-10bd-4669-b518-f8da4a156811-catalog-content\") pod \"redhat-marketplace-f5xvq\" (UID: \"f6346fe5-10bd-4669-b518-f8da4a156811\") " pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.576547 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6346fe5-10bd-4669-b518-f8da4a156811-utilities\") pod \"redhat-marketplace-f5xvq\" (UID: \"f6346fe5-10bd-4669-b518-f8da4a156811\") " pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.576656 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvfjk\" (UniqueName: \"kubernetes.io/projected/f6346fe5-10bd-4669-b518-f8da4a156811-kube-api-access-tvfjk\") pod \"redhat-marketplace-f5xvq\" (UID: \"f6346fe5-10bd-4669-b518-f8da4a156811\") " pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.677523 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6346fe5-10bd-4669-b518-f8da4a156811-catalog-content\") pod \"redhat-marketplace-f5xvq\" (UID: \"f6346fe5-10bd-4669-b518-f8da4a156811\") " pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.677585 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6346fe5-10bd-4669-b518-f8da4a156811-utilities\") pod \"redhat-marketplace-f5xvq\" (UID: \"f6346fe5-10bd-4669-b518-f8da4a156811\") " pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.677626 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvfjk\" (UniqueName: \"kubernetes.io/projected/f6346fe5-10bd-4669-b518-f8da4a156811-kube-api-access-tvfjk\") pod \"redhat-marketplace-f5xvq\" (UID: \"f6346fe5-10bd-4669-b518-f8da4a156811\") " pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.678299 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6346fe5-10bd-4669-b518-f8da4a156811-catalog-content\") pod \"redhat-marketplace-f5xvq\" (UID: \"f6346fe5-10bd-4669-b518-f8da4a156811\") " pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.678494 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6346fe5-10bd-4669-b518-f8da4a156811-utilities\") pod \"redhat-marketplace-f5xvq\" (UID: \"f6346fe5-10bd-4669-b518-f8da4a156811\") " pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.696692 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvfjk\" (UniqueName: \"kubernetes.io/projected/f6346fe5-10bd-4669-b518-f8da4a156811-kube-api-access-tvfjk\") pod \"redhat-marketplace-f5xvq\" (UID: \"f6346fe5-10bd-4669-b518-f8da4a156811\") " pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:34 crc kubenswrapper[4775]: I1002 02:17:34.821528 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:35 crc kubenswrapper[4775]: I1002 02:17:35.257560 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5xvq"] Oct 02 02:17:35 crc kubenswrapper[4775]: I1002 02:17:35.406797 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5xvq" event={"ID":"f6346fe5-10bd-4669-b518-f8da4a156811","Type":"ContainerStarted","Data":"0b83aae8c49e6976d76fdc91a1dec1fd60226d67b9b59b4ffdb9abcbbe564906"} Oct 02 02:17:36 crc kubenswrapper[4775]: I1002 02:17:36.418939 4775 generic.go:334] "Generic (PLEG): container finished" podID="f6346fe5-10bd-4669-b518-f8da4a156811" containerID="dfe9fc2d436ceee59db8d20782de230d3185c6bef73f3a4983f3b46c274801ce" exitCode=0 Oct 02 02:17:36 crc kubenswrapper[4775]: I1002 02:17:36.419095 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5xvq" event={"ID":"f6346fe5-10bd-4669-b518-f8da4a156811","Type":"ContainerDied","Data":"dfe9fc2d436ceee59db8d20782de230d3185c6bef73f3a4983f3b46c274801ce"} Oct 02 02:17:37 crc kubenswrapper[4775]: I1002 02:17:37.233486 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:17:37 crc kubenswrapper[4775]: I1002 02:17:37.233572 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:17:38 crc kubenswrapper[4775]: I1002 02:17:38.442116 4775 generic.go:334] "Generic (PLEG): container finished" podID="f6346fe5-10bd-4669-b518-f8da4a156811" containerID="2c46a2136cbe9b2d2ea4f364123a0276b38e8c3dd5024a463c8618cd67716d23" exitCode=0 Oct 02 02:17:38 crc kubenswrapper[4775]: I1002 02:17:38.442286 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5xvq" event={"ID":"f6346fe5-10bd-4669-b518-f8da4a156811","Type":"ContainerDied","Data":"2c46a2136cbe9b2d2ea4f364123a0276b38e8c3dd5024a463c8618cd67716d23"} Oct 02 02:17:39 crc kubenswrapper[4775]: I1002 02:17:39.455324 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5xvq" event={"ID":"f6346fe5-10bd-4669-b518-f8da4a156811","Type":"ContainerStarted","Data":"aead61afd74a2b243c34619ff1913049737847cfd1277a2a31c9e25bdcbd866b"} Oct 02 02:17:39 crc kubenswrapper[4775]: I1002 02:17:39.493249 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f5xvq" podStartSLOduration=3.077167038 podStartE2EDuration="5.49318017s" podCreationTimestamp="2025-10-02 02:17:34 +0000 UTC" firstStartedPulling="2025-10-02 02:17:36.422644692 +0000 UTC m=+2193.589388762" lastFinishedPulling="2025-10-02 02:17:38.838657814 +0000 UTC m=+2196.005401894" observedRunningTime="2025-10-02 02:17:39.483710637 +0000 UTC m=+2196.650454677" watchObservedRunningTime="2025-10-02 02:17:39.49318017 +0000 UTC m=+2196.659924240" Oct 02 02:17:44 crc kubenswrapper[4775]: I1002 02:17:44.822110 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:44 crc kubenswrapper[4775]: I1002 02:17:44.822887 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:44 crc kubenswrapper[4775]: I1002 02:17:44.902076 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:45 crc kubenswrapper[4775]: I1002 02:17:45.569931 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:45 crc kubenswrapper[4775]: I1002 02:17:45.636804 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5xvq"] Oct 02 02:17:47 crc kubenswrapper[4775]: I1002 02:17:47.525805 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f5xvq" podUID="f6346fe5-10bd-4669-b518-f8da4a156811" containerName="registry-server" containerID="cri-o://aead61afd74a2b243c34619ff1913049737847cfd1277a2a31c9e25bdcbd866b" gracePeriod=2 Oct 02 02:17:47 crc kubenswrapper[4775]: I1002 02:17:47.960796 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.099809 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvfjk\" (UniqueName: \"kubernetes.io/projected/f6346fe5-10bd-4669-b518-f8da4a156811-kube-api-access-tvfjk\") pod \"f6346fe5-10bd-4669-b518-f8da4a156811\" (UID: \"f6346fe5-10bd-4669-b518-f8da4a156811\") " Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.100005 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6346fe5-10bd-4669-b518-f8da4a156811-utilities\") pod \"f6346fe5-10bd-4669-b518-f8da4a156811\" (UID: \"f6346fe5-10bd-4669-b518-f8da4a156811\") " Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.100048 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6346fe5-10bd-4669-b518-f8da4a156811-catalog-content\") pod \"f6346fe5-10bd-4669-b518-f8da4a156811\" (UID: \"f6346fe5-10bd-4669-b518-f8da4a156811\") " Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.101678 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6346fe5-10bd-4669-b518-f8da4a156811-utilities" (OuterVolumeSpecName: "utilities") pod "f6346fe5-10bd-4669-b518-f8da4a156811" (UID: "f6346fe5-10bd-4669-b518-f8da4a156811"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.108270 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6346fe5-10bd-4669-b518-f8da4a156811-kube-api-access-tvfjk" (OuterVolumeSpecName: "kube-api-access-tvfjk") pod "f6346fe5-10bd-4669-b518-f8da4a156811" (UID: "f6346fe5-10bd-4669-b518-f8da4a156811"). InnerVolumeSpecName "kube-api-access-tvfjk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.120677 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6346fe5-10bd-4669-b518-f8da4a156811-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f6346fe5-10bd-4669-b518-f8da4a156811" (UID: "f6346fe5-10bd-4669-b518-f8da4a156811"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.201996 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvfjk\" (UniqueName: \"kubernetes.io/projected/f6346fe5-10bd-4669-b518-f8da4a156811-kube-api-access-tvfjk\") on node \"crc\" DevicePath \"\"" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.202061 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6346fe5-10bd-4669-b518-f8da4a156811-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.202074 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6346fe5-10bd-4669-b518-f8da4a156811-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.540037 4775 generic.go:334] "Generic (PLEG): container finished" podID="f6346fe5-10bd-4669-b518-f8da4a156811" containerID="aead61afd74a2b243c34619ff1913049737847cfd1277a2a31c9e25bdcbd866b" exitCode=0 Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.540107 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f5xvq" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.540102 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5xvq" event={"ID":"f6346fe5-10bd-4669-b518-f8da4a156811","Type":"ContainerDied","Data":"aead61afd74a2b243c34619ff1913049737847cfd1277a2a31c9e25bdcbd866b"} Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.540243 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f5xvq" event={"ID":"f6346fe5-10bd-4669-b518-f8da4a156811","Type":"ContainerDied","Data":"0b83aae8c49e6976d76fdc91a1dec1fd60226d67b9b59b4ffdb9abcbbe564906"} Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.540265 4775 scope.go:117] "RemoveContainer" containerID="aead61afd74a2b243c34619ff1913049737847cfd1277a2a31c9e25bdcbd866b" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.573415 4775 scope.go:117] "RemoveContainer" containerID="2c46a2136cbe9b2d2ea4f364123a0276b38e8c3dd5024a463c8618cd67716d23" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.587770 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5xvq"] Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.596055 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f5xvq"] Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.607339 4775 scope.go:117] "RemoveContainer" containerID="dfe9fc2d436ceee59db8d20782de230d3185c6bef73f3a4983f3b46c274801ce" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.645611 4775 scope.go:117] "RemoveContainer" containerID="aead61afd74a2b243c34619ff1913049737847cfd1277a2a31c9e25bdcbd866b" Oct 02 02:17:48 crc kubenswrapper[4775]: E1002 02:17:48.645977 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aead61afd74a2b243c34619ff1913049737847cfd1277a2a31c9e25bdcbd866b\": container with ID starting with aead61afd74a2b243c34619ff1913049737847cfd1277a2a31c9e25bdcbd866b not found: ID does not exist" containerID="aead61afd74a2b243c34619ff1913049737847cfd1277a2a31c9e25bdcbd866b" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.646019 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aead61afd74a2b243c34619ff1913049737847cfd1277a2a31c9e25bdcbd866b"} err="failed to get container status \"aead61afd74a2b243c34619ff1913049737847cfd1277a2a31c9e25bdcbd866b\": rpc error: code = NotFound desc = could not find container \"aead61afd74a2b243c34619ff1913049737847cfd1277a2a31c9e25bdcbd866b\": container with ID starting with aead61afd74a2b243c34619ff1913049737847cfd1277a2a31c9e25bdcbd866b not found: ID does not exist" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.646048 4775 scope.go:117] "RemoveContainer" containerID="2c46a2136cbe9b2d2ea4f364123a0276b38e8c3dd5024a463c8618cd67716d23" Oct 02 02:17:48 crc kubenswrapper[4775]: E1002 02:17:48.646410 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c46a2136cbe9b2d2ea4f364123a0276b38e8c3dd5024a463c8618cd67716d23\": container with ID starting with 2c46a2136cbe9b2d2ea4f364123a0276b38e8c3dd5024a463c8618cd67716d23 not found: ID does not exist" containerID="2c46a2136cbe9b2d2ea4f364123a0276b38e8c3dd5024a463c8618cd67716d23" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.646452 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c46a2136cbe9b2d2ea4f364123a0276b38e8c3dd5024a463c8618cd67716d23"} err="failed to get container status \"2c46a2136cbe9b2d2ea4f364123a0276b38e8c3dd5024a463c8618cd67716d23\": rpc error: code = NotFound desc = could not find container \"2c46a2136cbe9b2d2ea4f364123a0276b38e8c3dd5024a463c8618cd67716d23\": container with ID starting with 2c46a2136cbe9b2d2ea4f364123a0276b38e8c3dd5024a463c8618cd67716d23 not found: ID does not exist" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.646484 4775 scope.go:117] "RemoveContainer" containerID="dfe9fc2d436ceee59db8d20782de230d3185c6bef73f3a4983f3b46c274801ce" Oct 02 02:17:48 crc kubenswrapper[4775]: E1002 02:17:48.646723 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfe9fc2d436ceee59db8d20782de230d3185c6bef73f3a4983f3b46c274801ce\": container with ID starting with dfe9fc2d436ceee59db8d20782de230d3185c6bef73f3a4983f3b46c274801ce not found: ID does not exist" containerID="dfe9fc2d436ceee59db8d20782de230d3185c6bef73f3a4983f3b46c274801ce" Oct 02 02:17:48 crc kubenswrapper[4775]: I1002 02:17:48.646763 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfe9fc2d436ceee59db8d20782de230d3185c6bef73f3a4983f3b46c274801ce"} err="failed to get container status \"dfe9fc2d436ceee59db8d20782de230d3185c6bef73f3a4983f3b46c274801ce\": rpc error: code = NotFound desc = could not find container \"dfe9fc2d436ceee59db8d20782de230d3185c6bef73f3a4983f3b46c274801ce\": container with ID starting with dfe9fc2d436ceee59db8d20782de230d3185c6bef73f3a4983f3b46c274801ce not found: ID does not exist" Oct 02 02:17:49 crc kubenswrapper[4775]: I1002 02:17:49.794113 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6346fe5-10bd-4669-b518-f8da4a156811" path="/var/lib/kubelet/pods/f6346fe5-10bd-4669-b518-f8da4a156811/volumes" Oct 02 02:18:07 crc kubenswrapper[4775]: I1002 02:18:07.233629 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:18:07 crc kubenswrapper[4775]: I1002 02:18:07.234238 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:18:07 crc kubenswrapper[4775]: I1002 02:18:07.234295 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 02:18:07 crc kubenswrapper[4775]: I1002 02:18:07.235019 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:18:07 crc kubenswrapper[4775]: I1002 02:18:07.235143 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" gracePeriod=600 Oct 02 02:18:07 crc kubenswrapper[4775]: E1002 02:18:07.362430 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:18:07 crc kubenswrapper[4775]: I1002 02:18:07.722814 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" exitCode=0 Oct 02 02:18:07 crc kubenswrapper[4775]: I1002 02:18:07.722923 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317"} Oct 02 02:18:07 crc kubenswrapper[4775]: I1002 02:18:07.723297 4775 scope.go:117] "RemoveContainer" containerID="d47c09b02159d79a7c305b57f9ace268d1c9df5d8b7592ff15343a10d7d9bf60" Oct 02 02:18:07 crc kubenswrapper[4775]: I1002 02:18:07.724373 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:18:07 crc kubenswrapper[4775]: E1002 02:18:07.726287 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:18:19 crc kubenswrapper[4775]: I1002 02:18:19.765519 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:18:19 crc kubenswrapper[4775]: E1002 02:18:19.766359 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:18:34 crc kubenswrapper[4775]: I1002 02:18:34.766538 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:18:34 crc kubenswrapper[4775]: E1002 02:18:34.767563 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:18:45 crc kubenswrapper[4775]: I1002 02:18:45.765937 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:18:45 crc kubenswrapper[4775]: E1002 02:18:45.767008 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:18:58 crc kubenswrapper[4775]: I1002 02:18:58.765742 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:18:58 crc kubenswrapper[4775]: E1002 02:18:58.766510 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:19:13 crc kubenswrapper[4775]: I1002 02:19:13.773627 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:19:13 crc kubenswrapper[4775]: E1002 02:19:13.774523 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.656425 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kbx77"] Oct 02 02:19:25 crc kubenswrapper[4775]: E1002 02:19:25.657747 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6346fe5-10bd-4669-b518-f8da4a156811" containerName="extract-content" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.657773 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6346fe5-10bd-4669-b518-f8da4a156811" containerName="extract-content" Oct 02 02:19:25 crc kubenswrapper[4775]: E1002 02:19:25.657808 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6346fe5-10bd-4669-b518-f8da4a156811" containerName="extract-utilities" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.657823 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6346fe5-10bd-4669-b518-f8da4a156811" containerName="extract-utilities" Oct 02 02:19:25 crc kubenswrapper[4775]: E1002 02:19:25.657856 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6346fe5-10bd-4669-b518-f8da4a156811" containerName="registry-server" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.657874 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6346fe5-10bd-4669-b518-f8da4a156811" containerName="registry-server" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.658263 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6346fe5-10bd-4669-b518-f8da4a156811" containerName="registry-server" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.660407 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.678427 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kbx77"] Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.766078 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:19:25 crc kubenswrapper[4775]: E1002 02:19:25.766765 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.815061 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3103bdf-df5e-45c5-b453-90c903eb76ef-catalog-content\") pod \"certified-operators-kbx77\" (UID: \"b3103bdf-df5e-45c5-b453-90c903eb76ef\") " pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.815143 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxpnp\" (UniqueName: \"kubernetes.io/projected/b3103bdf-df5e-45c5-b453-90c903eb76ef-kube-api-access-fxpnp\") pod \"certified-operators-kbx77\" (UID: \"b3103bdf-df5e-45c5-b453-90c903eb76ef\") " pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.815238 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3103bdf-df5e-45c5-b453-90c903eb76ef-utilities\") pod \"certified-operators-kbx77\" (UID: \"b3103bdf-df5e-45c5-b453-90c903eb76ef\") " pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.917480 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3103bdf-df5e-45c5-b453-90c903eb76ef-catalog-content\") pod \"certified-operators-kbx77\" (UID: \"b3103bdf-df5e-45c5-b453-90c903eb76ef\") " pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.917573 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxpnp\" (UniqueName: \"kubernetes.io/projected/b3103bdf-df5e-45c5-b453-90c903eb76ef-kube-api-access-fxpnp\") pod \"certified-operators-kbx77\" (UID: \"b3103bdf-df5e-45c5-b453-90c903eb76ef\") " pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.917803 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3103bdf-df5e-45c5-b453-90c903eb76ef-utilities\") pod \"certified-operators-kbx77\" (UID: \"b3103bdf-df5e-45c5-b453-90c903eb76ef\") " pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.918401 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3103bdf-df5e-45c5-b453-90c903eb76ef-utilities\") pod \"certified-operators-kbx77\" (UID: \"b3103bdf-df5e-45c5-b453-90c903eb76ef\") " pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.918519 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3103bdf-df5e-45c5-b453-90c903eb76ef-catalog-content\") pod \"certified-operators-kbx77\" (UID: \"b3103bdf-df5e-45c5-b453-90c903eb76ef\") " pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:25 crc kubenswrapper[4775]: I1002 02:19:25.966048 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxpnp\" (UniqueName: \"kubernetes.io/projected/b3103bdf-df5e-45c5-b453-90c903eb76ef-kube-api-access-fxpnp\") pod \"certified-operators-kbx77\" (UID: \"b3103bdf-df5e-45c5-b453-90c903eb76ef\") " pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:26 crc kubenswrapper[4775]: I1002 02:19:26.048010 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:26 crc kubenswrapper[4775]: I1002 02:19:26.492545 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kbx77"] Oct 02 02:19:27 crc kubenswrapper[4775]: I1002 02:19:27.446105 4775 generic.go:334] "Generic (PLEG): container finished" podID="b3103bdf-df5e-45c5-b453-90c903eb76ef" containerID="c6e1348415ef3f4541ce0874b39e3658114940c983f094dec8b5fb0fbf15a0b7" exitCode=0 Oct 02 02:19:27 crc kubenswrapper[4775]: I1002 02:19:27.446173 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbx77" event={"ID":"b3103bdf-df5e-45c5-b453-90c903eb76ef","Type":"ContainerDied","Data":"c6e1348415ef3f4541ce0874b39e3658114940c983f094dec8b5fb0fbf15a0b7"} Oct 02 02:19:27 crc kubenswrapper[4775]: I1002 02:19:27.449138 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbx77" event={"ID":"b3103bdf-df5e-45c5-b453-90c903eb76ef","Type":"ContainerStarted","Data":"afe21aee8708d0e98b84029a9de2b521c3943cd3f6932546ea1c10ecde748bc4"} Oct 02 02:19:28 crc kubenswrapper[4775]: I1002 02:19:28.461073 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbx77" event={"ID":"b3103bdf-df5e-45c5-b453-90c903eb76ef","Type":"ContainerStarted","Data":"8200fd7c725a72769f26a54e91ba6309aeb5e8f9bdee8efa435bf5e4471d20b6"} Oct 02 02:19:29 crc kubenswrapper[4775]: I1002 02:19:29.471475 4775 generic.go:334] "Generic (PLEG): container finished" podID="b3103bdf-df5e-45c5-b453-90c903eb76ef" containerID="8200fd7c725a72769f26a54e91ba6309aeb5e8f9bdee8efa435bf5e4471d20b6" exitCode=0 Oct 02 02:19:29 crc kubenswrapper[4775]: I1002 02:19:29.471527 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbx77" event={"ID":"b3103bdf-df5e-45c5-b453-90c903eb76ef","Type":"ContainerDied","Data":"8200fd7c725a72769f26a54e91ba6309aeb5e8f9bdee8efa435bf5e4471d20b6"} Oct 02 02:19:30 crc kubenswrapper[4775]: I1002 02:19:30.482518 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbx77" event={"ID":"b3103bdf-df5e-45c5-b453-90c903eb76ef","Type":"ContainerStarted","Data":"24ddab5325593a3e406ba741d968db74b94aef82b99c7489decd125072ca1c40"} Oct 02 02:19:30 crc kubenswrapper[4775]: I1002 02:19:30.515483 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kbx77" podStartSLOduration=3.053833052 podStartE2EDuration="5.515451635s" podCreationTimestamp="2025-10-02 02:19:25 +0000 UTC" firstStartedPulling="2025-10-02 02:19:27.448479819 +0000 UTC m=+2304.615223899" lastFinishedPulling="2025-10-02 02:19:29.910098392 +0000 UTC m=+2307.076842482" observedRunningTime="2025-10-02 02:19:30.512360495 +0000 UTC m=+2307.679104585" watchObservedRunningTime="2025-10-02 02:19:30.515451635 +0000 UTC m=+2307.682195715" Oct 02 02:19:36 crc kubenswrapper[4775]: I1002 02:19:36.048947 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:36 crc kubenswrapper[4775]: I1002 02:19:36.050096 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:36 crc kubenswrapper[4775]: I1002 02:19:36.116704 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:36 crc kubenswrapper[4775]: I1002 02:19:36.626435 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:36 crc kubenswrapper[4775]: I1002 02:19:36.685009 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kbx77"] Oct 02 02:19:38 crc kubenswrapper[4775]: I1002 02:19:38.562468 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kbx77" podUID="b3103bdf-df5e-45c5-b453-90c903eb76ef" containerName="registry-server" containerID="cri-o://24ddab5325593a3e406ba741d968db74b94aef82b99c7489decd125072ca1c40" gracePeriod=2 Oct 02 02:19:38 crc kubenswrapper[4775]: I1002 02:19:38.765889 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:19:38 crc kubenswrapper[4775]: E1002 02:19:38.766307 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.027317 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.143179 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3103bdf-df5e-45c5-b453-90c903eb76ef-utilities\") pod \"b3103bdf-df5e-45c5-b453-90c903eb76ef\" (UID: \"b3103bdf-df5e-45c5-b453-90c903eb76ef\") " Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.143281 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3103bdf-df5e-45c5-b453-90c903eb76ef-catalog-content\") pod \"b3103bdf-df5e-45c5-b453-90c903eb76ef\" (UID: \"b3103bdf-df5e-45c5-b453-90c903eb76ef\") " Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.143354 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxpnp\" (UniqueName: \"kubernetes.io/projected/b3103bdf-df5e-45c5-b453-90c903eb76ef-kube-api-access-fxpnp\") pod \"b3103bdf-df5e-45c5-b453-90c903eb76ef\" (UID: \"b3103bdf-df5e-45c5-b453-90c903eb76ef\") " Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.145327 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3103bdf-df5e-45c5-b453-90c903eb76ef-utilities" (OuterVolumeSpecName: "utilities") pod "b3103bdf-df5e-45c5-b453-90c903eb76ef" (UID: "b3103bdf-df5e-45c5-b453-90c903eb76ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.153039 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3103bdf-df5e-45c5-b453-90c903eb76ef-kube-api-access-fxpnp" (OuterVolumeSpecName: "kube-api-access-fxpnp") pod "b3103bdf-df5e-45c5-b453-90c903eb76ef" (UID: "b3103bdf-df5e-45c5-b453-90c903eb76ef"). InnerVolumeSpecName "kube-api-access-fxpnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.244918 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3103bdf-df5e-45c5-b453-90c903eb76ef-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.245020 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxpnp\" (UniqueName: \"kubernetes.io/projected/b3103bdf-df5e-45c5-b453-90c903eb76ef-kube-api-access-fxpnp\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.579043 4775 generic.go:334] "Generic (PLEG): container finished" podID="b3103bdf-df5e-45c5-b453-90c903eb76ef" containerID="24ddab5325593a3e406ba741d968db74b94aef82b99c7489decd125072ca1c40" exitCode=0 Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.579145 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kbx77" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.580627 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbx77" event={"ID":"b3103bdf-df5e-45c5-b453-90c903eb76ef","Type":"ContainerDied","Data":"24ddab5325593a3e406ba741d968db74b94aef82b99c7489decd125072ca1c40"} Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.580791 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kbx77" event={"ID":"b3103bdf-df5e-45c5-b453-90c903eb76ef","Type":"ContainerDied","Data":"afe21aee8708d0e98b84029a9de2b521c3943cd3f6932546ea1c10ecde748bc4"} Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.580925 4775 scope.go:117] "RemoveContainer" containerID="24ddab5325593a3e406ba741d968db74b94aef82b99c7489decd125072ca1c40" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.617429 4775 scope.go:117] "RemoveContainer" containerID="8200fd7c725a72769f26a54e91ba6309aeb5e8f9bdee8efa435bf5e4471d20b6" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.663020 4775 scope.go:117] "RemoveContainer" containerID="c6e1348415ef3f4541ce0874b39e3658114940c983f094dec8b5fb0fbf15a0b7" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.692462 4775 scope.go:117] "RemoveContainer" containerID="24ddab5325593a3e406ba741d968db74b94aef82b99c7489decd125072ca1c40" Oct 02 02:19:39 crc kubenswrapper[4775]: E1002 02:19:39.693137 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24ddab5325593a3e406ba741d968db74b94aef82b99c7489decd125072ca1c40\": container with ID starting with 24ddab5325593a3e406ba741d968db74b94aef82b99c7489decd125072ca1c40 not found: ID does not exist" containerID="24ddab5325593a3e406ba741d968db74b94aef82b99c7489decd125072ca1c40" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.693206 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24ddab5325593a3e406ba741d968db74b94aef82b99c7489decd125072ca1c40"} err="failed to get container status \"24ddab5325593a3e406ba741d968db74b94aef82b99c7489decd125072ca1c40\": rpc error: code = NotFound desc = could not find container \"24ddab5325593a3e406ba741d968db74b94aef82b99c7489decd125072ca1c40\": container with ID starting with 24ddab5325593a3e406ba741d968db74b94aef82b99c7489decd125072ca1c40 not found: ID does not exist" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.693250 4775 scope.go:117] "RemoveContainer" containerID="8200fd7c725a72769f26a54e91ba6309aeb5e8f9bdee8efa435bf5e4471d20b6" Oct 02 02:19:39 crc kubenswrapper[4775]: E1002 02:19:39.693888 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8200fd7c725a72769f26a54e91ba6309aeb5e8f9bdee8efa435bf5e4471d20b6\": container with ID starting with 8200fd7c725a72769f26a54e91ba6309aeb5e8f9bdee8efa435bf5e4471d20b6 not found: ID does not exist" containerID="8200fd7c725a72769f26a54e91ba6309aeb5e8f9bdee8efa435bf5e4471d20b6" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.693928 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8200fd7c725a72769f26a54e91ba6309aeb5e8f9bdee8efa435bf5e4471d20b6"} err="failed to get container status \"8200fd7c725a72769f26a54e91ba6309aeb5e8f9bdee8efa435bf5e4471d20b6\": rpc error: code = NotFound desc = could not find container \"8200fd7c725a72769f26a54e91ba6309aeb5e8f9bdee8efa435bf5e4471d20b6\": container with ID starting with 8200fd7c725a72769f26a54e91ba6309aeb5e8f9bdee8efa435bf5e4471d20b6 not found: ID does not exist" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.693968 4775 scope.go:117] "RemoveContainer" containerID="c6e1348415ef3f4541ce0874b39e3658114940c983f094dec8b5fb0fbf15a0b7" Oct 02 02:19:39 crc kubenswrapper[4775]: E1002 02:19:39.694497 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6e1348415ef3f4541ce0874b39e3658114940c983f094dec8b5fb0fbf15a0b7\": container with ID starting with c6e1348415ef3f4541ce0874b39e3658114940c983f094dec8b5fb0fbf15a0b7 not found: ID does not exist" containerID="c6e1348415ef3f4541ce0874b39e3658114940c983f094dec8b5fb0fbf15a0b7" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.694550 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6e1348415ef3f4541ce0874b39e3658114940c983f094dec8b5fb0fbf15a0b7"} err="failed to get container status \"c6e1348415ef3f4541ce0874b39e3658114940c983f094dec8b5fb0fbf15a0b7\": rpc error: code = NotFound desc = could not find container \"c6e1348415ef3f4541ce0874b39e3658114940c983f094dec8b5fb0fbf15a0b7\": container with ID starting with c6e1348415ef3f4541ce0874b39e3658114940c983f094dec8b5fb0fbf15a0b7 not found: ID does not exist" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.803622 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3103bdf-df5e-45c5-b453-90c903eb76ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3103bdf-df5e-45c5-b453-90c903eb76ef" (UID: "b3103bdf-df5e-45c5-b453-90c903eb76ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.854872 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3103bdf-df5e-45c5-b453-90c903eb76ef-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.920603 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kbx77"] Oct 02 02:19:39 crc kubenswrapper[4775]: I1002 02:19:39.927904 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kbx77"] Oct 02 02:19:41 crc kubenswrapper[4775]: I1002 02:19:41.781661 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3103bdf-df5e-45c5-b453-90c903eb76ef" path="/var/lib/kubelet/pods/b3103bdf-df5e-45c5-b453-90c903eb76ef/volumes" Oct 02 02:19:53 crc kubenswrapper[4775]: I1002 02:19:53.773052 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:19:53 crc kubenswrapper[4775]: E1002 02:19:53.773663 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:20:06 crc kubenswrapper[4775]: I1002 02:20:06.765821 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:20:06 crc kubenswrapper[4775]: E1002 02:20:06.767010 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:20:20 crc kubenswrapper[4775]: I1002 02:20:20.765950 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:20:20 crc kubenswrapper[4775]: E1002 02:20:20.766840 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:20:32 crc kubenswrapper[4775]: I1002 02:20:32.765945 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:20:32 crc kubenswrapper[4775]: E1002 02:20:32.767621 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:20:46 crc kubenswrapper[4775]: I1002 02:20:46.765524 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:20:46 crc kubenswrapper[4775]: E1002 02:20:46.766280 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:20:58 crc kubenswrapper[4775]: I1002 02:20:58.765407 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:20:58 crc kubenswrapper[4775]: E1002 02:20:58.766284 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:21:10 crc kubenswrapper[4775]: I1002 02:21:10.766069 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:21:10 crc kubenswrapper[4775]: E1002 02:21:10.767180 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:21:22 crc kubenswrapper[4775]: I1002 02:21:22.765743 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:21:22 crc kubenswrapper[4775]: E1002 02:21:22.766803 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:21:37 crc kubenswrapper[4775]: I1002 02:21:37.764781 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:21:37 crc kubenswrapper[4775]: E1002 02:21:37.765824 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:21:48 crc kubenswrapper[4775]: I1002 02:21:48.766035 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:21:48 crc kubenswrapper[4775]: E1002 02:21:48.767089 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:22:01 crc kubenswrapper[4775]: I1002 02:22:01.765905 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:22:01 crc kubenswrapper[4775]: E1002 02:22:01.767275 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:22:14 crc kubenswrapper[4775]: I1002 02:22:14.765396 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:22:14 crc kubenswrapper[4775]: E1002 02:22:14.766491 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:22:26 crc kubenswrapper[4775]: I1002 02:22:26.765885 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:22:26 crc kubenswrapper[4775]: E1002 02:22:26.767020 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:22:40 crc kubenswrapper[4775]: I1002 02:22:40.765720 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:22:40 crc kubenswrapper[4775]: E1002 02:22:40.766646 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:22:52 crc kubenswrapper[4775]: I1002 02:22:52.764981 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:22:52 crc kubenswrapper[4775]: E1002 02:22:52.765900 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:23:06 crc kubenswrapper[4775]: I1002 02:23:06.766207 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:23:06 crc kubenswrapper[4775]: E1002 02:23:06.767741 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:23:20 crc kubenswrapper[4775]: I1002 02:23:20.766425 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:23:21 crc kubenswrapper[4775]: I1002 02:23:21.707539 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"3f87c732873861f7da11725e497cfd107b7bc6b663b852114f5d70f0a26de837"} Oct 02 02:25:37 crc kubenswrapper[4775]: I1002 02:25:37.234146 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:25:37 crc kubenswrapper[4775]: I1002 02:25:37.234922 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:25:43 crc kubenswrapper[4775]: I1002 02:25:43.968281 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9m6kb"] Oct 02 02:25:43 crc kubenswrapper[4775]: E1002 02:25:43.969537 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3103bdf-df5e-45c5-b453-90c903eb76ef" containerName="extract-content" Oct 02 02:25:43 crc kubenswrapper[4775]: I1002 02:25:43.969565 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3103bdf-df5e-45c5-b453-90c903eb76ef" containerName="extract-content" Oct 02 02:25:43 crc kubenswrapper[4775]: E1002 02:25:43.969610 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3103bdf-df5e-45c5-b453-90c903eb76ef" containerName="extract-utilities" Oct 02 02:25:43 crc kubenswrapper[4775]: I1002 02:25:43.969625 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3103bdf-df5e-45c5-b453-90c903eb76ef" containerName="extract-utilities" Oct 02 02:25:43 crc kubenswrapper[4775]: E1002 02:25:43.969652 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3103bdf-df5e-45c5-b453-90c903eb76ef" containerName="registry-server" Oct 02 02:25:43 crc kubenswrapper[4775]: I1002 02:25:43.969665 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3103bdf-df5e-45c5-b453-90c903eb76ef" containerName="registry-server" Oct 02 02:25:43 crc kubenswrapper[4775]: I1002 02:25:43.969935 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3103bdf-df5e-45c5-b453-90c903eb76ef" containerName="registry-server" Oct 02 02:25:43 crc kubenswrapper[4775]: I1002 02:25:43.971841 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:43 crc kubenswrapper[4775]: I1002 02:25:43.996910 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9m6kb"] Oct 02 02:25:44 crc kubenswrapper[4775]: I1002 02:25:44.088892 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d2589ef-f790-4428-9433-c2da78eb19ce-utilities\") pod \"redhat-operators-9m6kb\" (UID: \"0d2589ef-f790-4428-9433-c2da78eb19ce\") " pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:44 crc kubenswrapper[4775]: I1002 02:25:44.089021 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxcb4\" (UniqueName: \"kubernetes.io/projected/0d2589ef-f790-4428-9433-c2da78eb19ce-kube-api-access-bxcb4\") pod \"redhat-operators-9m6kb\" (UID: \"0d2589ef-f790-4428-9433-c2da78eb19ce\") " pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:44 crc kubenswrapper[4775]: I1002 02:25:44.089092 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d2589ef-f790-4428-9433-c2da78eb19ce-catalog-content\") pod \"redhat-operators-9m6kb\" (UID: \"0d2589ef-f790-4428-9433-c2da78eb19ce\") " pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:44 crc kubenswrapper[4775]: I1002 02:25:44.190234 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d2589ef-f790-4428-9433-c2da78eb19ce-utilities\") pod \"redhat-operators-9m6kb\" (UID: \"0d2589ef-f790-4428-9433-c2da78eb19ce\") " pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:44 crc kubenswrapper[4775]: I1002 02:25:44.190332 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxcb4\" (UniqueName: \"kubernetes.io/projected/0d2589ef-f790-4428-9433-c2da78eb19ce-kube-api-access-bxcb4\") pod \"redhat-operators-9m6kb\" (UID: \"0d2589ef-f790-4428-9433-c2da78eb19ce\") " pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:44 crc kubenswrapper[4775]: I1002 02:25:44.190416 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d2589ef-f790-4428-9433-c2da78eb19ce-catalog-content\") pod \"redhat-operators-9m6kb\" (UID: \"0d2589ef-f790-4428-9433-c2da78eb19ce\") " pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:44 crc kubenswrapper[4775]: I1002 02:25:44.190830 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d2589ef-f790-4428-9433-c2da78eb19ce-utilities\") pod \"redhat-operators-9m6kb\" (UID: \"0d2589ef-f790-4428-9433-c2da78eb19ce\") " pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:44 crc kubenswrapper[4775]: I1002 02:25:44.191334 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d2589ef-f790-4428-9433-c2da78eb19ce-catalog-content\") pod \"redhat-operators-9m6kb\" (UID: \"0d2589ef-f790-4428-9433-c2da78eb19ce\") " pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:44 crc kubenswrapper[4775]: I1002 02:25:44.217745 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxcb4\" (UniqueName: \"kubernetes.io/projected/0d2589ef-f790-4428-9433-c2da78eb19ce-kube-api-access-bxcb4\") pod \"redhat-operators-9m6kb\" (UID: \"0d2589ef-f790-4428-9433-c2da78eb19ce\") " pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:44 crc kubenswrapper[4775]: I1002 02:25:44.311087 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:44 crc kubenswrapper[4775]: I1002 02:25:44.761378 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9m6kb"] Oct 02 02:25:45 crc kubenswrapper[4775]: I1002 02:25:45.080477 4775 generic.go:334] "Generic (PLEG): container finished" podID="0d2589ef-f790-4428-9433-c2da78eb19ce" containerID="806d3f5c4b15951833aca0b3f15ab3a3b4fb0ad76bc623d1d2c4304ebc639c28" exitCode=0 Oct 02 02:25:45 crc kubenswrapper[4775]: I1002 02:25:45.080524 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9m6kb" event={"ID":"0d2589ef-f790-4428-9433-c2da78eb19ce","Type":"ContainerDied","Data":"806d3f5c4b15951833aca0b3f15ab3a3b4fb0ad76bc623d1d2c4304ebc639c28"} Oct 02 02:25:45 crc kubenswrapper[4775]: I1002 02:25:45.080843 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9m6kb" event={"ID":"0d2589ef-f790-4428-9433-c2da78eb19ce","Type":"ContainerStarted","Data":"5a8a677d66ebc28db4790ed81d48077e9c06aee2782a4cbb671ad1d643643321"} Oct 02 02:25:45 crc kubenswrapper[4775]: I1002 02:25:45.083621 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 02:25:47 crc kubenswrapper[4775]: I1002 02:25:47.116423 4775 generic.go:334] "Generic (PLEG): container finished" podID="0d2589ef-f790-4428-9433-c2da78eb19ce" containerID="6fdbfac6ba7048bf5fb29e22e4fa32df41069b84ec1bc2ac0a7c46ea8384d19e" exitCode=0 Oct 02 02:25:47 crc kubenswrapper[4775]: I1002 02:25:47.116560 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9m6kb" event={"ID":"0d2589ef-f790-4428-9433-c2da78eb19ce","Type":"ContainerDied","Data":"6fdbfac6ba7048bf5fb29e22e4fa32df41069b84ec1bc2ac0a7c46ea8384d19e"} Oct 02 02:25:48 crc kubenswrapper[4775]: I1002 02:25:48.133014 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9m6kb" event={"ID":"0d2589ef-f790-4428-9433-c2da78eb19ce","Type":"ContainerStarted","Data":"4a9b05e8ab66a780a48638d011aa3ea844877687a7c67b2ebdf275ad9638c617"} Oct 02 02:25:48 crc kubenswrapper[4775]: I1002 02:25:48.163651 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9m6kb" podStartSLOduration=2.6573146469999998 podStartE2EDuration="5.163626023s" podCreationTimestamp="2025-10-02 02:25:43 +0000 UTC" firstStartedPulling="2025-10-02 02:25:45.083340755 +0000 UTC m=+2682.250084815" lastFinishedPulling="2025-10-02 02:25:47.589652151 +0000 UTC m=+2684.756396191" observedRunningTime="2025-10-02 02:25:48.156015948 +0000 UTC m=+2685.322760058" watchObservedRunningTime="2025-10-02 02:25:48.163626023 +0000 UTC m=+2685.330370103" Oct 02 02:25:49 crc kubenswrapper[4775]: I1002 02:25:49.367175 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lkthq"] Oct 02 02:25:49 crc kubenswrapper[4775]: I1002 02:25:49.373117 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:25:49 crc kubenswrapper[4775]: I1002 02:25:49.381184 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lkthq"] Oct 02 02:25:49 crc kubenswrapper[4775]: I1002 02:25:49.475367 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q58cf\" (UniqueName: \"kubernetes.io/projected/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-kube-api-access-q58cf\") pod \"community-operators-lkthq\" (UID: \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\") " pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:25:49 crc kubenswrapper[4775]: I1002 02:25:49.475530 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-catalog-content\") pod \"community-operators-lkthq\" (UID: \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\") " pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:25:49 crc kubenswrapper[4775]: I1002 02:25:49.475590 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-utilities\") pod \"community-operators-lkthq\" (UID: \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\") " pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:25:49 crc kubenswrapper[4775]: I1002 02:25:49.576583 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-catalog-content\") pod \"community-operators-lkthq\" (UID: \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\") " pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:25:49 crc kubenswrapper[4775]: I1002 02:25:49.576937 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-utilities\") pod \"community-operators-lkthq\" (UID: \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\") " pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:25:49 crc kubenswrapper[4775]: I1002 02:25:49.577069 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q58cf\" (UniqueName: \"kubernetes.io/projected/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-kube-api-access-q58cf\") pod \"community-operators-lkthq\" (UID: \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\") " pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:25:49 crc kubenswrapper[4775]: I1002 02:25:49.577413 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-catalog-content\") pod \"community-operators-lkthq\" (UID: \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\") " pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:25:49 crc kubenswrapper[4775]: I1002 02:25:49.577533 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-utilities\") pod \"community-operators-lkthq\" (UID: \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\") " pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:25:49 crc kubenswrapper[4775]: I1002 02:25:49.607899 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q58cf\" (UniqueName: \"kubernetes.io/projected/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-kube-api-access-q58cf\") pod \"community-operators-lkthq\" (UID: \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\") " pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:25:49 crc kubenswrapper[4775]: I1002 02:25:49.718208 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:25:50 crc kubenswrapper[4775]: I1002 02:25:50.211043 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lkthq"] Oct 02 02:25:50 crc kubenswrapper[4775]: W1002 02:25:50.217397 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ea370e4_d04b_4a9b_a7ea_b31865cf86f2.slice/crio-272bf1af3b3a22ab0ca6a310559cf0b4fedb6d4012e4d6063949d9d07e348537 WatchSource:0}: Error finding container 272bf1af3b3a22ab0ca6a310559cf0b4fedb6d4012e4d6063949d9d07e348537: Status 404 returned error can't find the container with id 272bf1af3b3a22ab0ca6a310559cf0b4fedb6d4012e4d6063949d9d07e348537 Oct 02 02:25:51 crc kubenswrapper[4775]: I1002 02:25:51.158889 4775 generic.go:334] "Generic (PLEG): container finished" podID="6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" containerID="e3be5b07032ab308c8e968e7d5376b4419b4829a0cd850d87f763e4dbfed075f" exitCode=0 Oct 02 02:25:51 crc kubenswrapper[4775]: I1002 02:25:51.158992 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lkthq" event={"ID":"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2","Type":"ContainerDied","Data":"e3be5b07032ab308c8e968e7d5376b4419b4829a0cd850d87f763e4dbfed075f"} Oct 02 02:25:51 crc kubenswrapper[4775]: I1002 02:25:51.159219 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lkthq" event={"ID":"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2","Type":"ContainerStarted","Data":"272bf1af3b3a22ab0ca6a310559cf0b4fedb6d4012e4d6063949d9d07e348537"} Oct 02 02:25:52 crc kubenswrapper[4775]: I1002 02:25:52.169588 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lkthq" event={"ID":"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2","Type":"ContainerStarted","Data":"b01ab9ad80b6b7b66887d7d6b8ad79fac0b18480e2af176d277c202cc828692f"} Oct 02 02:25:53 crc kubenswrapper[4775]: I1002 02:25:53.181782 4775 generic.go:334] "Generic (PLEG): container finished" podID="6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" containerID="b01ab9ad80b6b7b66887d7d6b8ad79fac0b18480e2af176d277c202cc828692f" exitCode=0 Oct 02 02:25:53 crc kubenswrapper[4775]: I1002 02:25:53.181860 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lkthq" event={"ID":"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2","Type":"ContainerDied","Data":"b01ab9ad80b6b7b66887d7d6b8ad79fac0b18480e2af176d277c202cc828692f"} Oct 02 02:25:54 crc kubenswrapper[4775]: I1002 02:25:54.194391 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lkthq" event={"ID":"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2","Type":"ContainerStarted","Data":"832b4fd8c95c7ca1a65e7d72a7c16b5a7a453fdf8796410514bc2a47178c2f23"} Oct 02 02:25:54 crc kubenswrapper[4775]: I1002 02:25:54.221421 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lkthq" podStartSLOduration=2.806773559 podStartE2EDuration="5.221397193s" podCreationTimestamp="2025-10-02 02:25:49 +0000 UTC" firstStartedPulling="2025-10-02 02:25:51.164091874 +0000 UTC m=+2688.330835924" lastFinishedPulling="2025-10-02 02:25:53.578715518 +0000 UTC m=+2690.745459558" observedRunningTime="2025-10-02 02:25:54.214214329 +0000 UTC m=+2691.380958399" watchObservedRunningTime="2025-10-02 02:25:54.221397193 +0000 UTC m=+2691.388141263" Oct 02 02:25:54 crc kubenswrapper[4775]: I1002 02:25:54.311254 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:54 crc kubenswrapper[4775]: I1002 02:25:54.311347 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:54 crc kubenswrapper[4775]: I1002 02:25:54.379247 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:55 crc kubenswrapper[4775]: I1002 02:25:55.275340 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:55 crc kubenswrapper[4775]: I1002 02:25:55.952566 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9m6kb"] Oct 02 02:25:57 crc kubenswrapper[4775]: I1002 02:25:57.229574 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9m6kb" podUID="0d2589ef-f790-4428-9433-c2da78eb19ce" containerName="registry-server" containerID="cri-o://4a9b05e8ab66a780a48638d011aa3ea844877687a7c67b2ebdf275ad9638c617" gracePeriod=2 Oct 02 02:25:58 crc kubenswrapper[4775]: I1002 02:25:58.241632 4775 generic.go:334] "Generic (PLEG): container finished" podID="0d2589ef-f790-4428-9433-c2da78eb19ce" containerID="4a9b05e8ab66a780a48638d011aa3ea844877687a7c67b2ebdf275ad9638c617" exitCode=0 Oct 02 02:25:58 crc kubenswrapper[4775]: I1002 02:25:58.241763 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9m6kb" event={"ID":"0d2589ef-f790-4428-9433-c2da78eb19ce","Type":"ContainerDied","Data":"4a9b05e8ab66a780a48638d011aa3ea844877687a7c67b2ebdf275ad9638c617"} Oct 02 02:25:58 crc kubenswrapper[4775]: I1002 02:25:58.802715 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:58 crc kubenswrapper[4775]: I1002 02:25:58.933609 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d2589ef-f790-4428-9433-c2da78eb19ce-utilities\") pod \"0d2589ef-f790-4428-9433-c2da78eb19ce\" (UID: \"0d2589ef-f790-4428-9433-c2da78eb19ce\") " Oct 02 02:25:58 crc kubenswrapper[4775]: I1002 02:25:58.933732 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d2589ef-f790-4428-9433-c2da78eb19ce-catalog-content\") pod \"0d2589ef-f790-4428-9433-c2da78eb19ce\" (UID: \"0d2589ef-f790-4428-9433-c2da78eb19ce\") " Oct 02 02:25:58 crc kubenswrapper[4775]: I1002 02:25:58.933790 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxcb4\" (UniqueName: \"kubernetes.io/projected/0d2589ef-f790-4428-9433-c2da78eb19ce-kube-api-access-bxcb4\") pod \"0d2589ef-f790-4428-9433-c2da78eb19ce\" (UID: \"0d2589ef-f790-4428-9433-c2da78eb19ce\") " Oct 02 02:25:58 crc kubenswrapper[4775]: I1002 02:25:58.935491 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d2589ef-f790-4428-9433-c2da78eb19ce-utilities" (OuterVolumeSpecName: "utilities") pod "0d2589ef-f790-4428-9433-c2da78eb19ce" (UID: "0d2589ef-f790-4428-9433-c2da78eb19ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:25:58 crc kubenswrapper[4775]: I1002 02:25:58.943146 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d2589ef-f790-4428-9433-c2da78eb19ce-kube-api-access-bxcb4" (OuterVolumeSpecName: "kube-api-access-bxcb4") pod "0d2589ef-f790-4428-9433-c2da78eb19ce" (UID: "0d2589ef-f790-4428-9433-c2da78eb19ce"). InnerVolumeSpecName "kube-api-access-bxcb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.036209 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxcb4\" (UniqueName: \"kubernetes.io/projected/0d2589ef-f790-4428-9433-c2da78eb19ce-kube-api-access-bxcb4\") on node \"crc\" DevicePath \"\"" Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.036258 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d2589ef-f790-4428-9433-c2da78eb19ce-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.068813 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d2589ef-f790-4428-9433-c2da78eb19ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d2589ef-f790-4428-9433-c2da78eb19ce" (UID: "0d2589ef-f790-4428-9433-c2da78eb19ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.138073 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d2589ef-f790-4428-9433-c2da78eb19ce-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.259087 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9m6kb" event={"ID":"0d2589ef-f790-4428-9433-c2da78eb19ce","Type":"ContainerDied","Data":"5a8a677d66ebc28db4790ed81d48077e9c06aee2782a4cbb671ad1d643643321"} Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.259172 4775 scope.go:117] "RemoveContainer" containerID="4a9b05e8ab66a780a48638d011aa3ea844877687a7c67b2ebdf275ad9638c617" Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.259295 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9m6kb" Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.287181 4775 scope.go:117] "RemoveContainer" containerID="6fdbfac6ba7048bf5fb29e22e4fa32df41069b84ec1bc2ac0a7c46ea8384d19e" Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.323241 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9m6kb"] Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.324058 4775 scope.go:117] "RemoveContainer" containerID="806d3f5c4b15951833aca0b3f15ab3a3b4fb0ad76bc623d1d2c4304ebc639c28" Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.328759 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9m6kb"] Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.719015 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.719082 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.791059 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d2589ef-f790-4428-9433-c2da78eb19ce" path="/var/lib/kubelet/pods/0d2589ef-f790-4428-9433-c2da78eb19ce/volumes" Oct 02 02:25:59 crc kubenswrapper[4775]: I1002 02:25:59.803165 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:26:00 crc kubenswrapper[4775]: I1002 02:26:00.330090 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:26:01 crc kubenswrapper[4775]: I1002 02:26:01.557869 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lkthq"] Oct 02 02:26:02 crc kubenswrapper[4775]: I1002 02:26:02.291134 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lkthq" podUID="6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" containerName="registry-server" containerID="cri-o://832b4fd8c95c7ca1a65e7d72a7c16b5a7a453fdf8796410514bc2a47178c2f23" gracePeriod=2 Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.264214 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.303660 4775 generic.go:334] "Generic (PLEG): container finished" podID="6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" containerID="832b4fd8c95c7ca1a65e7d72a7c16b5a7a453fdf8796410514bc2a47178c2f23" exitCode=0 Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.303724 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lkthq" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.303764 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lkthq" event={"ID":"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2","Type":"ContainerDied","Data":"832b4fd8c95c7ca1a65e7d72a7c16b5a7a453fdf8796410514bc2a47178c2f23"} Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.303850 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lkthq" event={"ID":"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2","Type":"ContainerDied","Data":"272bf1af3b3a22ab0ca6a310559cf0b4fedb6d4012e4d6063949d9d07e348537"} Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.303885 4775 scope.go:117] "RemoveContainer" containerID="832b4fd8c95c7ca1a65e7d72a7c16b5a7a453fdf8796410514bc2a47178c2f23" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.328338 4775 scope.go:117] "RemoveContainer" containerID="b01ab9ad80b6b7b66887d7d6b8ad79fac0b18480e2af176d277c202cc828692f" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.346373 4775 scope.go:117] "RemoveContainer" containerID="e3be5b07032ab308c8e968e7d5376b4419b4829a0cd850d87f763e4dbfed075f" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.373748 4775 scope.go:117] "RemoveContainer" containerID="832b4fd8c95c7ca1a65e7d72a7c16b5a7a453fdf8796410514bc2a47178c2f23" Oct 02 02:26:03 crc kubenswrapper[4775]: E1002 02:26:03.374274 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"832b4fd8c95c7ca1a65e7d72a7c16b5a7a453fdf8796410514bc2a47178c2f23\": container with ID starting with 832b4fd8c95c7ca1a65e7d72a7c16b5a7a453fdf8796410514bc2a47178c2f23 not found: ID does not exist" containerID="832b4fd8c95c7ca1a65e7d72a7c16b5a7a453fdf8796410514bc2a47178c2f23" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.374339 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"832b4fd8c95c7ca1a65e7d72a7c16b5a7a453fdf8796410514bc2a47178c2f23"} err="failed to get container status \"832b4fd8c95c7ca1a65e7d72a7c16b5a7a453fdf8796410514bc2a47178c2f23\": rpc error: code = NotFound desc = could not find container \"832b4fd8c95c7ca1a65e7d72a7c16b5a7a453fdf8796410514bc2a47178c2f23\": container with ID starting with 832b4fd8c95c7ca1a65e7d72a7c16b5a7a453fdf8796410514bc2a47178c2f23 not found: ID does not exist" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.374379 4775 scope.go:117] "RemoveContainer" containerID="b01ab9ad80b6b7b66887d7d6b8ad79fac0b18480e2af176d277c202cc828692f" Oct 02 02:26:03 crc kubenswrapper[4775]: E1002 02:26:03.374944 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b01ab9ad80b6b7b66887d7d6b8ad79fac0b18480e2af176d277c202cc828692f\": container with ID starting with b01ab9ad80b6b7b66887d7d6b8ad79fac0b18480e2af176d277c202cc828692f not found: ID does not exist" containerID="b01ab9ad80b6b7b66887d7d6b8ad79fac0b18480e2af176d277c202cc828692f" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.375161 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b01ab9ad80b6b7b66887d7d6b8ad79fac0b18480e2af176d277c202cc828692f"} err="failed to get container status \"b01ab9ad80b6b7b66887d7d6b8ad79fac0b18480e2af176d277c202cc828692f\": rpc error: code = NotFound desc = could not find container \"b01ab9ad80b6b7b66887d7d6b8ad79fac0b18480e2af176d277c202cc828692f\": container with ID starting with b01ab9ad80b6b7b66887d7d6b8ad79fac0b18480e2af176d277c202cc828692f not found: ID does not exist" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.375314 4775 scope.go:117] "RemoveContainer" containerID="e3be5b07032ab308c8e968e7d5376b4419b4829a0cd850d87f763e4dbfed075f" Oct 02 02:26:03 crc kubenswrapper[4775]: E1002 02:26:03.375851 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3be5b07032ab308c8e968e7d5376b4419b4829a0cd850d87f763e4dbfed075f\": container with ID starting with e3be5b07032ab308c8e968e7d5376b4419b4829a0cd850d87f763e4dbfed075f not found: ID does not exist" containerID="e3be5b07032ab308c8e968e7d5376b4419b4829a0cd850d87f763e4dbfed075f" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.376560 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3be5b07032ab308c8e968e7d5376b4419b4829a0cd850d87f763e4dbfed075f"} err="failed to get container status \"e3be5b07032ab308c8e968e7d5376b4419b4829a0cd850d87f763e4dbfed075f\": rpc error: code = NotFound desc = could not find container \"e3be5b07032ab308c8e968e7d5376b4419b4829a0cd850d87f763e4dbfed075f\": container with ID starting with e3be5b07032ab308c8e968e7d5376b4419b4829a0cd850d87f763e4dbfed075f not found: ID does not exist" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.409237 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-catalog-content\") pod \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\" (UID: \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\") " Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.409463 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-utilities\") pod \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\" (UID: \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\") " Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.409641 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q58cf\" (UniqueName: \"kubernetes.io/projected/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-kube-api-access-q58cf\") pod \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\" (UID: \"6ea370e4-d04b-4a9b-a7ea-b31865cf86f2\") " Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.410796 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-utilities" (OuterVolumeSpecName: "utilities") pod "6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" (UID: "6ea370e4-d04b-4a9b-a7ea-b31865cf86f2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.415428 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-kube-api-access-q58cf" (OuterVolumeSpecName: "kube-api-access-q58cf") pod "6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" (UID: "6ea370e4-d04b-4a9b-a7ea-b31865cf86f2"). InnerVolumeSpecName "kube-api-access-q58cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.463642 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" (UID: "6ea370e4-d04b-4a9b-a7ea-b31865cf86f2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.511993 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.512160 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q58cf\" (UniqueName: \"kubernetes.io/projected/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-kube-api-access-q58cf\") on node \"crc\" DevicePath \"\"" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.512173 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.653872 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lkthq"] Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.662540 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lkthq"] Oct 02 02:26:03 crc kubenswrapper[4775]: I1002 02:26:03.789415 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" path="/var/lib/kubelet/pods/6ea370e4-d04b-4a9b-a7ea-b31865cf86f2/volumes" Oct 02 02:26:07 crc kubenswrapper[4775]: I1002 02:26:07.233776 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:26:07 crc kubenswrapper[4775]: I1002 02:26:07.234202 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:26:37 crc kubenswrapper[4775]: I1002 02:26:37.234259 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:26:37 crc kubenswrapper[4775]: I1002 02:26:37.234948 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:26:37 crc kubenswrapper[4775]: I1002 02:26:37.235047 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 02:26:37 crc kubenswrapper[4775]: I1002 02:26:37.236620 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3f87c732873861f7da11725e497cfd107b7bc6b663b852114f5d70f0a26de837"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:26:37 crc kubenswrapper[4775]: I1002 02:26:37.236830 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://3f87c732873861f7da11725e497cfd107b7bc6b663b852114f5d70f0a26de837" gracePeriod=600 Oct 02 02:26:37 crc kubenswrapper[4775]: I1002 02:26:37.638327 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="3f87c732873861f7da11725e497cfd107b7bc6b663b852114f5d70f0a26de837" exitCode=0 Oct 02 02:26:37 crc kubenswrapper[4775]: I1002 02:26:37.638433 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"3f87c732873861f7da11725e497cfd107b7bc6b663b852114f5d70f0a26de837"} Oct 02 02:26:37 crc kubenswrapper[4775]: I1002 02:26:37.638687 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d"} Oct 02 02:26:37 crc kubenswrapper[4775]: I1002 02:26:37.638722 4775 scope.go:117] "RemoveContainer" containerID="c9bb2e4b0c6b4f113294aea7d27f69860f764e64d8d9f2bf18bf3c7569b21317" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.174249 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h2rs6"] Oct 02 02:28:26 crc kubenswrapper[4775]: E1002 02:28:26.175470 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" containerName="registry-server" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.175491 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" containerName="registry-server" Oct 02 02:28:26 crc kubenswrapper[4775]: E1002 02:28:26.175515 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" containerName="extract-utilities" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.175523 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" containerName="extract-utilities" Oct 02 02:28:26 crc kubenswrapper[4775]: E1002 02:28:26.175533 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d2589ef-f790-4428-9433-c2da78eb19ce" containerName="extract-utilities" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.175552 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d2589ef-f790-4428-9433-c2da78eb19ce" containerName="extract-utilities" Oct 02 02:28:26 crc kubenswrapper[4775]: E1002 02:28:26.175571 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d2589ef-f790-4428-9433-c2da78eb19ce" containerName="registry-server" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.175578 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d2589ef-f790-4428-9433-c2da78eb19ce" containerName="registry-server" Oct 02 02:28:26 crc kubenswrapper[4775]: E1002 02:28:26.175590 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" containerName="extract-content" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.175597 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" containerName="extract-content" Oct 02 02:28:26 crc kubenswrapper[4775]: E1002 02:28:26.175619 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d2589ef-f790-4428-9433-c2da78eb19ce" containerName="extract-content" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.175628 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d2589ef-f790-4428-9433-c2da78eb19ce" containerName="extract-content" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.175843 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d2589ef-f790-4428-9433-c2da78eb19ce" containerName="registry-server" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.175868 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ea370e4-d04b-4a9b-a7ea-b31865cf86f2" containerName="registry-server" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.177273 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.194512 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2rs6"] Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.333164 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnk9z\" (UniqueName: \"kubernetes.io/projected/b50612db-5c36-47a5-b4ee-44f754df5de5-kube-api-access-mnk9z\") pod \"redhat-marketplace-h2rs6\" (UID: \"b50612db-5c36-47a5-b4ee-44f754df5de5\") " pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.333209 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b50612db-5c36-47a5-b4ee-44f754df5de5-utilities\") pod \"redhat-marketplace-h2rs6\" (UID: \"b50612db-5c36-47a5-b4ee-44f754df5de5\") " pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.333301 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b50612db-5c36-47a5-b4ee-44f754df5de5-catalog-content\") pod \"redhat-marketplace-h2rs6\" (UID: \"b50612db-5c36-47a5-b4ee-44f754df5de5\") " pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.434189 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b50612db-5c36-47a5-b4ee-44f754df5de5-catalog-content\") pod \"redhat-marketplace-h2rs6\" (UID: \"b50612db-5c36-47a5-b4ee-44f754df5de5\") " pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.434285 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnk9z\" (UniqueName: \"kubernetes.io/projected/b50612db-5c36-47a5-b4ee-44f754df5de5-kube-api-access-mnk9z\") pod \"redhat-marketplace-h2rs6\" (UID: \"b50612db-5c36-47a5-b4ee-44f754df5de5\") " pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.434310 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b50612db-5c36-47a5-b4ee-44f754df5de5-utilities\") pod \"redhat-marketplace-h2rs6\" (UID: \"b50612db-5c36-47a5-b4ee-44f754df5de5\") " pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.434846 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b50612db-5c36-47a5-b4ee-44f754df5de5-catalog-content\") pod \"redhat-marketplace-h2rs6\" (UID: \"b50612db-5c36-47a5-b4ee-44f754df5de5\") " pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.434891 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b50612db-5c36-47a5-b4ee-44f754df5de5-utilities\") pod \"redhat-marketplace-h2rs6\" (UID: \"b50612db-5c36-47a5-b4ee-44f754df5de5\") " pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.463007 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnk9z\" (UniqueName: \"kubernetes.io/projected/b50612db-5c36-47a5-b4ee-44f754df5de5-kube-api-access-mnk9z\") pod \"redhat-marketplace-h2rs6\" (UID: \"b50612db-5c36-47a5-b4ee-44f754df5de5\") " pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.521704 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:26 crc kubenswrapper[4775]: I1002 02:28:26.767261 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2rs6"] Oct 02 02:28:27 crc kubenswrapper[4775]: I1002 02:28:27.673103 4775 generic.go:334] "Generic (PLEG): container finished" podID="b50612db-5c36-47a5-b4ee-44f754df5de5" containerID="f07cb704dc654a550059f35b5adf90b13e80e27fb48f73290a151e82c01ff0a4" exitCode=0 Oct 02 02:28:27 crc kubenswrapper[4775]: I1002 02:28:27.673204 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2rs6" event={"ID":"b50612db-5c36-47a5-b4ee-44f754df5de5","Type":"ContainerDied","Data":"f07cb704dc654a550059f35b5adf90b13e80e27fb48f73290a151e82c01ff0a4"} Oct 02 02:28:27 crc kubenswrapper[4775]: I1002 02:28:27.673348 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2rs6" event={"ID":"b50612db-5c36-47a5-b4ee-44f754df5de5","Type":"ContainerStarted","Data":"0877fc74e163c468332cb3fedb1dae22a93c4eb5f5080251417f9d2f72b57f91"} Oct 02 02:28:29 crc kubenswrapper[4775]: I1002 02:28:29.689379 4775 generic.go:334] "Generic (PLEG): container finished" podID="b50612db-5c36-47a5-b4ee-44f754df5de5" containerID="ce81b20fc937e8b6f77e8bfa2d851d7b00eab44736ea5772fe8f01868ad92629" exitCode=0 Oct 02 02:28:29 crc kubenswrapper[4775]: I1002 02:28:29.689426 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2rs6" event={"ID":"b50612db-5c36-47a5-b4ee-44f754df5de5","Type":"ContainerDied","Data":"ce81b20fc937e8b6f77e8bfa2d851d7b00eab44736ea5772fe8f01868ad92629"} Oct 02 02:28:30 crc kubenswrapper[4775]: I1002 02:28:30.698143 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2rs6" event={"ID":"b50612db-5c36-47a5-b4ee-44f754df5de5","Type":"ContainerStarted","Data":"4f17579d257ece22cc90927112e396da9c3829195ef7d07e542fb794bd228609"} Oct 02 02:28:30 crc kubenswrapper[4775]: I1002 02:28:30.717563 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h2rs6" podStartSLOduration=2.24511407 podStartE2EDuration="4.717540983s" podCreationTimestamp="2025-10-02 02:28:26 +0000 UTC" firstStartedPulling="2025-10-02 02:28:27.675415495 +0000 UTC m=+2844.842159535" lastFinishedPulling="2025-10-02 02:28:30.147842368 +0000 UTC m=+2847.314586448" observedRunningTime="2025-10-02 02:28:30.716186058 +0000 UTC m=+2847.882930108" watchObservedRunningTime="2025-10-02 02:28:30.717540983 +0000 UTC m=+2847.884285023" Oct 02 02:28:36 crc kubenswrapper[4775]: I1002 02:28:36.522597 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:36 crc kubenswrapper[4775]: I1002 02:28:36.523059 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:36 crc kubenswrapper[4775]: I1002 02:28:36.568492 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:36 crc kubenswrapper[4775]: I1002 02:28:36.807473 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:36 crc kubenswrapper[4775]: I1002 02:28:36.878307 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2rs6"] Oct 02 02:28:37 crc kubenswrapper[4775]: I1002 02:28:37.233427 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:28:37 crc kubenswrapper[4775]: I1002 02:28:37.233511 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:28:38 crc kubenswrapper[4775]: I1002 02:28:38.764305 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h2rs6" podUID="b50612db-5c36-47a5-b4ee-44f754df5de5" containerName="registry-server" containerID="cri-o://4f17579d257ece22cc90927112e396da9c3829195ef7d07e542fb794bd228609" gracePeriod=2 Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.252648 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.324056 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b50612db-5c36-47a5-b4ee-44f754df5de5-catalog-content\") pod \"b50612db-5c36-47a5-b4ee-44f754df5de5\" (UID: \"b50612db-5c36-47a5-b4ee-44f754df5de5\") " Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.324272 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnk9z\" (UniqueName: \"kubernetes.io/projected/b50612db-5c36-47a5-b4ee-44f754df5de5-kube-api-access-mnk9z\") pod \"b50612db-5c36-47a5-b4ee-44f754df5de5\" (UID: \"b50612db-5c36-47a5-b4ee-44f754df5de5\") " Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.324340 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b50612db-5c36-47a5-b4ee-44f754df5de5-utilities\") pod \"b50612db-5c36-47a5-b4ee-44f754df5de5\" (UID: \"b50612db-5c36-47a5-b4ee-44f754df5de5\") " Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.326071 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b50612db-5c36-47a5-b4ee-44f754df5de5-utilities" (OuterVolumeSpecName: "utilities") pod "b50612db-5c36-47a5-b4ee-44f754df5de5" (UID: "b50612db-5c36-47a5-b4ee-44f754df5de5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.332158 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b50612db-5c36-47a5-b4ee-44f754df5de5-kube-api-access-mnk9z" (OuterVolumeSpecName: "kube-api-access-mnk9z") pod "b50612db-5c36-47a5-b4ee-44f754df5de5" (UID: "b50612db-5c36-47a5-b4ee-44f754df5de5"). InnerVolumeSpecName "kube-api-access-mnk9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.350635 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b50612db-5c36-47a5-b4ee-44f754df5de5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b50612db-5c36-47a5-b4ee-44f754df5de5" (UID: "b50612db-5c36-47a5-b4ee-44f754df5de5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.425709 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnk9z\" (UniqueName: \"kubernetes.io/projected/b50612db-5c36-47a5-b4ee-44f754df5de5-kube-api-access-mnk9z\") on node \"crc\" DevicePath \"\"" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.425762 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b50612db-5c36-47a5-b4ee-44f754df5de5-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.425773 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b50612db-5c36-47a5-b4ee-44f754df5de5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.777800 4775 generic.go:334] "Generic (PLEG): container finished" podID="b50612db-5c36-47a5-b4ee-44f754df5de5" containerID="4f17579d257ece22cc90927112e396da9c3829195ef7d07e542fb794bd228609" exitCode=0 Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.777848 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h2rs6" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.777892 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2rs6" event={"ID":"b50612db-5c36-47a5-b4ee-44f754df5de5","Type":"ContainerDied","Data":"4f17579d257ece22cc90927112e396da9c3829195ef7d07e542fb794bd228609"} Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.777991 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2rs6" event={"ID":"b50612db-5c36-47a5-b4ee-44f754df5de5","Type":"ContainerDied","Data":"0877fc74e163c468332cb3fedb1dae22a93c4eb5f5080251417f9d2f72b57f91"} Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.778034 4775 scope.go:117] "RemoveContainer" containerID="4f17579d257ece22cc90927112e396da9c3829195ef7d07e542fb794bd228609" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.821079 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2rs6"] Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.823912 4775 scope.go:117] "RemoveContainer" containerID="ce81b20fc937e8b6f77e8bfa2d851d7b00eab44736ea5772fe8f01868ad92629" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.829042 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2rs6"] Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.842047 4775 scope.go:117] "RemoveContainer" containerID="f07cb704dc654a550059f35b5adf90b13e80e27fb48f73290a151e82c01ff0a4" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.866436 4775 scope.go:117] "RemoveContainer" containerID="4f17579d257ece22cc90927112e396da9c3829195ef7d07e542fb794bd228609" Oct 02 02:28:39 crc kubenswrapper[4775]: E1002 02:28:39.867021 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f17579d257ece22cc90927112e396da9c3829195ef7d07e542fb794bd228609\": container with ID starting with 4f17579d257ece22cc90927112e396da9c3829195ef7d07e542fb794bd228609 not found: ID does not exist" containerID="4f17579d257ece22cc90927112e396da9c3829195ef7d07e542fb794bd228609" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.867062 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f17579d257ece22cc90927112e396da9c3829195ef7d07e542fb794bd228609"} err="failed to get container status \"4f17579d257ece22cc90927112e396da9c3829195ef7d07e542fb794bd228609\": rpc error: code = NotFound desc = could not find container \"4f17579d257ece22cc90927112e396da9c3829195ef7d07e542fb794bd228609\": container with ID starting with 4f17579d257ece22cc90927112e396da9c3829195ef7d07e542fb794bd228609 not found: ID does not exist" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.867082 4775 scope.go:117] "RemoveContainer" containerID="ce81b20fc937e8b6f77e8bfa2d851d7b00eab44736ea5772fe8f01868ad92629" Oct 02 02:28:39 crc kubenswrapper[4775]: E1002 02:28:39.867403 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce81b20fc937e8b6f77e8bfa2d851d7b00eab44736ea5772fe8f01868ad92629\": container with ID starting with ce81b20fc937e8b6f77e8bfa2d851d7b00eab44736ea5772fe8f01868ad92629 not found: ID does not exist" containerID="ce81b20fc937e8b6f77e8bfa2d851d7b00eab44736ea5772fe8f01868ad92629" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.867443 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce81b20fc937e8b6f77e8bfa2d851d7b00eab44736ea5772fe8f01868ad92629"} err="failed to get container status \"ce81b20fc937e8b6f77e8bfa2d851d7b00eab44736ea5772fe8f01868ad92629\": rpc error: code = NotFound desc = could not find container \"ce81b20fc937e8b6f77e8bfa2d851d7b00eab44736ea5772fe8f01868ad92629\": container with ID starting with ce81b20fc937e8b6f77e8bfa2d851d7b00eab44736ea5772fe8f01868ad92629 not found: ID does not exist" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.867462 4775 scope.go:117] "RemoveContainer" containerID="f07cb704dc654a550059f35b5adf90b13e80e27fb48f73290a151e82c01ff0a4" Oct 02 02:28:39 crc kubenswrapper[4775]: E1002 02:28:39.868231 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f07cb704dc654a550059f35b5adf90b13e80e27fb48f73290a151e82c01ff0a4\": container with ID starting with f07cb704dc654a550059f35b5adf90b13e80e27fb48f73290a151e82c01ff0a4 not found: ID does not exist" containerID="f07cb704dc654a550059f35b5adf90b13e80e27fb48f73290a151e82c01ff0a4" Oct 02 02:28:39 crc kubenswrapper[4775]: I1002 02:28:39.868294 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f07cb704dc654a550059f35b5adf90b13e80e27fb48f73290a151e82c01ff0a4"} err="failed to get container status \"f07cb704dc654a550059f35b5adf90b13e80e27fb48f73290a151e82c01ff0a4\": rpc error: code = NotFound desc = could not find container \"f07cb704dc654a550059f35b5adf90b13e80e27fb48f73290a151e82c01ff0a4\": container with ID starting with f07cb704dc654a550059f35b5adf90b13e80e27fb48f73290a151e82c01ff0a4 not found: ID does not exist" Oct 02 02:28:41 crc kubenswrapper[4775]: I1002 02:28:41.780905 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b50612db-5c36-47a5-b4ee-44f754df5de5" path="/var/lib/kubelet/pods/b50612db-5c36-47a5-b4ee-44f754df5de5/volumes" Oct 02 02:29:07 crc kubenswrapper[4775]: I1002 02:29:07.233405 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:29:07 crc kubenswrapper[4775]: I1002 02:29:07.233806 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.233901 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.235010 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.235053 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.235646 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.235718 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" gracePeriod=600 Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.293679 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jrj46"] Oct 02 02:29:37 crc kubenswrapper[4775]: E1002 02:29:37.294622 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b50612db-5c36-47a5-b4ee-44f754df5de5" containerName="extract-content" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.294651 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b50612db-5c36-47a5-b4ee-44f754df5de5" containerName="extract-content" Oct 02 02:29:37 crc kubenswrapper[4775]: E1002 02:29:37.294678 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b50612db-5c36-47a5-b4ee-44f754df5de5" containerName="registry-server" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.294691 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b50612db-5c36-47a5-b4ee-44f754df5de5" containerName="registry-server" Oct 02 02:29:37 crc kubenswrapper[4775]: E1002 02:29:37.294710 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b50612db-5c36-47a5-b4ee-44f754df5de5" containerName="extract-utilities" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.294724 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b50612db-5c36-47a5-b4ee-44f754df5de5" containerName="extract-utilities" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.295026 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b50612db-5c36-47a5-b4ee-44f754df5de5" containerName="registry-server" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.296681 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.313499 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jrj46"] Oct 02 02:29:37 crc kubenswrapper[4775]: E1002 02:29:37.365797 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.465671 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5pqk\" (UniqueName: \"kubernetes.io/projected/160fac8a-4e73-4cc9-9881-52dd305efe5a-kube-api-access-g5pqk\") pod \"certified-operators-jrj46\" (UID: \"160fac8a-4e73-4cc9-9881-52dd305efe5a\") " pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.465790 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/160fac8a-4e73-4cc9-9881-52dd305efe5a-catalog-content\") pod \"certified-operators-jrj46\" (UID: \"160fac8a-4e73-4cc9-9881-52dd305efe5a\") " pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.465835 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/160fac8a-4e73-4cc9-9881-52dd305efe5a-utilities\") pod \"certified-operators-jrj46\" (UID: \"160fac8a-4e73-4cc9-9881-52dd305efe5a\") " pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.567567 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5pqk\" (UniqueName: \"kubernetes.io/projected/160fac8a-4e73-4cc9-9881-52dd305efe5a-kube-api-access-g5pqk\") pod \"certified-operators-jrj46\" (UID: \"160fac8a-4e73-4cc9-9881-52dd305efe5a\") " pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.567649 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/160fac8a-4e73-4cc9-9881-52dd305efe5a-catalog-content\") pod \"certified-operators-jrj46\" (UID: \"160fac8a-4e73-4cc9-9881-52dd305efe5a\") " pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.567690 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/160fac8a-4e73-4cc9-9881-52dd305efe5a-utilities\") pod \"certified-operators-jrj46\" (UID: \"160fac8a-4e73-4cc9-9881-52dd305efe5a\") " pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.568409 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/160fac8a-4e73-4cc9-9881-52dd305efe5a-utilities\") pod \"certified-operators-jrj46\" (UID: \"160fac8a-4e73-4cc9-9881-52dd305efe5a\") " pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.569285 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/160fac8a-4e73-4cc9-9881-52dd305efe5a-catalog-content\") pod \"certified-operators-jrj46\" (UID: \"160fac8a-4e73-4cc9-9881-52dd305efe5a\") " pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.591004 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5pqk\" (UniqueName: \"kubernetes.io/projected/160fac8a-4e73-4cc9-9881-52dd305efe5a-kube-api-access-g5pqk\") pod \"certified-operators-jrj46\" (UID: \"160fac8a-4e73-4cc9-9881-52dd305efe5a\") " pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:37 crc kubenswrapper[4775]: I1002 02:29:37.620249 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:38 crc kubenswrapper[4775]: I1002 02:29:38.116539 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jrj46"] Oct 02 02:29:38 crc kubenswrapper[4775]: I1002 02:29:38.322077 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" exitCode=0 Oct 02 02:29:38 crc kubenswrapper[4775]: I1002 02:29:38.322125 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d"} Oct 02 02:29:38 crc kubenswrapper[4775]: I1002 02:29:38.322216 4775 scope.go:117] "RemoveContainer" containerID="3f87c732873861f7da11725e497cfd107b7bc6b663b852114f5d70f0a26de837" Oct 02 02:29:38 crc kubenswrapper[4775]: I1002 02:29:38.323040 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:29:38 crc kubenswrapper[4775]: E1002 02:29:38.323496 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:29:38 crc kubenswrapper[4775]: I1002 02:29:38.324232 4775 generic.go:334] "Generic (PLEG): container finished" podID="160fac8a-4e73-4cc9-9881-52dd305efe5a" containerID="b67b7f49b4834faae37582630a7a65382b170025a47f0cea992743973328e14b" exitCode=0 Oct 02 02:29:38 crc kubenswrapper[4775]: I1002 02:29:38.324276 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrj46" event={"ID":"160fac8a-4e73-4cc9-9881-52dd305efe5a","Type":"ContainerDied","Data":"b67b7f49b4834faae37582630a7a65382b170025a47f0cea992743973328e14b"} Oct 02 02:29:38 crc kubenswrapper[4775]: I1002 02:29:38.324314 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrj46" event={"ID":"160fac8a-4e73-4cc9-9881-52dd305efe5a","Type":"ContainerStarted","Data":"5dee06a71ac8d77250831010e2e3486ad1036bb2e0c8dddb9b93273e7d5dcc18"} Oct 02 02:29:39 crc kubenswrapper[4775]: I1002 02:29:39.334552 4775 generic.go:334] "Generic (PLEG): container finished" podID="160fac8a-4e73-4cc9-9881-52dd305efe5a" containerID="ed05868a49a37b595378169ec8370663ea05ecf140167d433dda41350aab5748" exitCode=0 Oct 02 02:29:39 crc kubenswrapper[4775]: I1002 02:29:39.335167 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrj46" event={"ID":"160fac8a-4e73-4cc9-9881-52dd305efe5a","Type":"ContainerDied","Data":"ed05868a49a37b595378169ec8370663ea05ecf140167d433dda41350aab5748"} Oct 02 02:29:40 crc kubenswrapper[4775]: I1002 02:29:40.348157 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrj46" event={"ID":"160fac8a-4e73-4cc9-9881-52dd305efe5a","Type":"ContainerStarted","Data":"951592201a75936dd74a65c85ba03cd335c4d072b4f6643e50ff61c6406cec6b"} Oct 02 02:29:40 crc kubenswrapper[4775]: I1002 02:29:40.368414 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jrj46" podStartSLOduration=1.8419509120000002 podStartE2EDuration="3.368396479s" podCreationTimestamp="2025-10-02 02:29:37 +0000 UTC" firstStartedPulling="2025-10-02 02:29:38.326349206 +0000 UTC m=+2915.493093256" lastFinishedPulling="2025-10-02 02:29:39.852794783 +0000 UTC m=+2917.019538823" observedRunningTime="2025-10-02 02:29:40.367614049 +0000 UTC m=+2917.534358129" watchObservedRunningTime="2025-10-02 02:29:40.368396479 +0000 UTC m=+2917.535140519" Oct 02 02:29:47 crc kubenswrapper[4775]: I1002 02:29:47.620773 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:47 crc kubenswrapper[4775]: I1002 02:29:47.621524 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:47 crc kubenswrapper[4775]: I1002 02:29:47.688860 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:48 crc kubenswrapper[4775]: I1002 02:29:48.472737 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:48 crc kubenswrapper[4775]: I1002 02:29:48.541608 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jrj46"] Oct 02 02:29:50 crc kubenswrapper[4775]: I1002 02:29:50.446885 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jrj46" podUID="160fac8a-4e73-4cc9-9881-52dd305efe5a" containerName="registry-server" containerID="cri-o://951592201a75936dd74a65c85ba03cd335c4d072b4f6643e50ff61c6406cec6b" gracePeriod=2 Oct 02 02:29:50 crc kubenswrapper[4775]: I1002 02:29:50.947496 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.084610 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/160fac8a-4e73-4cc9-9881-52dd305efe5a-catalog-content\") pod \"160fac8a-4e73-4cc9-9881-52dd305efe5a\" (UID: \"160fac8a-4e73-4cc9-9881-52dd305efe5a\") " Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.084692 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/160fac8a-4e73-4cc9-9881-52dd305efe5a-utilities\") pod \"160fac8a-4e73-4cc9-9881-52dd305efe5a\" (UID: \"160fac8a-4e73-4cc9-9881-52dd305efe5a\") " Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.084732 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5pqk\" (UniqueName: \"kubernetes.io/projected/160fac8a-4e73-4cc9-9881-52dd305efe5a-kube-api-access-g5pqk\") pod \"160fac8a-4e73-4cc9-9881-52dd305efe5a\" (UID: \"160fac8a-4e73-4cc9-9881-52dd305efe5a\") " Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.086187 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/160fac8a-4e73-4cc9-9881-52dd305efe5a-utilities" (OuterVolumeSpecName: "utilities") pod "160fac8a-4e73-4cc9-9881-52dd305efe5a" (UID: "160fac8a-4e73-4cc9-9881-52dd305efe5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.090651 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/160fac8a-4e73-4cc9-9881-52dd305efe5a-kube-api-access-g5pqk" (OuterVolumeSpecName: "kube-api-access-g5pqk") pod "160fac8a-4e73-4cc9-9881-52dd305efe5a" (UID: "160fac8a-4e73-4cc9-9881-52dd305efe5a"). InnerVolumeSpecName "kube-api-access-g5pqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.134397 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/160fac8a-4e73-4cc9-9881-52dd305efe5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "160fac8a-4e73-4cc9-9881-52dd305efe5a" (UID: "160fac8a-4e73-4cc9-9881-52dd305efe5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.187053 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/160fac8a-4e73-4cc9-9881-52dd305efe5a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.187104 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/160fac8a-4e73-4cc9-9881-52dd305efe5a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.187124 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5pqk\" (UniqueName: \"kubernetes.io/projected/160fac8a-4e73-4cc9-9881-52dd305efe5a-kube-api-access-g5pqk\") on node \"crc\" DevicePath \"\"" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.460878 4775 generic.go:334] "Generic (PLEG): container finished" podID="160fac8a-4e73-4cc9-9881-52dd305efe5a" containerID="951592201a75936dd74a65c85ba03cd335c4d072b4f6643e50ff61c6406cec6b" exitCode=0 Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.460938 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrj46" event={"ID":"160fac8a-4e73-4cc9-9881-52dd305efe5a","Type":"ContainerDied","Data":"951592201a75936dd74a65c85ba03cd335c4d072b4f6643e50ff61c6406cec6b"} Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.460973 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jrj46" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.461031 4775 scope.go:117] "RemoveContainer" containerID="951592201a75936dd74a65c85ba03cd335c4d072b4f6643e50ff61c6406cec6b" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.461013 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jrj46" event={"ID":"160fac8a-4e73-4cc9-9881-52dd305efe5a","Type":"ContainerDied","Data":"5dee06a71ac8d77250831010e2e3486ad1036bb2e0c8dddb9b93273e7d5dcc18"} Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.501103 4775 scope.go:117] "RemoveContainer" containerID="ed05868a49a37b595378169ec8370663ea05ecf140167d433dda41350aab5748" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.512552 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jrj46"] Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.522268 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jrj46"] Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.544356 4775 scope.go:117] "RemoveContainer" containerID="b67b7f49b4834faae37582630a7a65382b170025a47f0cea992743973328e14b" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.571717 4775 scope.go:117] "RemoveContainer" containerID="951592201a75936dd74a65c85ba03cd335c4d072b4f6643e50ff61c6406cec6b" Oct 02 02:29:51 crc kubenswrapper[4775]: E1002 02:29:51.572150 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"951592201a75936dd74a65c85ba03cd335c4d072b4f6643e50ff61c6406cec6b\": container with ID starting with 951592201a75936dd74a65c85ba03cd335c4d072b4f6643e50ff61c6406cec6b not found: ID does not exist" containerID="951592201a75936dd74a65c85ba03cd335c4d072b4f6643e50ff61c6406cec6b" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.572184 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"951592201a75936dd74a65c85ba03cd335c4d072b4f6643e50ff61c6406cec6b"} err="failed to get container status \"951592201a75936dd74a65c85ba03cd335c4d072b4f6643e50ff61c6406cec6b\": rpc error: code = NotFound desc = could not find container \"951592201a75936dd74a65c85ba03cd335c4d072b4f6643e50ff61c6406cec6b\": container with ID starting with 951592201a75936dd74a65c85ba03cd335c4d072b4f6643e50ff61c6406cec6b not found: ID does not exist" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.572204 4775 scope.go:117] "RemoveContainer" containerID="ed05868a49a37b595378169ec8370663ea05ecf140167d433dda41350aab5748" Oct 02 02:29:51 crc kubenswrapper[4775]: E1002 02:29:51.572840 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed05868a49a37b595378169ec8370663ea05ecf140167d433dda41350aab5748\": container with ID starting with ed05868a49a37b595378169ec8370663ea05ecf140167d433dda41350aab5748 not found: ID does not exist" containerID="ed05868a49a37b595378169ec8370663ea05ecf140167d433dda41350aab5748" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.572860 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed05868a49a37b595378169ec8370663ea05ecf140167d433dda41350aab5748"} err="failed to get container status \"ed05868a49a37b595378169ec8370663ea05ecf140167d433dda41350aab5748\": rpc error: code = NotFound desc = could not find container \"ed05868a49a37b595378169ec8370663ea05ecf140167d433dda41350aab5748\": container with ID starting with ed05868a49a37b595378169ec8370663ea05ecf140167d433dda41350aab5748 not found: ID does not exist" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.572875 4775 scope.go:117] "RemoveContainer" containerID="b67b7f49b4834faae37582630a7a65382b170025a47f0cea992743973328e14b" Oct 02 02:29:51 crc kubenswrapper[4775]: E1002 02:29:51.573331 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b67b7f49b4834faae37582630a7a65382b170025a47f0cea992743973328e14b\": container with ID starting with b67b7f49b4834faae37582630a7a65382b170025a47f0cea992743973328e14b not found: ID does not exist" containerID="b67b7f49b4834faae37582630a7a65382b170025a47f0cea992743973328e14b" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.573366 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b67b7f49b4834faae37582630a7a65382b170025a47f0cea992743973328e14b"} err="failed to get container status \"b67b7f49b4834faae37582630a7a65382b170025a47f0cea992743973328e14b\": rpc error: code = NotFound desc = could not find container \"b67b7f49b4834faae37582630a7a65382b170025a47f0cea992743973328e14b\": container with ID starting with b67b7f49b4834faae37582630a7a65382b170025a47f0cea992743973328e14b not found: ID does not exist" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.765658 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:29:51 crc kubenswrapper[4775]: E1002 02:29:51.766074 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:29:51 crc kubenswrapper[4775]: I1002 02:29:51.782177 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="160fac8a-4e73-4cc9-9881-52dd305efe5a" path="/var/lib/kubelet/pods/160fac8a-4e73-4cc9-9881-52dd305efe5a/volumes" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.189311 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95"] Oct 02 02:30:00 crc kubenswrapper[4775]: E1002 02:30:00.190163 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="160fac8a-4e73-4cc9-9881-52dd305efe5a" containerName="extract-utilities" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.190181 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="160fac8a-4e73-4cc9-9881-52dd305efe5a" containerName="extract-utilities" Oct 02 02:30:00 crc kubenswrapper[4775]: E1002 02:30:00.190196 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="160fac8a-4e73-4cc9-9881-52dd305efe5a" containerName="registry-server" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.190204 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="160fac8a-4e73-4cc9-9881-52dd305efe5a" containerName="registry-server" Oct 02 02:30:00 crc kubenswrapper[4775]: E1002 02:30:00.190227 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="160fac8a-4e73-4cc9-9881-52dd305efe5a" containerName="extract-content" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.190233 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="160fac8a-4e73-4cc9-9881-52dd305efe5a" containerName="extract-content" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.190409 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="160fac8a-4e73-4cc9-9881-52dd305efe5a" containerName="registry-server" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.190885 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.193083 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.193906 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.209104 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95"] Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.333764 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crsqp\" (UniqueName: \"kubernetes.io/projected/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-kube-api-access-crsqp\") pod \"collect-profiles-29322870-6lc95\" (UID: \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.334426 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-config-volume\") pod \"collect-profiles-29322870-6lc95\" (UID: \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.334675 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-secret-volume\") pod \"collect-profiles-29322870-6lc95\" (UID: \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.436186 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-config-volume\") pod \"collect-profiles-29322870-6lc95\" (UID: \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.436522 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-secret-volume\") pod \"collect-profiles-29322870-6lc95\" (UID: \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.436681 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crsqp\" (UniqueName: \"kubernetes.io/projected/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-kube-api-access-crsqp\") pod \"collect-profiles-29322870-6lc95\" (UID: \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.437054 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-config-volume\") pod \"collect-profiles-29322870-6lc95\" (UID: \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.443735 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-secret-volume\") pod \"collect-profiles-29322870-6lc95\" (UID: \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.456944 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crsqp\" (UniqueName: \"kubernetes.io/projected/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-kube-api-access-crsqp\") pod \"collect-profiles-29322870-6lc95\" (UID: \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.513421 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" Oct 02 02:30:00 crc kubenswrapper[4775]: I1002 02:30:00.936557 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95"] Oct 02 02:30:01 crc kubenswrapper[4775]: I1002 02:30:01.556049 4775 generic.go:334] "Generic (PLEG): container finished" podID="aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5" containerID="eae24cc858e46a80e35ec9e64496af5a44f475e2b250320c7378f243778f2f5e" exitCode=0 Oct 02 02:30:01 crc kubenswrapper[4775]: I1002 02:30:01.556114 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" event={"ID":"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5","Type":"ContainerDied","Data":"eae24cc858e46a80e35ec9e64496af5a44f475e2b250320c7378f243778f2f5e"} Oct 02 02:30:01 crc kubenswrapper[4775]: I1002 02:30:01.556486 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" event={"ID":"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5","Type":"ContainerStarted","Data":"71500c8914a598be83293fe43efda7103e41574c858241dbf2bd806aa38bccf0"} Oct 02 02:30:02 crc kubenswrapper[4775]: I1002 02:30:02.872594 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" Oct 02 02:30:02 crc kubenswrapper[4775]: I1002 02:30:02.975254 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-config-volume\") pod \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\" (UID: \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\") " Oct 02 02:30:02 crc kubenswrapper[4775]: I1002 02:30:02.975580 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crsqp\" (UniqueName: \"kubernetes.io/projected/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-kube-api-access-crsqp\") pod \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\" (UID: \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\") " Oct 02 02:30:02 crc kubenswrapper[4775]: I1002 02:30:02.975727 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-secret-volume\") pod \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\" (UID: \"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5\") " Oct 02 02:30:02 crc kubenswrapper[4775]: I1002 02:30:02.976099 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-config-volume" (OuterVolumeSpecName: "config-volume") pod "aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5" (UID: "aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:30:02 crc kubenswrapper[4775]: I1002 02:30:02.976316 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:30:02 crc kubenswrapper[4775]: I1002 02:30:02.981489 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5" (UID: "aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:30:02 crc kubenswrapper[4775]: I1002 02:30:02.982231 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-kube-api-access-crsqp" (OuterVolumeSpecName: "kube-api-access-crsqp") pod "aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5" (UID: "aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5"). InnerVolumeSpecName "kube-api-access-crsqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:30:03 crc kubenswrapper[4775]: I1002 02:30:03.077690 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crsqp\" (UniqueName: \"kubernetes.io/projected/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-kube-api-access-crsqp\") on node \"crc\" DevicePath \"\"" Oct 02 02:30:03 crc kubenswrapper[4775]: I1002 02:30:03.077833 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:30:03 crc kubenswrapper[4775]: I1002 02:30:03.574698 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" event={"ID":"aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5","Type":"ContainerDied","Data":"71500c8914a598be83293fe43efda7103e41574c858241dbf2bd806aa38bccf0"} Oct 02 02:30:03 crc kubenswrapper[4775]: I1002 02:30:03.574753 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71500c8914a598be83293fe43efda7103e41574c858241dbf2bd806aa38bccf0" Oct 02 02:30:03 crc kubenswrapper[4775]: I1002 02:30:03.574772 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95" Oct 02 02:30:03 crc kubenswrapper[4775]: I1002 02:30:03.770287 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:30:03 crc kubenswrapper[4775]: E1002 02:30:03.770577 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:30:03 crc kubenswrapper[4775]: I1002 02:30:03.951756 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2"] Oct 02 02:30:03 crc kubenswrapper[4775]: I1002 02:30:03.959690 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322825-pjpz2"] Oct 02 02:30:05 crc kubenswrapper[4775]: I1002 02:30:05.778881 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="406d7724-813a-4299-90d3-c48c9f5b139e" path="/var/lib/kubelet/pods/406d7724-813a-4299-90d3-c48c9f5b139e/volumes" Oct 02 02:30:09 crc kubenswrapper[4775]: I1002 02:30:09.268770 4775 scope.go:117] "RemoveContainer" containerID="9ff66b564c8f4c4ce9b4c0bb0f26bd91e314c00fafbb2218aefca1f7f91394cc" Oct 02 02:30:16 crc kubenswrapper[4775]: I1002 02:30:16.765792 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:30:16 crc kubenswrapper[4775]: E1002 02:30:16.766885 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:30:28 crc kubenswrapper[4775]: I1002 02:30:28.766360 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:30:28 crc kubenswrapper[4775]: E1002 02:30:28.767091 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:30:39 crc kubenswrapper[4775]: I1002 02:30:39.766024 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:30:39 crc kubenswrapper[4775]: E1002 02:30:39.768007 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:30:50 crc kubenswrapper[4775]: I1002 02:30:50.765361 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:30:50 crc kubenswrapper[4775]: E1002 02:30:50.766360 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:31:01 crc kubenswrapper[4775]: I1002 02:31:01.766567 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:31:01 crc kubenswrapper[4775]: E1002 02:31:01.768839 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:31:16 crc kubenswrapper[4775]: I1002 02:31:16.765298 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:31:16 crc kubenswrapper[4775]: E1002 02:31:16.766340 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:31:31 crc kubenswrapper[4775]: I1002 02:31:31.765893 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:31:31 crc kubenswrapper[4775]: E1002 02:31:31.766883 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:31:45 crc kubenswrapper[4775]: I1002 02:31:45.766199 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:31:45 crc kubenswrapper[4775]: E1002 02:31:45.767478 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:31:58 crc kubenswrapper[4775]: I1002 02:31:58.765389 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:31:58 crc kubenswrapper[4775]: E1002 02:31:58.766286 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:32:09 crc kubenswrapper[4775]: I1002 02:32:09.765872 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:32:09 crc kubenswrapper[4775]: E1002 02:32:09.766939 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:32:22 crc kubenswrapper[4775]: I1002 02:32:22.766409 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:32:22 crc kubenswrapper[4775]: E1002 02:32:22.767684 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:32:33 crc kubenswrapper[4775]: I1002 02:32:33.775123 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:32:33 crc kubenswrapper[4775]: E1002 02:32:33.776227 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:32:47 crc kubenswrapper[4775]: I1002 02:32:47.766584 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:32:47 crc kubenswrapper[4775]: E1002 02:32:47.767703 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:33:01 crc kubenswrapper[4775]: I1002 02:33:01.765288 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:33:01 crc kubenswrapper[4775]: E1002 02:33:01.767693 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:33:16 crc kubenswrapper[4775]: I1002 02:33:16.765358 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:33:16 crc kubenswrapper[4775]: E1002 02:33:16.766283 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:33:27 crc kubenswrapper[4775]: I1002 02:33:27.765747 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:33:27 crc kubenswrapper[4775]: E1002 02:33:27.766676 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:33:38 crc kubenswrapper[4775]: I1002 02:33:38.764840 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:33:38 crc kubenswrapper[4775]: E1002 02:33:38.765596 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:33:50 crc kubenswrapper[4775]: I1002 02:33:50.765668 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:33:50 crc kubenswrapper[4775]: E1002 02:33:50.766647 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:34:02 crc kubenswrapper[4775]: I1002 02:34:02.765178 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:34:02 crc kubenswrapper[4775]: E1002 02:34:02.767152 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:34:14 crc kubenswrapper[4775]: I1002 02:34:14.765649 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:34:14 crc kubenswrapper[4775]: E1002 02:34:14.766802 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:34:29 crc kubenswrapper[4775]: I1002 02:34:29.765436 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:34:29 crc kubenswrapper[4775]: E1002 02:34:29.767417 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:34:41 crc kubenswrapper[4775]: I1002 02:34:41.765130 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:34:42 crc kubenswrapper[4775]: I1002 02:34:42.143779 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"dbcebe30237c32f26c9ca1f57bfd5e7480a1a0e125a47d39113cbb556f02abc5"} Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.000911 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-692bb"] Oct 02 02:36:24 crc kubenswrapper[4775]: E1002 02:36:24.002269 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5" containerName="collect-profiles" Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.002299 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5" containerName="collect-profiles" Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.002703 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5" containerName="collect-profiles" Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.005148 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.025202 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-692bb"] Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.095599 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e1912d-1a71-4372-b6ef-2e08a53bf302-utilities\") pod \"community-operators-692bb\" (UID: \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\") " pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.095687 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e1912d-1a71-4372-b6ef-2e08a53bf302-catalog-content\") pod \"community-operators-692bb\" (UID: \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\") " pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.095974 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7jwg\" (UniqueName: \"kubernetes.io/projected/d5e1912d-1a71-4372-b6ef-2e08a53bf302-kube-api-access-b7jwg\") pod \"community-operators-692bb\" (UID: \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\") " pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.197709 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e1912d-1a71-4372-b6ef-2e08a53bf302-utilities\") pod \"community-operators-692bb\" (UID: \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\") " pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.197756 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e1912d-1a71-4372-b6ef-2e08a53bf302-catalog-content\") pod \"community-operators-692bb\" (UID: \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\") " pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.197785 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7jwg\" (UniqueName: \"kubernetes.io/projected/d5e1912d-1a71-4372-b6ef-2e08a53bf302-kube-api-access-b7jwg\") pod \"community-operators-692bb\" (UID: \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\") " pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.198432 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e1912d-1a71-4372-b6ef-2e08a53bf302-utilities\") pod \"community-operators-692bb\" (UID: \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\") " pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.198552 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e1912d-1a71-4372-b6ef-2e08a53bf302-catalog-content\") pod \"community-operators-692bb\" (UID: \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\") " pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.227910 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7jwg\" (UniqueName: \"kubernetes.io/projected/d5e1912d-1a71-4372-b6ef-2e08a53bf302-kube-api-access-b7jwg\") pod \"community-operators-692bb\" (UID: \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\") " pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.344241 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:24 crc kubenswrapper[4775]: I1002 02:36:24.879329 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-692bb"] Oct 02 02:36:25 crc kubenswrapper[4775]: I1002 02:36:25.120673 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-692bb" event={"ID":"d5e1912d-1a71-4372-b6ef-2e08a53bf302","Type":"ContainerStarted","Data":"b636d45284117b0104471b217c2815367f671e9886b754da90143c7a9d0dca02"} Oct 02 02:36:25 crc kubenswrapper[4775]: I1002 02:36:25.121241 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-692bb" event={"ID":"d5e1912d-1a71-4372-b6ef-2e08a53bf302","Type":"ContainerStarted","Data":"b5d84bca7f6f09497301c9f0def46352fff4f3522bd2442905d3134ee69046fe"} Oct 02 02:36:26 crc kubenswrapper[4775]: I1002 02:36:26.132664 4775 generic.go:334] "Generic (PLEG): container finished" podID="d5e1912d-1a71-4372-b6ef-2e08a53bf302" containerID="b636d45284117b0104471b217c2815367f671e9886b754da90143c7a9d0dca02" exitCode=0 Oct 02 02:36:26 crc kubenswrapper[4775]: I1002 02:36:26.132729 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-692bb" event={"ID":"d5e1912d-1a71-4372-b6ef-2e08a53bf302","Type":"ContainerDied","Data":"b636d45284117b0104471b217c2815367f671e9886b754da90143c7a9d0dca02"} Oct 02 02:36:26 crc kubenswrapper[4775]: I1002 02:36:26.136405 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 02:36:27 crc kubenswrapper[4775]: I1002 02:36:27.145535 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-692bb" event={"ID":"d5e1912d-1a71-4372-b6ef-2e08a53bf302","Type":"ContainerStarted","Data":"16667e1e0bee889e946396dc9b1f69b182bff8a2e697c94988676cf1736881e8"} Oct 02 02:36:28 crc kubenswrapper[4775]: I1002 02:36:28.158910 4775 generic.go:334] "Generic (PLEG): container finished" podID="d5e1912d-1a71-4372-b6ef-2e08a53bf302" containerID="16667e1e0bee889e946396dc9b1f69b182bff8a2e697c94988676cf1736881e8" exitCode=0 Oct 02 02:36:28 crc kubenswrapper[4775]: I1002 02:36:28.159007 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-692bb" event={"ID":"d5e1912d-1a71-4372-b6ef-2e08a53bf302","Type":"ContainerDied","Data":"16667e1e0bee889e946396dc9b1f69b182bff8a2e697c94988676cf1736881e8"} Oct 02 02:36:29 crc kubenswrapper[4775]: I1002 02:36:29.170333 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-692bb" event={"ID":"d5e1912d-1a71-4372-b6ef-2e08a53bf302","Type":"ContainerStarted","Data":"3aca8518f963d128419f5d2cf13f20caec5124289589ab9987ec2fbaf894d105"} Oct 02 02:36:29 crc kubenswrapper[4775]: I1002 02:36:29.190182 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-692bb" podStartSLOduration=3.72611261 podStartE2EDuration="6.190166524s" podCreationTimestamp="2025-10-02 02:36:23 +0000 UTC" firstStartedPulling="2025-10-02 02:36:26.136099571 +0000 UTC m=+3323.302843621" lastFinishedPulling="2025-10-02 02:36:28.600153485 +0000 UTC m=+3325.766897535" observedRunningTime="2025-10-02 02:36:29.187489895 +0000 UTC m=+3326.354233945" watchObservedRunningTime="2025-10-02 02:36:29.190166524 +0000 UTC m=+3326.356910574" Oct 02 02:36:32 crc kubenswrapper[4775]: I1002 02:36:32.366405 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fc9xs"] Oct 02 02:36:32 crc kubenswrapper[4775]: I1002 02:36:32.369439 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:32 crc kubenswrapper[4775]: I1002 02:36:32.388866 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fc9xs"] Oct 02 02:36:32 crc kubenswrapper[4775]: I1002 02:36:32.528301 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81b45fe-739c-4acf-a91d-21d565484ec9-utilities\") pod \"redhat-operators-fc9xs\" (UID: \"e81b45fe-739c-4acf-a91d-21d565484ec9\") " pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:32 crc kubenswrapper[4775]: I1002 02:36:32.528425 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81b45fe-739c-4acf-a91d-21d565484ec9-catalog-content\") pod \"redhat-operators-fc9xs\" (UID: \"e81b45fe-739c-4acf-a91d-21d565484ec9\") " pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:32 crc kubenswrapper[4775]: I1002 02:36:32.528452 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8hwx\" (UniqueName: \"kubernetes.io/projected/e81b45fe-739c-4acf-a91d-21d565484ec9-kube-api-access-t8hwx\") pod \"redhat-operators-fc9xs\" (UID: \"e81b45fe-739c-4acf-a91d-21d565484ec9\") " pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:32 crc kubenswrapper[4775]: I1002 02:36:32.629679 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8hwx\" (UniqueName: \"kubernetes.io/projected/e81b45fe-739c-4acf-a91d-21d565484ec9-kube-api-access-t8hwx\") pod \"redhat-operators-fc9xs\" (UID: \"e81b45fe-739c-4acf-a91d-21d565484ec9\") " pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:32 crc kubenswrapper[4775]: I1002 02:36:32.630040 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81b45fe-739c-4acf-a91d-21d565484ec9-utilities\") pod \"redhat-operators-fc9xs\" (UID: \"e81b45fe-739c-4acf-a91d-21d565484ec9\") " pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:32 crc kubenswrapper[4775]: I1002 02:36:32.630187 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81b45fe-739c-4acf-a91d-21d565484ec9-catalog-content\") pod \"redhat-operators-fc9xs\" (UID: \"e81b45fe-739c-4acf-a91d-21d565484ec9\") " pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:32 crc kubenswrapper[4775]: I1002 02:36:32.630792 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81b45fe-739c-4acf-a91d-21d565484ec9-catalog-content\") pod \"redhat-operators-fc9xs\" (UID: \"e81b45fe-739c-4acf-a91d-21d565484ec9\") " pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:32 crc kubenswrapper[4775]: I1002 02:36:32.630791 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81b45fe-739c-4acf-a91d-21d565484ec9-utilities\") pod \"redhat-operators-fc9xs\" (UID: \"e81b45fe-739c-4acf-a91d-21d565484ec9\") " pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:32 crc kubenswrapper[4775]: I1002 02:36:32.656639 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8hwx\" (UniqueName: \"kubernetes.io/projected/e81b45fe-739c-4acf-a91d-21d565484ec9-kube-api-access-t8hwx\") pod \"redhat-operators-fc9xs\" (UID: \"e81b45fe-739c-4acf-a91d-21d565484ec9\") " pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:32 crc kubenswrapper[4775]: I1002 02:36:32.704010 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:33 crc kubenswrapper[4775]: I1002 02:36:33.158672 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fc9xs"] Oct 02 02:36:33 crc kubenswrapper[4775]: I1002 02:36:33.221656 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fc9xs" event={"ID":"e81b45fe-739c-4acf-a91d-21d565484ec9","Type":"ContainerStarted","Data":"ca09ae34219042cc0a05697bfc1896563631e9d37b4b8a2d12534d0c950660bd"} Oct 02 02:36:34 crc kubenswrapper[4775]: I1002 02:36:34.233632 4775 generic.go:334] "Generic (PLEG): container finished" podID="e81b45fe-739c-4acf-a91d-21d565484ec9" containerID="996523ccd2c4efa0dbe886319e68d00160c9c0e388e002e5509dd02cdffa323f" exitCode=0 Oct 02 02:36:34 crc kubenswrapper[4775]: I1002 02:36:34.233699 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fc9xs" event={"ID":"e81b45fe-739c-4acf-a91d-21d565484ec9","Type":"ContainerDied","Data":"996523ccd2c4efa0dbe886319e68d00160c9c0e388e002e5509dd02cdffa323f"} Oct 02 02:36:34 crc kubenswrapper[4775]: I1002 02:36:34.345152 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:34 crc kubenswrapper[4775]: I1002 02:36:34.345242 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:34 crc kubenswrapper[4775]: I1002 02:36:34.417420 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:35 crc kubenswrapper[4775]: I1002 02:36:35.243862 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fc9xs" event={"ID":"e81b45fe-739c-4acf-a91d-21d565484ec9","Type":"ContainerStarted","Data":"495315bdb9cbddbec4b6a8e26be38d5a3b18afad02fed9be93b5192cabc0371f"} Oct 02 02:36:35 crc kubenswrapper[4775]: I1002 02:36:35.313319 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:36 crc kubenswrapper[4775]: I1002 02:36:36.256028 4775 generic.go:334] "Generic (PLEG): container finished" podID="e81b45fe-739c-4acf-a91d-21d565484ec9" containerID="495315bdb9cbddbec4b6a8e26be38d5a3b18afad02fed9be93b5192cabc0371f" exitCode=0 Oct 02 02:36:36 crc kubenswrapper[4775]: I1002 02:36:36.256158 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fc9xs" event={"ID":"e81b45fe-739c-4acf-a91d-21d565484ec9","Type":"ContainerDied","Data":"495315bdb9cbddbec4b6a8e26be38d5a3b18afad02fed9be93b5192cabc0371f"} Oct 02 02:36:36 crc kubenswrapper[4775]: I1002 02:36:36.760443 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-692bb"] Oct 02 02:36:37 crc kubenswrapper[4775]: I1002 02:36:37.267923 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fc9xs" event={"ID":"e81b45fe-739c-4acf-a91d-21d565484ec9","Type":"ContainerStarted","Data":"13eebac5e5319765d24268c613c310a4c79ee4e7dc1521a9990ef5ae36ed700f"} Oct 02 02:36:37 crc kubenswrapper[4775]: I1002 02:36:37.268118 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-692bb" podUID="d5e1912d-1a71-4372-b6ef-2e08a53bf302" containerName="registry-server" containerID="cri-o://3aca8518f963d128419f5d2cf13f20caec5124289589ab9987ec2fbaf894d105" gracePeriod=2 Oct 02 02:36:37 crc kubenswrapper[4775]: I1002 02:36:37.297892 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fc9xs" podStartSLOduration=2.764311246 podStartE2EDuration="5.297866822s" podCreationTimestamp="2025-10-02 02:36:32 +0000 UTC" firstStartedPulling="2025-10-02 02:36:34.235793892 +0000 UTC m=+3331.402537962" lastFinishedPulling="2025-10-02 02:36:36.769349468 +0000 UTC m=+3333.936093538" observedRunningTime="2025-10-02 02:36:37.293753616 +0000 UTC m=+3334.460497656" watchObservedRunningTime="2025-10-02 02:36:37.297866822 +0000 UTC m=+3334.464610892" Oct 02 02:36:37 crc kubenswrapper[4775]: I1002 02:36:37.742178 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:37 crc kubenswrapper[4775]: I1002 02:36:37.810043 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e1912d-1a71-4372-b6ef-2e08a53bf302-utilities\") pod \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\" (UID: \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\") " Oct 02 02:36:37 crc kubenswrapper[4775]: I1002 02:36:37.810174 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e1912d-1a71-4372-b6ef-2e08a53bf302-catalog-content\") pod \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\" (UID: \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\") " Oct 02 02:36:37 crc kubenswrapper[4775]: I1002 02:36:37.810283 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7jwg\" (UniqueName: \"kubernetes.io/projected/d5e1912d-1a71-4372-b6ef-2e08a53bf302-kube-api-access-b7jwg\") pod \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\" (UID: \"d5e1912d-1a71-4372-b6ef-2e08a53bf302\") " Oct 02 02:36:37 crc kubenswrapper[4775]: I1002 02:36:37.812554 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5e1912d-1a71-4372-b6ef-2e08a53bf302-utilities" (OuterVolumeSpecName: "utilities") pod "d5e1912d-1a71-4372-b6ef-2e08a53bf302" (UID: "d5e1912d-1a71-4372-b6ef-2e08a53bf302"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:36:37 crc kubenswrapper[4775]: I1002 02:36:37.817023 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5e1912d-1a71-4372-b6ef-2e08a53bf302-kube-api-access-b7jwg" (OuterVolumeSpecName: "kube-api-access-b7jwg") pod "d5e1912d-1a71-4372-b6ef-2e08a53bf302" (UID: "d5e1912d-1a71-4372-b6ef-2e08a53bf302"). InnerVolumeSpecName "kube-api-access-b7jwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:36:37 crc kubenswrapper[4775]: I1002 02:36:37.890694 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5e1912d-1a71-4372-b6ef-2e08a53bf302-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5e1912d-1a71-4372-b6ef-2e08a53bf302" (UID: "d5e1912d-1a71-4372-b6ef-2e08a53bf302"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:36:37 crc kubenswrapper[4775]: I1002 02:36:37.912307 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5e1912d-1a71-4372-b6ef-2e08a53bf302-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:36:37 crc kubenswrapper[4775]: I1002 02:36:37.912345 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7jwg\" (UniqueName: \"kubernetes.io/projected/d5e1912d-1a71-4372-b6ef-2e08a53bf302-kube-api-access-b7jwg\") on node \"crc\" DevicePath \"\"" Oct 02 02:36:37 crc kubenswrapper[4775]: I1002 02:36:37.912359 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5e1912d-1a71-4372-b6ef-2e08a53bf302-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.279841 4775 generic.go:334] "Generic (PLEG): container finished" podID="d5e1912d-1a71-4372-b6ef-2e08a53bf302" containerID="3aca8518f963d128419f5d2cf13f20caec5124289589ab9987ec2fbaf894d105" exitCode=0 Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.279948 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-692bb" event={"ID":"d5e1912d-1a71-4372-b6ef-2e08a53bf302","Type":"ContainerDied","Data":"3aca8518f963d128419f5d2cf13f20caec5124289589ab9987ec2fbaf894d105"} Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.280004 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-692bb" Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.280073 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-692bb" event={"ID":"d5e1912d-1a71-4372-b6ef-2e08a53bf302","Type":"ContainerDied","Data":"b5d84bca7f6f09497301c9f0def46352fff4f3522bd2442905d3134ee69046fe"} Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.280115 4775 scope.go:117] "RemoveContainer" containerID="3aca8518f963d128419f5d2cf13f20caec5124289589ab9987ec2fbaf894d105" Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.313166 4775 scope.go:117] "RemoveContainer" containerID="16667e1e0bee889e946396dc9b1f69b182bff8a2e697c94988676cf1736881e8" Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.348692 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-692bb"] Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.355385 4775 scope.go:117] "RemoveContainer" containerID="b636d45284117b0104471b217c2815367f671e9886b754da90143c7a9d0dca02" Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.359041 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-692bb"] Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.391585 4775 scope.go:117] "RemoveContainer" containerID="3aca8518f963d128419f5d2cf13f20caec5124289589ab9987ec2fbaf894d105" Oct 02 02:36:38 crc kubenswrapper[4775]: E1002 02:36:38.396694 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3aca8518f963d128419f5d2cf13f20caec5124289589ab9987ec2fbaf894d105\": container with ID starting with 3aca8518f963d128419f5d2cf13f20caec5124289589ab9987ec2fbaf894d105 not found: ID does not exist" containerID="3aca8518f963d128419f5d2cf13f20caec5124289589ab9987ec2fbaf894d105" Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.396762 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3aca8518f963d128419f5d2cf13f20caec5124289589ab9987ec2fbaf894d105"} err="failed to get container status \"3aca8518f963d128419f5d2cf13f20caec5124289589ab9987ec2fbaf894d105\": rpc error: code = NotFound desc = could not find container \"3aca8518f963d128419f5d2cf13f20caec5124289589ab9987ec2fbaf894d105\": container with ID starting with 3aca8518f963d128419f5d2cf13f20caec5124289589ab9987ec2fbaf894d105 not found: ID does not exist" Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.396806 4775 scope.go:117] "RemoveContainer" containerID="16667e1e0bee889e946396dc9b1f69b182bff8a2e697c94988676cf1736881e8" Oct 02 02:36:38 crc kubenswrapper[4775]: E1002 02:36:38.397320 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16667e1e0bee889e946396dc9b1f69b182bff8a2e697c94988676cf1736881e8\": container with ID starting with 16667e1e0bee889e946396dc9b1f69b182bff8a2e697c94988676cf1736881e8 not found: ID does not exist" containerID="16667e1e0bee889e946396dc9b1f69b182bff8a2e697c94988676cf1736881e8" Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.397363 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16667e1e0bee889e946396dc9b1f69b182bff8a2e697c94988676cf1736881e8"} err="failed to get container status \"16667e1e0bee889e946396dc9b1f69b182bff8a2e697c94988676cf1736881e8\": rpc error: code = NotFound desc = could not find container \"16667e1e0bee889e946396dc9b1f69b182bff8a2e697c94988676cf1736881e8\": container with ID starting with 16667e1e0bee889e946396dc9b1f69b182bff8a2e697c94988676cf1736881e8 not found: ID does not exist" Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.397395 4775 scope.go:117] "RemoveContainer" containerID="b636d45284117b0104471b217c2815367f671e9886b754da90143c7a9d0dca02" Oct 02 02:36:38 crc kubenswrapper[4775]: E1002 02:36:38.397717 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b636d45284117b0104471b217c2815367f671e9886b754da90143c7a9d0dca02\": container with ID starting with b636d45284117b0104471b217c2815367f671e9886b754da90143c7a9d0dca02 not found: ID does not exist" containerID="b636d45284117b0104471b217c2815367f671e9886b754da90143c7a9d0dca02" Oct 02 02:36:38 crc kubenswrapper[4775]: I1002 02:36:38.397747 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b636d45284117b0104471b217c2815367f671e9886b754da90143c7a9d0dca02"} err="failed to get container status \"b636d45284117b0104471b217c2815367f671e9886b754da90143c7a9d0dca02\": rpc error: code = NotFound desc = could not find container \"b636d45284117b0104471b217c2815367f671e9886b754da90143c7a9d0dca02\": container with ID starting with b636d45284117b0104471b217c2815367f671e9886b754da90143c7a9d0dca02 not found: ID does not exist" Oct 02 02:36:39 crc kubenswrapper[4775]: I1002 02:36:39.783087 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5e1912d-1a71-4372-b6ef-2e08a53bf302" path="/var/lib/kubelet/pods/d5e1912d-1a71-4372-b6ef-2e08a53bf302/volumes" Oct 02 02:36:42 crc kubenswrapper[4775]: I1002 02:36:42.705111 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:42 crc kubenswrapper[4775]: I1002 02:36:42.705192 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:43 crc kubenswrapper[4775]: I1002 02:36:43.783754 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fc9xs" podUID="e81b45fe-739c-4acf-a91d-21d565484ec9" containerName="registry-server" probeResult="failure" output=< Oct 02 02:36:43 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 02:36:43 crc kubenswrapper[4775]: > Oct 02 02:36:52 crc kubenswrapper[4775]: I1002 02:36:52.777970 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:52 crc kubenswrapper[4775]: I1002 02:36:52.842375 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:53 crc kubenswrapper[4775]: I1002 02:36:53.028735 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fc9xs"] Oct 02 02:36:54 crc kubenswrapper[4775]: I1002 02:36:54.440329 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fc9xs" podUID="e81b45fe-739c-4acf-a91d-21d565484ec9" containerName="registry-server" containerID="cri-o://13eebac5e5319765d24268c613c310a4c79ee4e7dc1521a9990ef5ae36ed700f" gracePeriod=2 Oct 02 02:36:54 crc kubenswrapper[4775]: I1002 02:36:54.865341 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:54 crc kubenswrapper[4775]: I1002 02:36:54.993706 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81b45fe-739c-4acf-a91d-21d565484ec9-utilities\") pod \"e81b45fe-739c-4acf-a91d-21d565484ec9\" (UID: \"e81b45fe-739c-4acf-a91d-21d565484ec9\") " Oct 02 02:36:54 crc kubenswrapper[4775]: I1002 02:36:54.993777 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8hwx\" (UniqueName: \"kubernetes.io/projected/e81b45fe-739c-4acf-a91d-21d565484ec9-kube-api-access-t8hwx\") pod \"e81b45fe-739c-4acf-a91d-21d565484ec9\" (UID: \"e81b45fe-739c-4acf-a91d-21d565484ec9\") " Oct 02 02:36:54 crc kubenswrapper[4775]: I1002 02:36:54.993901 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81b45fe-739c-4acf-a91d-21d565484ec9-catalog-content\") pod \"e81b45fe-739c-4acf-a91d-21d565484ec9\" (UID: \"e81b45fe-739c-4acf-a91d-21d565484ec9\") " Oct 02 02:36:54 crc kubenswrapper[4775]: I1002 02:36:54.995310 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e81b45fe-739c-4acf-a91d-21d565484ec9-utilities" (OuterVolumeSpecName: "utilities") pod "e81b45fe-739c-4acf-a91d-21d565484ec9" (UID: "e81b45fe-739c-4acf-a91d-21d565484ec9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.000082 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e81b45fe-739c-4acf-a91d-21d565484ec9-kube-api-access-t8hwx" (OuterVolumeSpecName: "kube-api-access-t8hwx") pod "e81b45fe-739c-4acf-a91d-21d565484ec9" (UID: "e81b45fe-739c-4acf-a91d-21d565484ec9"). InnerVolumeSpecName "kube-api-access-t8hwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.095740 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81b45fe-739c-4acf-a91d-21d565484ec9-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.095790 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8hwx\" (UniqueName: \"kubernetes.io/projected/e81b45fe-739c-4acf-a91d-21d565484ec9-kube-api-access-t8hwx\") on node \"crc\" DevicePath \"\"" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.108989 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e81b45fe-739c-4acf-a91d-21d565484ec9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e81b45fe-739c-4acf-a91d-21d565484ec9" (UID: "e81b45fe-739c-4acf-a91d-21d565484ec9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.197772 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81b45fe-739c-4acf-a91d-21d565484ec9-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.451832 4775 generic.go:334] "Generic (PLEG): container finished" podID="e81b45fe-739c-4acf-a91d-21d565484ec9" containerID="13eebac5e5319765d24268c613c310a4c79ee4e7dc1521a9990ef5ae36ed700f" exitCode=0 Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.451922 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fc9xs" event={"ID":"e81b45fe-739c-4acf-a91d-21d565484ec9","Type":"ContainerDied","Data":"13eebac5e5319765d24268c613c310a4c79ee4e7dc1521a9990ef5ae36ed700f"} Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.452016 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fc9xs" event={"ID":"e81b45fe-739c-4acf-a91d-21d565484ec9","Type":"ContainerDied","Data":"ca09ae34219042cc0a05697bfc1896563631e9d37b4b8a2d12534d0c950660bd"} Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.452015 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fc9xs" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.452140 4775 scope.go:117] "RemoveContainer" containerID="13eebac5e5319765d24268c613c310a4c79ee4e7dc1521a9990ef5ae36ed700f" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.489479 4775 scope.go:117] "RemoveContainer" containerID="495315bdb9cbddbec4b6a8e26be38d5a3b18afad02fed9be93b5192cabc0371f" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.523498 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fc9xs"] Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.532766 4775 scope.go:117] "RemoveContainer" containerID="996523ccd2c4efa0dbe886319e68d00160c9c0e388e002e5509dd02cdffa323f" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.536433 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fc9xs"] Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.572908 4775 scope.go:117] "RemoveContainer" containerID="13eebac5e5319765d24268c613c310a4c79ee4e7dc1521a9990ef5ae36ed700f" Oct 02 02:36:55 crc kubenswrapper[4775]: E1002 02:36:55.573995 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13eebac5e5319765d24268c613c310a4c79ee4e7dc1521a9990ef5ae36ed700f\": container with ID starting with 13eebac5e5319765d24268c613c310a4c79ee4e7dc1521a9990ef5ae36ed700f not found: ID does not exist" containerID="13eebac5e5319765d24268c613c310a4c79ee4e7dc1521a9990ef5ae36ed700f" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.574050 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13eebac5e5319765d24268c613c310a4c79ee4e7dc1521a9990ef5ae36ed700f"} err="failed to get container status \"13eebac5e5319765d24268c613c310a4c79ee4e7dc1521a9990ef5ae36ed700f\": rpc error: code = NotFound desc = could not find container \"13eebac5e5319765d24268c613c310a4c79ee4e7dc1521a9990ef5ae36ed700f\": container with ID starting with 13eebac5e5319765d24268c613c310a4c79ee4e7dc1521a9990ef5ae36ed700f not found: ID does not exist" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.574081 4775 scope.go:117] "RemoveContainer" containerID="495315bdb9cbddbec4b6a8e26be38d5a3b18afad02fed9be93b5192cabc0371f" Oct 02 02:36:55 crc kubenswrapper[4775]: E1002 02:36:55.574710 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"495315bdb9cbddbec4b6a8e26be38d5a3b18afad02fed9be93b5192cabc0371f\": container with ID starting with 495315bdb9cbddbec4b6a8e26be38d5a3b18afad02fed9be93b5192cabc0371f not found: ID does not exist" containerID="495315bdb9cbddbec4b6a8e26be38d5a3b18afad02fed9be93b5192cabc0371f" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.574754 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"495315bdb9cbddbec4b6a8e26be38d5a3b18afad02fed9be93b5192cabc0371f"} err="failed to get container status \"495315bdb9cbddbec4b6a8e26be38d5a3b18afad02fed9be93b5192cabc0371f\": rpc error: code = NotFound desc = could not find container \"495315bdb9cbddbec4b6a8e26be38d5a3b18afad02fed9be93b5192cabc0371f\": container with ID starting with 495315bdb9cbddbec4b6a8e26be38d5a3b18afad02fed9be93b5192cabc0371f not found: ID does not exist" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.574786 4775 scope.go:117] "RemoveContainer" containerID="996523ccd2c4efa0dbe886319e68d00160c9c0e388e002e5509dd02cdffa323f" Oct 02 02:36:55 crc kubenswrapper[4775]: E1002 02:36:55.575172 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"996523ccd2c4efa0dbe886319e68d00160c9c0e388e002e5509dd02cdffa323f\": container with ID starting with 996523ccd2c4efa0dbe886319e68d00160c9c0e388e002e5509dd02cdffa323f not found: ID does not exist" containerID="996523ccd2c4efa0dbe886319e68d00160c9c0e388e002e5509dd02cdffa323f" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.575198 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"996523ccd2c4efa0dbe886319e68d00160c9c0e388e002e5509dd02cdffa323f"} err="failed to get container status \"996523ccd2c4efa0dbe886319e68d00160c9c0e388e002e5509dd02cdffa323f\": rpc error: code = NotFound desc = could not find container \"996523ccd2c4efa0dbe886319e68d00160c9c0e388e002e5509dd02cdffa323f\": container with ID starting with 996523ccd2c4efa0dbe886319e68d00160c9c0e388e002e5509dd02cdffa323f not found: ID does not exist" Oct 02 02:36:55 crc kubenswrapper[4775]: I1002 02:36:55.775304 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e81b45fe-739c-4acf-a91d-21d565484ec9" path="/var/lib/kubelet/pods/e81b45fe-739c-4acf-a91d-21d565484ec9/volumes" Oct 02 02:37:07 crc kubenswrapper[4775]: I1002 02:37:07.233489 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:37:07 crc kubenswrapper[4775]: I1002 02:37:07.234013 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:37:37 crc kubenswrapper[4775]: I1002 02:37:37.233171 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:37:37 crc kubenswrapper[4775]: I1002 02:37:37.234144 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:38:07 crc kubenswrapper[4775]: I1002 02:38:07.234018 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:38:07 crc kubenswrapper[4775]: I1002 02:38:07.234652 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:38:07 crc kubenswrapper[4775]: I1002 02:38:07.234712 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 02:38:07 crc kubenswrapper[4775]: I1002 02:38:07.235467 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dbcebe30237c32f26c9ca1f57bfd5e7480a1a0e125a47d39113cbb556f02abc5"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:38:07 crc kubenswrapper[4775]: I1002 02:38:07.235563 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://dbcebe30237c32f26c9ca1f57bfd5e7480a1a0e125a47d39113cbb556f02abc5" gracePeriod=600 Oct 02 02:38:08 crc kubenswrapper[4775]: I1002 02:38:08.125666 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="dbcebe30237c32f26c9ca1f57bfd5e7480a1a0e125a47d39113cbb556f02abc5" exitCode=0 Oct 02 02:38:08 crc kubenswrapper[4775]: I1002 02:38:08.125733 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"dbcebe30237c32f26c9ca1f57bfd5e7480a1a0e125a47d39113cbb556f02abc5"} Oct 02 02:38:08 crc kubenswrapper[4775]: I1002 02:38:08.126465 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95"} Oct 02 02:38:08 crc kubenswrapper[4775]: I1002 02:38:08.126510 4775 scope.go:117] "RemoveContainer" containerID="d73c0711b0fb9227191ecb2b1770c3604d42c2fbaa5ce8c474bce8f77c484a0d" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.618883 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gglgv"] Oct 02 02:38:38 crc kubenswrapper[4775]: E1002 02:38:38.619977 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81b45fe-739c-4acf-a91d-21d565484ec9" containerName="registry-server" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.619999 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81b45fe-739c-4acf-a91d-21d565484ec9" containerName="registry-server" Oct 02 02:38:38 crc kubenswrapper[4775]: E1002 02:38:38.620014 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5e1912d-1a71-4372-b6ef-2e08a53bf302" containerName="registry-server" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.620025 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5e1912d-1a71-4372-b6ef-2e08a53bf302" containerName="registry-server" Oct 02 02:38:38 crc kubenswrapper[4775]: E1002 02:38:38.620051 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81b45fe-739c-4acf-a91d-21d565484ec9" containerName="extract-utilities" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.620063 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81b45fe-739c-4acf-a91d-21d565484ec9" containerName="extract-utilities" Oct 02 02:38:38 crc kubenswrapper[4775]: E1002 02:38:38.620079 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5e1912d-1a71-4372-b6ef-2e08a53bf302" containerName="extract-content" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.620090 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5e1912d-1a71-4372-b6ef-2e08a53bf302" containerName="extract-content" Oct 02 02:38:38 crc kubenswrapper[4775]: E1002 02:38:38.620110 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5e1912d-1a71-4372-b6ef-2e08a53bf302" containerName="extract-utilities" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.620121 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5e1912d-1a71-4372-b6ef-2e08a53bf302" containerName="extract-utilities" Oct 02 02:38:38 crc kubenswrapper[4775]: E1002 02:38:38.620155 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81b45fe-739c-4acf-a91d-21d565484ec9" containerName="extract-content" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.620167 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81b45fe-739c-4acf-a91d-21d565484ec9" containerName="extract-content" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.620389 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e81b45fe-739c-4acf-a91d-21d565484ec9" containerName="registry-server" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.620416 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5e1912d-1a71-4372-b6ef-2e08a53bf302" containerName="registry-server" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.621943 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.632050 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gglgv"] Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.748730 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tkg2\" (UniqueName: \"kubernetes.io/projected/689d8065-f5a3-4e3d-a786-45837c12e20f-kube-api-access-7tkg2\") pod \"redhat-marketplace-gglgv\" (UID: \"689d8065-f5a3-4e3d-a786-45837c12e20f\") " pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.749063 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/689d8065-f5a3-4e3d-a786-45837c12e20f-utilities\") pod \"redhat-marketplace-gglgv\" (UID: \"689d8065-f5a3-4e3d-a786-45837c12e20f\") " pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.749142 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/689d8065-f5a3-4e3d-a786-45837c12e20f-catalog-content\") pod \"redhat-marketplace-gglgv\" (UID: \"689d8065-f5a3-4e3d-a786-45837c12e20f\") " pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.850583 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/689d8065-f5a3-4e3d-a786-45837c12e20f-catalog-content\") pod \"redhat-marketplace-gglgv\" (UID: \"689d8065-f5a3-4e3d-a786-45837c12e20f\") " pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.850751 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/689d8065-f5a3-4e3d-a786-45837c12e20f-utilities\") pod \"redhat-marketplace-gglgv\" (UID: \"689d8065-f5a3-4e3d-a786-45837c12e20f\") " pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.850777 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tkg2\" (UniqueName: \"kubernetes.io/projected/689d8065-f5a3-4e3d-a786-45837c12e20f-kube-api-access-7tkg2\") pod \"redhat-marketplace-gglgv\" (UID: \"689d8065-f5a3-4e3d-a786-45837c12e20f\") " pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.851168 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/689d8065-f5a3-4e3d-a786-45837c12e20f-catalog-content\") pod \"redhat-marketplace-gglgv\" (UID: \"689d8065-f5a3-4e3d-a786-45837c12e20f\") " pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.851505 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/689d8065-f5a3-4e3d-a786-45837c12e20f-utilities\") pod \"redhat-marketplace-gglgv\" (UID: \"689d8065-f5a3-4e3d-a786-45837c12e20f\") " pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:38 crc kubenswrapper[4775]: I1002 02:38:38.874750 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tkg2\" (UniqueName: \"kubernetes.io/projected/689d8065-f5a3-4e3d-a786-45837c12e20f-kube-api-access-7tkg2\") pod \"redhat-marketplace-gglgv\" (UID: \"689d8065-f5a3-4e3d-a786-45837c12e20f\") " pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:39 crc kubenswrapper[4775]: I1002 02:38:39.003723 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:39 crc kubenswrapper[4775]: I1002 02:38:39.486009 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gglgv"] Oct 02 02:38:40 crc kubenswrapper[4775]: I1002 02:38:40.451664 4775 generic.go:334] "Generic (PLEG): container finished" podID="689d8065-f5a3-4e3d-a786-45837c12e20f" containerID="1b7890eccfbc88863ea5bdc9d4e0321a25a4a5397e9f751ee9fad350ab66f357" exitCode=0 Oct 02 02:38:40 crc kubenswrapper[4775]: I1002 02:38:40.451722 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gglgv" event={"ID":"689d8065-f5a3-4e3d-a786-45837c12e20f","Type":"ContainerDied","Data":"1b7890eccfbc88863ea5bdc9d4e0321a25a4a5397e9f751ee9fad350ab66f357"} Oct 02 02:38:40 crc kubenswrapper[4775]: I1002 02:38:40.451755 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gglgv" event={"ID":"689d8065-f5a3-4e3d-a786-45837c12e20f","Type":"ContainerStarted","Data":"24a7c4708bf19c39e4407eab7cf90178ac35770d558bc4a384216a9e7f3a51a4"} Oct 02 02:38:42 crc kubenswrapper[4775]: I1002 02:38:42.472607 4775 generic.go:334] "Generic (PLEG): container finished" podID="689d8065-f5a3-4e3d-a786-45837c12e20f" containerID="f8c8d0e97055f4347d51c5ae4872c2f0ed0b3841adbc259a8ef364c53062384c" exitCode=0 Oct 02 02:38:42 crc kubenswrapper[4775]: I1002 02:38:42.475373 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gglgv" event={"ID":"689d8065-f5a3-4e3d-a786-45837c12e20f","Type":"ContainerDied","Data":"f8c8d0e97055f4347d51c5ae4872c2f0ed0b3841adbc259a8ef364c53062384c"} Oct 02 02:38:43 crc kubenswrapper[4775]: I1002 02:38:43.487983 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gglgv" event={"ID":"689d8065-f5a3-4e3d-a786-45837c12e20f","Type":"ContainerStarted","Data":"356301d0493da34100f94fafaf3f4d95323acc53e62149c7ef26304e1a31c517"} Oct 02 02:38:43 crc kubenswrapper[4775]: I1002 02:38:43.517521 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gglgv" podStartSLOduration=3.072404248 podStartE2EDuration="5.517484524s" podCreationTimestamp="2025-10-02 02:38:38 +0000 UTC" firstStartedPulling="2025-10-02 02:38:40.454471271 +0000 UTC m=+3457.621215351" lastFinishedPulling="2025-10-02 02:38:42.899551557 +0000 UTC m=+3460.066295627" observedRunningTime="2025-10-02 02:38:43.51228392 +0000 UTC m=+3460.679027960" watchObservedRunningTime="2025-10-02 02:38:43.517484524 +0000 UTC m=+3460.684228624" Oct 02 02:38:49 crc kubenswrapper[4775]: I1002 02:38:49.004227 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:49 crc kubenswrapper[4775]: I1002 02:38:49.004639 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:49 crc kubenswrapper[4775]: I1002 02:38:49.075101 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:49 crc kubenswrapper[4775]: I1002 02:38:49.596558 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:49 crc kubenswrapper[4775]: I1002 02:38:49.674441 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gglgv"] Oct 02 02:38:51 crc kubenswrapper[4775]: I1002 02:38:51.557073 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gglgv" podUID="689d8065-f5a3-4e3d-a786-45837c12e20f" containerName="registry-server" containerID="cri-o://356301d0493da34100f94fafaf3f4d95323acc53e62149c7ef26304e1a31c517" gracePeriod=2 Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.007499 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.061737 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/689d8065-f5a3-4e3d-a786-45837c12e20f-catalog-content\") pod \"689d8065-f5a3-4e3d-a786-45837c12e20f\" (UID: \"689d8065-f5a3-4e3d-a786-45837c12e20f\") " Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.061861 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tkg2\" (UniqueName: \"kubernetes.io/projected/689d8065-f5a3-4e3d-a786-45837c12e20f-kube-api-access-7tkg2\") pod \"689d8065-f5a3-4e3d-a786-45837c12e20f\" (UID: \"689d8065-f5a3-4e3d-a786-45837c12e20f\") " Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.062034 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/689d8065-f5a3-4e3d-a786-45837c12e20f-utilities\") pod \"689d8065-f5a3-4e3d-a786-45837c12e20f\" (UID: \"689d8065-f5a3-4e3d-a786-45837c12e20f\") " Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.063184 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/689d8065-f5a3-4e3d-a786-45837c12e20f-utilities" (OuterVolumeSpecName: "utilities") pod "689d8065-f5a3-4e3d-a786-45837c12e20f" (UID: "689d8065-f5a3-4e3d-a786-45837c12e20f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.071071 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/689d8065-f5a3-4e3d-a786-45837c12e20f-kube-api-access-7tkg2" (OuterVolumeSpecName: "kube-api-access-7tkg2") pod "689d8065-f5a3-4e3d-a786-45837c12e20f" (UID: "689d8065-f5a3-4e3d-a786-45837c12e20f"). InnerVolumeSpecName "kube-api-access-7tkg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.087300 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/689d8065-f5a3-4e3d-a786-45837c12e20f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "689d8065-f5a3-4e3d-a786-45837c12e20f" (UID: "689d8065-f5a3-4e3d-a786-45837c12e20f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.164323 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/689d8065-f5a3-4e3d-a786-45837c12e20f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.164358 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/689d8065-f5a3-4e3d-a786-45837c12e20f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.164374 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tkg2\" (UniqueName: \"kubernetes.io/projected/689d8065-f5a3-4e3d-a786-45837c12e20f-kube-api-access-7tkg2\") on node \"crc\" DevicePath \"\"" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.569174 4775 generic.go:334] "Generic (PLEG): container finished" podID="689d8065-f5a3-4e3d-a786-45837c12e20f" containerID="356301d0493da34100f94fafaf3f4d95323acc53e62149c7ef26304e1a31c517" exitCode=0 Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.569252 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gglgv" event={"ID":"689d8065-f5a3-4e3d-a786-45837c12e20f","Type":"ContainerDied","Data":"356301d0493da34100f94fafaf3f4d95323acc53e62149c7ef26304e1a31c517"} Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.569295 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gglgv" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.569314 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gglgv" event={"ID":"689d8065-f5a3-4e3d-a786-45837c12e20f","Type":"ContainerDied","Data":"24a7c4708bf19c39e4407eab7cf90178ac35770d558bc4a384216a9e7f3a51a4"} Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.569354 4775 scope.go:117] "RemoveContainer" containerID="356301d0493da34100f94fafaf3f4d95323acc53e62149c7ef26304e1a31c517" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.605925 4775 scope.go:117] "RemoveContainer" containerID="f8c8d0e97055f4347d51c5ae4872c2f0ed0b3841adbc259a8ef364c53062384c" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.630135 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gglgv"] Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.631621 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gglgv"] Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.643805 4775 scope.go:117] "RemoveContainer" containerID="1b7890eccfbc88863ea5bdc9d4e0321a25a4a5397e9f751ee9fad350ab66f357" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.693779 4775 scope.go:117] "RemoveContainer" containerID="356301d0493da34100f94fafaf3f4d95323acc53e62149c7ef26304e1a31c517" Oct 02 02:38:52 crc kubenswrapper[4775]: E1002 02:38:52.694460 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"356301d0493da34100f94fafaf3f4d95323acc53e62149c7ef26304e1a31c517\": container with ID starting with 356301d0493da34100f94fafaf3f4d95323acc53e62149c7ef26304e1a31c517 not found: ID does not exist" containerID="356301d0493da34100f94fafaf3f4d95323acc53e62149c7ef26304e1a31c517" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.694502 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"356301d0493da34100f94fafaf3f4d95323acc53e62149c7ef26304e1a31c517"} err="failed to get container status \"356301d0493da34100f94fafaf3f4d95323acc53e62149c7ef26304e1a31c517\": rpc error: code = NotFound desc = could not find container \"356301d0493da34100f94fafaf3f4d95323acc53e62149c7ef26304e1a31c517\": container with ID starting with 356301d0493da34100f94fafaf3f4d95323acc53e62149c7ef26304e1a31c517 not found: ID does not exist" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.694527 4775 scope.go:117] "RemoveContainer" containerID="f8c8d0e97055f4347d51c5ae4872c2f0ed0b3841adbc259a8ef364c53062384c" Oct 02 02:38:52 crc kubenswrapper[4775]: E1002 02:38:52.694935 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8c8d0e97055f4347d51c5ae4872c2f0ed0b3841adbc259a8ef364c53062384c\": container with ID starting with f8c8d0e97055f4347d51c5ae4872c2f0ed0b3841adbc259a8ef364c53062384c not found: ID does not exist" containerID="f8c8d0e97055f4347d51c5ae4872c2f0ed0b3841adbc259a8ef364c53062384c" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.695011 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8c8d0e97055f4347d51c5ae4872c2f0ed0b3841adbc259a8ef364c53062384c"} err="failed to get container status \"f8c8d0e97055f4347d51c5ae4872c2f0ed0b3841adbc259a8ef364c53062384c\": rpc error: code = NotFound desc = could not find container \"f8c8d0e97055f4347d51c5ae4872c2f0ed0b3841adbc259a8ef364c53062384c\": container with ID starting with f8c8d0e97055f4347d51c5ae4872c2f0ed0b3841adbc259a8ef364c53062384c not found: ID does not exist" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.695060 4775 scope.go:117] "RemoveContainer" containerID="1b7890eccfbc88863ea5bdc9d4e0321a25a4a5397e9f751ee9fad350ab66f357" Oct 02 02:38:52 crc kubenswrapper[4775]: E1002 02:38:52.695549 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b7890eccfbc88863ea5bdc9d4e0321a25a4a5397e9f751ee9fad350ab66f357\": container with ID starting with 1b7890eccfbc88863ea5bdc9d4e0321a25a4a5397e9f751ee9fad350ab66f357 not found: ID does not exist" containerID="1b7890eccfbc88863ea5bdc9d4e0321a25a4a5397e9f751ee9fad350ab66f357" Oct 02 02:38:52 crc kubenswrapper[4775]: I1002 02:38:52.695630 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b7890eccfbc88863ea5bdc9d4e0321a25a4a5397e9f751ee9fad350ab66f357"} err="failed to get container status \"1b7890eccfbc88863ea5bdc9d4e0321a25a4a5397e9f751ee9fad350ab66f357\": rpc error: code = NotFound desc = could not find container \"1b7890eccfbc88863ea5bdc9d4e0321a25a4a5397e9f751ee9fad350ab66f357\": container with ID starting with 1b7890eccfbc88863ea5bdc9d4e0321a25a4a5397e9f751ee9fad350ab66f357 not found: ID does not exist" Oct 02 02:38:53 crc kubenswrapper[4775]: I1002 02:38:53.783210 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="689d8065-f5a3-4e3d-a786-45837c12e20f" path="/var/lib/kubelet/pods/689d8065-f5a3-4e3d-a786-45837c12e20f/volumes" Oct 02 02:39:50 crc kubenswrapper[4775]: I1002 02:39:50.830265 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bvt4g"] Oct 02 02:39:50 crc kubenswrapper[4775]: E1002 02:39:50.831633 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689d8065-f5a3-4e3d-a786-45837c12e20f" containerName="registry-server" Oct 02 02:39:50 crc kubenswrapper[4775]: I1002 02:39:50.831663 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="689d8065-f5a3-4e3d-a786-45837c12e20f" containerName="registry-server" Oct 02 02:39:50 crc kubenswrapper[4775]: E1002 02:39:50.831724 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689d8065-f5a3-4e3d-a786-45837c12e20f" containerName="extract-content" Oct 02 02:39:50 crc kubenswrapper[4775]: I1002 02:39:50.831742 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="689d8065-f5a3-4e3d-a786-45837c12e20f" containerName="extract-content" Oct 02 02:39:50 crc kubenswrapper[4775]: E1002 02:39:50.831773 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="689d8065-f5a3-4e3d-a786-45837c12e20f" containerName="extract-utilities" Oct 02 02:39:50 crc kubenswrapper[4775]: I1002 02:39:50.831791 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="689d8065-f5a3-4e3d-a786-45837c12e20f" containerName="extract-utilities" Oct 02 02:39:50 crc kubenswrapper[4775]: I1002 02:39:50.834220 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="689d8065-f5a3-4e3d-a786-45837c12e20f" containerName="registry-server" Oct 02 02:39:50 crc kubenswrapper[4775]: I1002 02:39:50.836225 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:39:50 crc kubenswrapper[4775]: I1002 02:39:50.851711 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bvt4g"] Oct 02 02:39:50 crc kubenswrapper[4775]: I1002 02:39:50.960754 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-catalog-content\") pod \"certified-operators-bvt4g\" (UID: \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\") " pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:39:50 crc kubenswrapper[4775]: I1002 02:39:50.960831 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-utilities\") pod \"certified-operators-bvt4g\" (UID: \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\") " pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:39:50 crc kubenswrapper[4775]: I1002 02:39:50.960861 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxgn6\" (UniqueName: \"kubernetes.io/projected/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-kube-api-access-wxgn6\") pod \"certified-operators-bvt4g\" (UID: \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\") " pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:39:51 crc kubenswrapper[4775]: I1002 02:39:51.062801 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-utilities\") pod \"certified-operators-bvt4g\" (UID: \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\") " pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:39:51 crc kubenswrapper[4775]: I1002 02:39:51.062866 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxgn6\" (UniqueName: \"kubernetes.io/projected/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-kube-api-access-wxgn6\") pod \"certified-operators-bvt4g\" (UID: \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\") " pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:39:51 crc kubenswrapper[4775]: I1002 02:39:51.063001 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-catalog-content\") pod \"certified-operators-bvt4g\" (UID: \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\") " pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:39:51 crc kubenswrapper[4775]: I1002 02:39:51.063319 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-utilities\") pod \"certified-operators-bvt4g\" (UID: \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\") " pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:39:51 crc kubenswrapper[4775]: I1002 02:39:51.063680 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-catalog-content\") pod \"certified-operators-bvt4g\" (UID: \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\") " pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:39:51 crc kubenswrapper[4775]: I1002 02:39:51.087573 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxgn6\" (UniqueName: \"kubernetes.io/projected/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-kube-api-access-wxgn6\") pod \"certified-operators-bvt4g\" (UID: \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\") " pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:39:51 crc kubenswrapper[4775]: I1002 02:39:51.165936 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:39:51 crc kubenswrapper[4775]: I1002 02:39:51.614629 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bvt4g"] Oct 02 02:39:52 crc kubenswrapper[4775]: I1002 02:39:52.193502 4775 generic.go:334] "Generic (PLEG): container finished" podID="cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" containerID="b13f8c38067c388bcb0fa4091c5eb6a52a38b90bc25ca5eb159dc93e2ba2ea80" exitCode=0 Oct 02 02:39:52 crc kubenswrapper[4775]: I1002 02:39:52.193608 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bvt4g" event={"ID":"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a","Type":"ContainerDied","Data":"b13f8c38067c388bcb0fa4091c5eb6a52a38b90bc25ca5eb159dc93e2ba2ea80"} Oct 02 02:39:52 crc kubenswrapper[4775]: I1002 02:39:52.193880 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bvt4g" event={"ID":"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a","Type":"ContainerStarted","Data":"e781f332d79c4f7740cfc149e1def442df31eddec594cf3d654264a6c40a7e05"} Oct 02 02:39:53 crc kubenswrapper[4775]: I1002 02:39:53.202693 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bvt4g" event={"ID":"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a","Type":"ContainerStarted","Data":"6d547a8db8dee75662f5d4cc401ba34e49eb08b88cc2313d99bc4180659b0624"} Oct 02 02:39:54 crc kubenswrapper[4775]: I1002 02:39:54.216050 4775 generic.go:334] "Generic (PLEG): container finished" podID="cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" containerID="6d547a8db8dee75662f5d4cc401ba34e49eb08b88cc2313d99bc4180659b0624" exitCode=0 Oct 02 02:39:54 crc kubenswrapper[4775]: I1002 02:39:54.217119 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bvt4g" event={"ID":"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a","Type":"ContainerDied","Data":"6d547a8db8dee75662f5d4cc401ba34e49eb08b88cc2313d99bc4180659b0624"} Oct 02 02:39:55 crc kubenswrapper[4775]: I1002 02:39:55.229212 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bvt4g" event={"ID":"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a","Type":"ContainerStarted","Data":"49956f577ea366c3d05acdc0f537db078b750fb40de8cfcd0951eeb4fe47c246"} Oct 02 02:39:55 crc kubenswrapper[4775]: I1002 02:39:55.257071 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bvt4g" podStartSLOduration=2.674940784 podStartE2EDuration="5.257056381s" podCreationTimestamp="2025-10-02 02:39:50 +0000 UTC" firstStartedPulling="2025-10-02 02:39:52.195208438 +0000 UTC m=+3529.361952518" lastFinishedPulling="2025-10-02 02:39:54.777324045 +0000 UTC m=+3531.944068115" observedRunningTime="2025-10-02 02:39:55.254185337 +0000 UTC m=+3532.420929377" watchObservedRunningTime="2025-10-02 02:39:55.257056381 +0000 UTC m=+3532.423800421" Oct 02 02:40:01 crc kubenswrapper[4775]: I1002 02:40:01.167019 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:40:01 crc kubenswrapper[4775]: I1002 02:40:01.167288 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:40:01 crc kubenswrapper[4775]: I1002 02:40:01.251127 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:40:01 crc kubenswrapper[4775]: I1002 02:40:01.343734 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:40:01 crc kubenswrapper[4775]: I1002 02:40:01.503995 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bvt4g"] Oct 02 02:40:03 crc kubenswrapper[4775]: I1002 02:40:03.310031 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bvt4g" podUID="cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" containerName="registry-server" containerID="cri-o://49956f577ea366c3d05acdc0f537db078b750fb40de8cfcd0951eeb4fe47c246" gracePeriod=2 Oct 02 02:40:03 crc kubenswrapper[4775]: I1002 02:40:03.749608 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:40:03 crc kubenswrapper[4775]: I1002 02:40:03.862087 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-utilities\") pod \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\" (UID: \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\") " Oct 02 02:40:03 crc kubenswrapper[4775]: I1002 02:40:03.862133 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxgn6\" (UniqueName: \"kubernetes.io/projected/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-kube-api-access-wxgn6\") pod \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\" (UID: \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\") " Oct 02 02:40:03 crc kubenswrapper[4775]: I1002 02:40:03.862246 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-catalog-content\") pod \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\" (UID: \"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a\") " Oct 02 02:40:03 crc kubenswrapper[4775]: I1002 02:40:03.863344 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-utilities" (OuterVolumeSpecName: "utilities") pod "cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" (UID: "cde78e6e-adf0-47e8-b7f0-d1ded6cd880a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:40:03 crc kubenswrapper[4775]: I1002 02:40:03.879134 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-kube-api-access-wxgn6" (OuterVolumeSpecName: "kube-api-access-wxgn6") pod "cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" (UID: "cde78e6e-adf0-47e8-b7f0-d1ded6cd880a"). InnerVolumeSpecName "kube-api-access-wxgn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:40:03 crc kubenswrapper[4775]: I1002 02:40:03.964319 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:40:03 crc kubenswrapper[4775]: I1002 02:40:03.964360 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxgn6\" (UniqueName: \"kubernetes.io/projected/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-kube-api-access-wxgn6\") on node \"crc\" DevicePath \"\"" Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.319487 4775 generic.go:334] "Generic (PLEG): container finished" podID="cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" containerID="49956f577ea366c3d05acdc0f537db078b750fb40de8cfcd0951eeb4fe47c246" exitCode=0 Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.319533 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bvt4g" event={"ID":"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a","Type":"ContainerDied","Data":"49956f577ea366c3d05acdc0f537db078b750fb40de8cfcd0951eeb4fe47c246"} Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.319566 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bvt4g" event={"ID":"cde78e6e-adf0-47e8-b7f0-d1ded6cd880a","Type":"ContainerDied","Data":"e781f332d79c4f7740cfc149e1def442df31eddec594cf3d654264a6c40a7e05"} Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.319666 4775 scope.go:117] "RemoveContainer" containerID="49956f577ea366c3d05acdc0f537db078b750fb40de8cfcd0951eeb4fe47c246" Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.321182 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bvt4g" Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.346339 4775 scope.go:117] "RemoveContainer" containerID="6d547a8db8dee75662f5d4cc401ba34e49eb08b88cc2313d99bc4180659b0624" Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.366507 4775 scope.go:117] "RemoveContainer" containerID="b13f8c38067c388bcb0fa4091c5eb6a52a38b90bc25ca5eb159dc93e2ba2ea80" Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.396223 4775 scope.go:117] "RemoveContainer" containerID="49956f577ea366c3d05acdc0f537db078b750fb40de8cfcd0951eeb4fe47c246" Oct 02 02:40:04 crc kubenswrapper[4775]: E1002 02:40:04.396710 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49956f577ea366c3d05acdc0f537db078b750fb40de8cfcd0951eeb4fe47c246\": container with ID starting with 49956f577ea366c3d05acdc0f537db078b750fb40de8cfcd0951eeb4fe47c246 not found: ID does not exist" containerID="49956f577ea366c3d05acdc0f537db078b750fb40de8cfcd0951eeb4fe47c246" Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.396750 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49956f577ea366c3d05acdc0f537db078b750fb40de8cfcd0951eeb4fe47c246"} err="failed to get container status \"49956f577ea366c3d05acdc0f537db078b750fb40de8cfcd0951eeb4fe47c246\": rpc error: code = NotFound desc = could not find container \"49956f577ea366c3d05acdc0f537db078b750fb40de8cfcd0951eeb4fe47c246\": container with ID starting with 49956f577ea366c3d05acdc0f537db078b750fb40de8cfcd0951eeb4fe47c246 not found: ID does not exist" Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.396776 4775 scope.go:117] "RemoveContainer" containerID="6d547a8db8dee75662f5d4cc401ba34e49eb08b88cc2313d99bc4180659b0624" Oct 02 02:40:04 crc kubenswrapper[4775]: E1002 02:40:04.397084 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d547a8db8dee75662f5d4cc401ba34e49eb08b88cc2313d99bc4180659b0624\": container with ID starting with 6d547a8db8dee75662f5d4cc401ba34e49eb08b88cc2313d99bc4180659b0624 not found: ID does not exist" containerID="6d547a8db8dee75662f5d4cc401ba34e49eb08b88cc2313d99bc4180659b0624" Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.397220 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d547a8db8dee75662f5d4cc401ba34e49eb08b88cc2313d99bc4180659b0624"} err="failed to get container status \"6d547a8db8dee75662f5d4cc401ba34e49eb08b88cc2313d99bc4180659b0624\": rpc error: code = NotFound desc = could not find container \"6d547a8db8dee75662f5d4cc401ba34e49eb08b88cc2313d99bc4180659b0624\": container with ID starting with 6d547a8db8dee75662f5d4cc401ba34e49eb08b88cc2313d99bc4180659b0624 not found: ID does not exist" Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.397329 4775 scope.go:117] "RemoveContainer" containerID="b13f8c38067c388bcb0fa4091c5eb6a52a38b90bc25ca5eb159dc93e2ba2ea80" Oct 02 02:40:04 crc kubenswrapper[4775]: E1002 02:40:04.397657 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b13f8c38067c388bcb0fa4091c5eb6a52a38b90bc25ca5eb159dc93e2ba2ea80\": container with ID starting with b13f8c38067c388bcb0fa4091c5eb6a52a38b90bc25ca5eb159dc93e2ba2ea80 not found: ID does not exist" containerID="b13f8c38067c388bcb0fa4091c5eb6a52a38b90bc25ca5eb159dc93e2ba2ea80" Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.397680 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b13f8c38067c388bcb0fa4091c5eb6a52a38b90bc25ca5eb159dc93e2ba2ea80"} err="failed to get container status \"b13f8c38067c388bcb0fa4091c5eb6a52a38b90bc25ca5eb159dc93e2ba2ea80\": rpc error: code = NotFound desc = could not find container \"b13f8c38067c388bcb0fa4091c5eb6a52a38b90bc25ca5eb159dc93e2ba2ea80\": container with ID starting with b13f8c38067c388bcb0fa4091c5eb6a52a38b90bc25ca5eb159dc93e2ba2ea80 not found: ID does not exist" Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.434581 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" (UID: "cde78e6e-adf0-47e8-b7f0-d1ded6cd880a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.471612 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.676560 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bvt4g"] Oct 02 02:40:04 crc kubenswrapper[4775]: I1002 02:40:04.688989 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bvt4g"] Oct 02 02:40:05 crc kubenswrapper[4775]: I1002 02:40:05.778986 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" path="/var/lib/kubelet/pods/cde78e6e-adf0-47e8-b7f0-d1ded6cd880a/volumes" Oct 02 02:40:07 crc kubenswrapper[4775]: I1002 02:40:07.234262 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:40:07 crc kubenswrapper[4775]: I1002 02:40:07.234639 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:40:37 crc kubenswrapper[4775]: I1002 02:40:37.233788 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:40:37 crc kubenswrapper[4775]: I1002 02:40:37.234664 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:41:07 crc kubenswrapper[4775]: I1002 02:41:07.234089 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:41:07 crc kubenswrapper[4775]: I1002 02:41:07.234801 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:41:07 crc kubenswrapper[4775]: I1002 02:41:07.234867 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 02:41:07 crc kubenswrapper[4775]: I1002 02:41:07.235739 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:41:07 crc kubenswrapper[4775]: I1002 02:41:07.235833 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" gracePeriod=600 Oct 02 02:41:07 crc kubenswrapper[4775]: E1002 02:41:07.372804 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:41:07 crc kubenswrapper[4775]: I1002 02:41:07.935726 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" exitCode=0 Oct 02 02:41:07 crc kubenswrapper[4775]: I1002 02:41:07.935819 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95"} Oct 02 02:41:07 crc kubenswrapper[4775]: I1002 02:41:07.935908 4775 scope.go:117] "RemoveContainer" containerID="dbcebe30237c32f26c9ca1f57bfd5e7480a1a0e125a47d39113cbb556f02abc5" Oct 02 02:41:07 crc kubenswrapper[4775]: I1002 02:41:07.937039 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:41:07 crc kubenswrapper[4775]: E1002 02:41:07.937710 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:41:22 crc kubenswrapper[4775]: I1002 02:41:22.765748 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:41:22 crc kubenswrapper[4775]: E1002 02:41:22.766725 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:41:37 crc kubenswrapper[4775]: I1002 02:41:37.765077 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:41:37 crc kubenswrapper[4775]: E1002 02:41:37.766112 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:41:49 crc kubenswrapper[4775]: I1002 02:41:49.766328 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:41:49 crc kubenswrapper[4775]: E1002 02:41:49.767723 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:42:00 crc kubenswrapper[4775]: I1002 02:42:00.765121 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:42:00 crc kubenswrapper[4775]: E1002 02:42:00.766145 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:42:14 crc kubenswrapper[4775]: I1002 02:42:14.765677 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:42:14 crc kubenswrapper[4775]: E1002 02:42:14.766542 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:42:25 crc kubenswrapper[4775]: I1002 02:42:25.766354 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:42:25 crc kubenswrapper[4775]: E1002 02:42:25.770222 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:42:38 crc kubenswrapper[4775]: I1002 02:42:38.765202 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:42:38 crc kubenswrapper[4775]: E1002 02:42:38.765879 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:42:49 crc kubenswrapper[4775]: I1002 02:42:49.766134 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:42:49 crc kubenswrapper[4775]: E1002 02:42:49.768588 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:43:02 crc kubenswrapper[4775]: I1002 02:43:02.766007 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:43:02 crc kubenswrapper[4775]: E1002 02:43:02.768322 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:43:14 crc kubenswrapper[4775]: I1002 02:43:14.765127 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:43:14 crc kubenswrapper[4775]: E1002 02:43:14.766075 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:43:28 crc kubenswrapper[4775]: I1002 02:43:28.765196 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:43:28 crc kubenswrapper[4775]: E1002 02:43:28.766324 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:43:43 crc kubenswrapper[4775]: I1002 02:43:43.773453 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:43:43 crc kubenswrapper[4775]: E1002 02:43:43.774365 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:43:55 crc kubenswrapper[4775]: I1002 02:43:55.765213 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:43:55 crc kubenswrapper[4775]: E1002 02:43:55.766281 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:44:06 crc kubenswrapper[4775]: I1002 02:44:06.765993 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:44:06 crc kubenswrapper[4775]: E1002 02:44:06.767066 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:44:19 crc kubenswrapper[4775]: I1002 02:44:19.766126 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:44:19 crc kubenswrapper[4775]: E1002 02:44:19.767213 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:44:31 crc kubenswrapper[4775]: I1002 02:44:31.766101 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:44:31 crc kubenswrapper[4775]: E1002 02:44:31.767031 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:44:46 crc kubenswrapper[4775]: I1002 02:44:46.765892 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:44:46 crc kubenswrapper[4775]: E1002 02:44:46.767085 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:44:58 crc kubenswrapper[4775]: I1002 02:44:58.766314 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:44:58 crc kubenswrapper[4775]: E1002 02:44:58.766862 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.156675 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl"] Oct 02 02:45:00 crc kubenswrapper[4775]: E1002 02:45:00.156989 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" containerName="extract-utilities" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.157004 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" containerName="extract-utilities" Oct 02 02:45:00 crc kubenswrapper[4775]: E1002 02:45:00.157017 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" containerName="registry-server" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.157023 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" containerName="registry-server" Oct 02 02:45:00 crc kubenswrapper[4775]: E1002 02:45:00.157035 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" containerName="extract-content" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.157043 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" containerName="extract-content" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.157185 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="cde78e6e-adf0-47e8-b7f0-d1ded6cd880a" containerName="registry-server" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.157936 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.162585 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.163879 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.177598 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl"] Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.207369 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db403783-c3a4-4423-8cc9-16d1b49c2dd3-secret-volume\") pod \"collect-profiles-29322885-mptjl\" (UID: \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.207451 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t552h\" (UniqueName: \"kubernetes.io/projected/db403783-c3a4-4423-8cc9-16d1b49c2dd3-kube-api-access-t552h\") pod \"collect-profiles-29322885-mptjl\" (UID: \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.207485 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db403783-c3a4-4423-8cc9-16d1b49c2dd3-config-volume\") pod \"collect-profiles-29322885-mptjl\" (UID: \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.308262 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t552h\" (UniqueName: \"kubernetes.io/projected/db403783-c3a4-4423-8cc9-16d1b49c2dd3-kube-api-access-t552h\") pod \"collect-profiles-29322885-mptjl\" (UID: \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.308338 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db403783-c3a4-4423-8cc9-16d1b49c2dd3-config-volume\") pod \"collect-profiles-29322885-mptjl\" (UID: \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.308494 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db403783-c3a4-4423-8cc9-16d1b49c2dd3-secret-volume\") pod \"collect-profiles-29322885-mptjl\" (UID: \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.310704 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db403783-c3a4-4423-8cc9-16d1b49c2dd3-config-volume\") pod \"collect-profiles-29322885-mptjl\" (UID: \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.318574 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db403783-c3a4-4423-8cc9-16d1b49c2dd3-secret-volume\") pod \"collect-profiles-29322885-mptjl\" (UID: \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.334653 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t552h\" (UniqueName: \"kubernetes.io/projected/db403783-c3a4-4423-8cc9-16d1b49c2dd3-kube-api-access-t552h\") pod \"collect-profiles-29322885-mptjl\" (UID: \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.488177 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" Oct 02 02:45:00 crc kubenswrapper[4775]: I1002 02:45:00.768625 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl"] Oct 02 02:45:01 crc kubenswrapper[4775]: I1002 02:45:01.102175 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" event={"ID":"db403783-c3a4-4423-8cc9-16d1b49c2dd3","Type":"ContainerStarted","Data":"40b8b46191a6a080a0e34cb85e82b0ecebd35223ea9e8b197cd83e77f16af080"} Oct 02 02:45:01 crc kubenswrapper[4775]: I1002 02:45:01.102670 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" event={"ID":"db403783-c3a4-4423-8cc9-16d1b49c2dd3","Type":"ContainerStarted","Data":"a8767febd94b7f11a827ab705c4d7d73c18efd6a5c92ccbb36eb8efcbfba506d"} Oct 02 02:45:02 crc kubenswrapper[4775]: I1002 02:45:02.110831 4775 generic.go:334] "Generic (PLEG): container finished" podID="db403783-c3a4-4423-8cc9-16d1b49c2dd3" containerID="40b8b46191a6a080a0e34cb85e82b0ecebd35223ea9e8b197cd83e77f16af080" exitCode=0 Oct 02 02:45:02 crc kubenswrapper[4775]: I1002 02:45:02.110885 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" event={"ID":"db403783-c3a4-4423-8cc9-16d1b49c2dd3","Type":"ContainerDied","Data":"40b8b46191a6a080a0e34cb85e82b0ecebd35223ea9e8b197cd83e77f16af080"} Oct 02 02:45:03 crc kubenswrapper[4775]: I1002 02:45:03.495171 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" Oct 02 02:45:03 crc kubenswrapper[4775]: I1002 02:45:03.654923 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db403783-c3a4-4423-8cc9-16d1b49c2dd3-secret-volume\") pod \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\" (UID: \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\") " Oct 02 02:45:03 crc kubenswrapper[4775]: I1002 02:45:03.655162 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db403783-c3a4-4423-8cc9-16d1b49c2dd3-config-volume\") pod \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\" (UID: \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\") " Oct 02 02:45:03 crc kubenswrapper[4775]: I1002 02:45:03.655231 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t552h\" (UniqueName: \"kubernetes.io/projected/db403783-c3a4-4423-8cc9-16d1b49c2dd3-kube-api-access-t552h\") pod \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\" (UID: \"db403783-c3a4-4423-8cc9-16d1b49c2dd3\") " Oct 02 02:45:03 crc kubenswrapper[4775]: I1002 02:45:03.656107 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db403783-c3a4-4423-8cc9-16d1b49c2dd3-config-volume" (OuterVolumeSpecName: "config-volume") pod "db403783-c3a4-4423-8cc9-16d1b49c2dd3" (UID: "db403783-c3a4-4423-8cc9-16d1b49c2dd3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:45:03 crc kubenswrapper[4775]: I1002 02:45:03.656454 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/db403783-c3a4-4423-8cc9-16d1b49c2dd3-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:45:03 crc kubenswrapper[4775]: I1002 02:45:03.664169 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db403783-c3a4-4423-8cc9-16d1b49c2dd3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "db403783-c3a4-4423-8cc9-16d1b49c2dd3" (UID: "db403783-c3a4-4423-8cc9-16d1b49c2dd3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:45:03 crc kubenswrapper[4775]: I1002 02:45:03.665908 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db403783-c3a4-4423-8cc9-16d1b49c2dd3-kube-api-access-t552h" (OuterVolumeSpecName: "kube-api-access-t552h") pod "db403783-c3a4-4423-8cc9-16d1b49c2dd3" (UID: "db403783-c3a4-4423-8cc9-16d1b49c2dd3"). InnerVolumeSpecName "kube-api-access-t552h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:45:03 crc kubenswrapper[4775]: I1002 02:45:03.758213 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t552h\" (UniqueName: \"kubernetes.io/projected/db403783-c3a4-4423-8cc9-16d1b49c2dd3-kube-api-access-t552h\") on node \"crc\" DevicePath \"\"" Oct 02 02:45:03 crc kubenswrapper[4775]: I1002 02:45:03.758263 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/db403783-c3a4-4423-8cc9-16d1b49c2dd3-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 02:45:04 crc kubenswrapper[4775]: I1002 02:45:04.130847 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" event={"ID":"db403783-c3a4-4423-8cc9-16d1b49c2dd3","Type":"ContainerDied","Data":"a8767febd94b7f11a827ab705c4d7d73c18efd6a5c92ccbb36eb8efcbfba506d"} Oct 02 02:45:04 crc kubenswrapper[4775]: I1002 02:45:04.131292 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8767febd94b7f11a827ab705c4d7d73c18efd6a5c92ccbb36eb8efcbfba506d" Oct 02 02:45:04 crc kubenswrapper[4775]: I1002 02:45:04.130996 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl" Oct 02 02:45:04 crc kubenswrapper[4775]: I1002 02:45:04.602559 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7"] Oct 02 02:45:04 crc kubenswrapper[4775]: I1002 02:45:04.609129 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322840-w98q7"] Oct 02 02:45:05 crc kubenswrapper[4775]: I1002 02:45:05.783908 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adbfa780-e6fe-402d-a040-175271a4b1e1" path="/var/lib/kubelet/pods/adbfa780-e6fe-402d-a040-175271a4b1e1/volumes" Oct 02 02:45:09 crc kubenswrapper[4775]: I1002 02:45:09.699370 4775 scope.go:117] "RemoveContainer" containerID="b296aa4012103f2996652ab29adbbb7c34e7f48034623ae89d178928e0a69556" Oct 02 02:45:11 crc kubenswrapper[4775]: I1002 02:45:11.766546 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:45:11 crc kubenswrapper[4775]: E1002 02:45:11.768169 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:45:26 crc kubenswrapper[4775]: I1002 02:45:26.764935 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:45:26 crc kubenswrapper[4775]: E1002 02:45:26.765653 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:45:40 crc kubenswrapper[4775]: I1002 02:45:40.765681 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:45:40 crc kubenswrapper[4775]: E1002 02:45:40.768199 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:45:53 crc kubenswrapper[4775]: I1002 02:45:53.775755 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:45:53 crc kubenswrapper[4775]: E1002 02:45:53.778635 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:46:05 crc kubenswrapper[4775]: I1002 02:46:05.765146 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:46:05 crc kubenswrapper[4775]: E1002 02:46:05.765813 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:46:19 crc kubenswrapper[4775]: I1002 02:46:19.765785 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:46:20 crc kubenswrapper[4775]: I1002 02:46:20.838861 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"701c2b59237bce746d64de4b4003ac93de1140e75c06122e0289272297fc7b80"} Oct 02 02:46:42 crc kubenswrapper[4775]: I1002 02:46:42.723319 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xgn4k"] Oct 02 02:46:42 crc kubenswrapper[4775]: E1002 02:46:42.724660 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db403783-c3a4-4423-8cc9-16d1b49c2dd3" containerName="collect-profiles" Oct 02 02:46:42 crc kubenswrapper[4775]: I1002 02:46:42.724685 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="db403783-c3a4-4423-8cc9-16d1b49c2dd3" containerName="collect-profiles" Oct 02 02:46:42 crc kubenswrapper[4775]: I1002 02:46:42.725025 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="db403783-c3a4-4423-8cc9-16d1b49c2dd3" containerName="collect-profiles" Oct 02 02:46:42 crc kubenswrapper[4775]: I1002 02:46:42.726992 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:42 crc kubenswrapper[4775]: I1002 02:46:42.757365 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xgn4k"] Oct 02 02:46:42 crc kubenswrapper[4775]: I1002 02:46:42.824419 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dfad766-13d7-407a-82fc-92dcc742e8b5-catalog-content\") pod \"redhat-operators-xgn4k\" (UID: \"3dfad766-13d7-407a-82fc-92dcc742e8b5\") " pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:42 crc kubenswrapper[4775]: I1002 02:46:42.824546 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjcmz\" (UniqueName: \"kubernetes.io/projected/3dfad766-13d7-407a-82fc-92dcc742e8b5-kube-api-access-rjcmz\") pod \"redhat-operators-xgn4k\" (UID: \"3dfad766-13d7-407a-82fc-92dcc742e8b5\") " pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:42 crc kubenswrapper[4775]: I1002 02:46:42.824753 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dfad766-13d7-407a-82fc-92dcc742e8b5-utilities\") pod \"redhat-operators-xgn4k\" (UID: \"3dfad766-13d7-407a-82fc-92dcc742e8b5\") " pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:42 crc kubenswrapper[4775]: I1002 02:46:42.926777 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dfad766-13d7-407a-82fc-92dcc742e8b5-utilities\") pod \"redhat-operators-xgn4k\" (UID: \"3dfad766-13d7-407a-82fc-92dcc742e8b5\") " pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:42 crc kubenswrapper[4775]: I1002 02:46:42.926883 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dfad766-13d7-407a-82fc-92dcc742e8b5-catalog-content\") pod \"redhat-operators-xgn4k\" (UID: \"3dfad766-13d7-407a-82fc-92dcc742e8b5\") " pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:42 crc kubenswrapper[4775]: I1002 02:46:42.926987 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjcmz\" (UniqueName: \"kubernetes.io/projected/3dfad766-13d7-407a-82fc-92dcc742e8b5-kube-api-access-rjcmz\") pod \"redhat-operators-xgn4k\" (UID: \"3dfad766-13d7-407a-82fc-92dcc742e8b5\") " pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:42 crc kubenswrapper[4775]: I1002 02:46:42.927887 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dfad766-13d7-407a-82fc-92dcc742e8b5-utilities\") pod \"redhat-operators-xgn4k\" (UID: \"3dfad766-13d7-407a-82fc-92dcc742e8b5\") " pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:42 crc kubenswrapper[4775]: I1002 02:46:42.927919 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dfad766-13d7-407a-82fc-92dcc742e8b5-catalog-content\") pod \"redhat-operators-xgn4k\" (UID: \"3dfad766-13d7-407a-82fc-92dcc742e8b5\") " pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:42 crc kubenswrapper[4775]: I1002 02:46:42.951006 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjcmz\" (UniqueName: \"kubernetes.io/projected/3dfad766-13d7-407a-82fc-92dcc742e8b5-kube-api-access-rjcmz\") pod \"redhat-operators-xgn4k\" (UID: \"3dfad766-13d7-407a-82fc-92dcc742e8b5\") " pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:43 crc kubenswrapper[4775]: I1002 02:46:43.069831 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:43 crc kubenswrapper[4775]: I1002 02:46:43.324521 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xgn4k"] Oct 02 02:46:44 crc kubenswrapper[4775]: I1002 02:46:44.078796 4775 generic.go:334] "Generic (PLEG): container finished" podID="3dfad766-13d7-407a-82fc-92dcc742e8b5" containerID="233d0b12e0323c74355607b0122b1d23656696e27a7f858e69a93952772d8696" exitCode=0 Oct 02 02:46:44 crc kubenswrapper[4775]: I1002 02:46:44.078887 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgn4k" event={"ID":"3dfad766-13d7-407a-82fc-92dcc742e8b5","Type":"ContainerDied","Data":"233d0b12e0323c74355607b0122b1d23656696e27a7f858e69a93952772d8696"} Oct 02 02:46:44 crc kubenswrapper[4775]: I1002 02:46:44.079138 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgn4k" event={"ID":"3dfad766-13d7-407a-82fc-92dcc742e8b5","Type":"ContainerStarted","Data":"4987a67c628031a7b8eee279de20b313e441ccce68dab23f3ef2fefd127c60c9"} Oct 02 02:46:44 crc kubenswrapper[4775]: I1002 02:46:44.082401 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 02:46:45 crc kubenswrapper[4775]: I1002 02:46:45.090699 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgn4k" event={"ID":"3dfad766-13d7-407a-82fc-92dcc742e8b5","Type":"ContainerStarted","Data":"21059a4c1e5aca257baef10bb911b604ea0c1aafdad84ee4b899e80ddc717f65"} Oct 02 02:46:46 crc kubenswrapper[4775]: I1002 02:46:46.103037 4775 generic.go:334] "Generic (PLEG): container finished" podID="3dfad766-13d7-407a-82fc-92dcc742e8b5" containerID="21059a4c1e5aca257baef10bb911b604ea0c1aafdad84ee4b899e80ddc717f65" exitCode=0 Oct 02 02:46:46 crc kubenswrapper[4775]: I1002 02:46:46.103178 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgn4k" event={"ID":"3dfad766-13d7-407a-82fc-92dcc742e8b5","Type":"ContainerDied","Data":"21059a4c1e5aca257baef10bb911b604ea0c1aafdad84ee4b899e80ddc717f65"} Oct 02 02:46:47 crc kubenswrapper[4775]: I1002 02:46:47.117012 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgn4k" event={"ID":"3dfad766-13d7-407a-82fc-92dcc742e8b5","Type":"ContainerStarted","Data":"e786135d3055585c30135bd96c70088585621c8575980f43035805a50ba30039"} Oct 02 02:46:47 crc kubenswrapper[4775]: I1002 02:46:47.152228 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xgn4k" podStartSLOduration=2.697400366 podStartE2EDuration="5.152195269s" podCreationTimestamp="2025-10-02 02:46:42 +0000 UTC" firstStartedPulling="2025-10-02 02:46:44.08206256 +0000 UTC m=+3941.248806600" lastFinishedPulling="2025-10-02 02:46:46.536857423 +0000 UTC m=+3943.703601503" observedRunningTime="2025-10-02 02:46:47.142479316 +0000 UTC m=+3944.309223436" watchObservedRunningTime="2025-10-02 02:46:47.152195269 +0000 UTC m=+3944.318939349" Oct 02 02:46:53 crc kubenswrapper[4775]: I1002 02:46:53.071056 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:53 crc kubenswrapper[4775]: I1002 02:46:53.071896 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:53 crc kubenswrapper[4775]: I1002 02:46:53.153920 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:53 crc kubenswrapper[4775]: I1002 02:46:53.247284 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:53 crc kubenswrapper[4775]: I1002 02:46:53.404311 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xgn4k"] Oct 02 02:46:55 crc kubenswrapper[4775]: I1002 02:46:55.188404 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xgn4k" podUID="3dfad766-13d7-407a-82fc-92dcc742e8b5" containerName="registry-server" containerID="cri-o://e786135d3055585c30135bd96c70088585621c8575980f43035805a50ba30039" gracePeriod=2 Oct 02 02:46:55 crc kubenswrapper[4775]: I1002 02:46:55.978146 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.026450 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dfad766-13d7-407a-82fc-92dcc742e8b5-utilities\") pod \"3dfad766-13d7-407a-82fc-92dcc742e8b5\" (UID: \"3dfad766-13d7-407a-82fc-92dcc742e8b5\") " Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.026544 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dfad766-13d7-407a-82fc-92dcc742e8b5-catalog-content\") pod \"3dfad766-13d7-407a-82fc-92dcc742e8b5\" (UID: \"3dfad766-13d7-407a-82fc-92dcc742e8b5\") " Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.026604 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjcmz\" (UniqueName: \"kubernetes.io/projected/3dfad766-13d7-407a-82fc-92dcc742e8b5-kube-api-access-rjcmz\") pod \"3dfad766-13d7-407a-82fc-92dcc742e8b5\" (UID: \"3dfad766-13d7-407a-82fc-92dcc742e8b5\") " Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.027666 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dfad766-13d7-407a-82fc-92dcc742e8b5-utilities" (OuterVolumeSpecName: "utilities") pod "3dfad766-13d7-407a-82fc-92dcc742e8b5" (UID: "3dfad766-13d7-407a-82fc-92dcc742e8b5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.032988 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dfad766-13d7-407a-82fc-92dcc742e8b5-kube-api-access-rjcmz" (OuterVolumeSpecName: "kube-api-access-rjcmz") pod "3dfad766-13d7-407a-82fc-92dcc742e8b5" (UID: "3dfad766-13d7-407a-82fc-92dcc742e8b5"). InnerVolumeSpecName "kube-api-access-rjcmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.129253 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3dfad766-13d7-407a-82fc-92dcc742e8b5-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.129294 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjcmz\" (UniqueName: \"kubernetes.io/projected/3dfad766-13d7-407a-82fc-92dcc742e8b5-kube-api-access-rjcmz\") on node \"crc\" DevicePath \"\"" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.199907 4775 generic.go:334] "Generic (PLEG): container finished" podID="3dfad766-13d7-407a-82fc-92dcc742e8b5" containerID="e786135d3055585c30135bd96c70088585621c8575980f43035805a50ba30039" exitCode=0 Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.199991 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgn4k" event={"ID":"3dfad766-13d7-407a-82fc-92dcc742e8b5","Type":"ContainerDied","Data":"e786135d3055585c30135bd96c70088585621c8575980f43035805a50ba30039"} Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.199997 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xgn4k" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.200029 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xgn4k" event={"ID":"3dfad766-13d7-407a-82fc-92dcc742e8b5","Type":"ContainerDied","Data":"4987a67c628031a7b8eee279de20b313e441ccce68dab23f3ef2fefd127c60c9"} Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.200044 4775 scope.go:117] "RemoveContainer" containerID="e786135d3055585c30135bd96c70088585621c8575980f43035805a50ba30039" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.228851 4775 scope.go:117] "RemoveContainer" containerID="21059a4c1e5aca257baef10bb911b604ea0c1aafdad84ee4b899e80ddc717f65" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.250327 4775 scope.go:117] "RemoveContainer" containerID="233d0b12e0323c74355607b0122b1d23656696e27a7f858e69a93952772d8696" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.272261 4775 scope.go:117] "RemoveContainer" containerID="e786135d3055585c30135bd96c70088585621c8575980f43035805a50ba30039" Oct 02 02:46:56 crc kubenswrapper[4775]: E1002 02:46:56.272767 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e786135d3055585c30135bd96c70088585621c8575980f43035805a50ba30039\": container with ID starting with e786135d3055585c30135bd96c70088585621c8575980f43035805a50ba30039 not found: ID does not exist" containerID="e786135d3055585c30135bd96c70088585621c8575980f43035805a50ba30039" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.272815 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e786135d3055585c30135bd96c70088585621c8575980f43035805a50ba30039"} err="failed to get container status \"e786135d3055585c30135bd96c70088585621c8575980f43035805a50ba30039\": rpc error: code = NotFound desc = could not find container \"e786135d3055585c30135bd96c70088585621c8575980f43035805a50ba30039\": container with ID starting with e786135d3055585c30135bd96c70088585621c8575980f43035805a50ba30039 not found: ID does not exist" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.272848 4775 scope.go:117] "RemoveContainer" containerID="21059a4c1e5aca257baef10bb911b604ea0c1aafdad84ee4b899e80ddc717f65" Oct 02 02:46:56 crc kubenswrapper[4775]: E1002 02:46:56.273250 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21059a4c1e5aca257baef10bb911b604ea0c1aafdad84ee4b899e80ddc717f65\": container with ID starting with 21059a4c1e5aca257baef10bb911b604ea0c1aafdad84ee4b899e80ddc717f65 not found: ID does not exist" containerID="21059a4c1e5aca257baef10bb911b604ea0c1aafdad84ee4b899e80ddc717f65" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.273277 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21059a4c1e5aca257baef10bb911b604ea0c1aafdad84ee4b899e80ddc717f65"} err="failed to get container status \"21059a4c1e5aca257baef10bb911b604ea0c1aafdad84ee4b899e80ddc717f65\": rpc error: code = NotFound desc = could not find container \"21059a4c1e5aca257baef10bb911b604ea0c1aafdad84ee4b899e80ddc717f65\": container with ID starting with 21059a4c1e5aca257baef10bb911b604ea0c1aafdad84ee4b899e80ddc717f65 not found: ID does not exist" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.273301 4775 scope.go:117] "RemoveContainer" containerID="233d0b12e0323c74355607b0122b1d23656696e27a7f858e69a93952772d8696" Oct 02 02:46:56 crc kubenswrapper[4775]: E1002 02:46:56.273568 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"233d0b12e0323c74355607b0122b1d23656696e27a7f858e69a93952772d8696\": container with ID starting with 233d0b12e0323c74355607b0122b1d23656696e27a7f858e69a93952772d8696 not found: ID does not exist" containerID="233d0b12e0323c74355607b0122b1d23656696e27a7f858e69a93952772d8696" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.273598 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"233d0b12e0323c74355607b0122b1d23656696e27a7f858e69a93952772d8696"} err="failed to get container status \"233d0b12e0323c74355607b0122b1d23656696e27a7f858e69a93952772d8696\": rpc error: code = NotFound desc = could not find container \"233d0b12e0323c74355607b0122b1d23656696e27a7f858e69a93952772d8696\": container with ID starting with 233d0b12e0323c74355607b0122b1d23656696e27a7f858e69a93952772d8696 not found: ID does not exist" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.560024 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dfad766-13d7-407a-82fc-92dcc742e8b5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3dfad766-13d7-407a-82fc-92dcc742e8b5" (UID: "3dfad766-13d7-407a-82fc-92dcc742e8b5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.637451 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3dfad766-13d7-407a-82fc-92dcc742e8b5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.843098 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xgn4k"] Oct 02 02:46:56 crc kubenswrapper[4775]: I1002 02:46:56.848708 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xgn4k"] Oct 02 02:46:57 crc kubenswrapper[4775]: I1002 02:46:57.779098 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dfad766-13d7-407a-82fc-92dcc742e8b5" path="/var/lib/kubelet/pods/3dfad766-13d7-407a-82fc-92dcc742e8b5/volumes" Oct 02 02:47:48 crc kubenswrapper[4775]: I1002 02:47:48.996700 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5v2w8"] Oct 02 02:47:48 crc kubenswrapper[4775]: E1002 02:47:48.998062 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dfad766-13d7-407a-82fc-92dcc742e8b5" containerName="registry-server" Oct 02 02:47:48 crc kubenswrapper[4775]: I1002 02:47:48.998100 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dfad766-13d7-407a-82fc-92dcc742e8b5" containerName="registry-server" Oct 02 02:47:48 crc kubenswrapper[4775]: E1002 02:47:48.998133 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dfad766-13d7-407a-82fc-92dcc742e8b5" containerName="extract-utilities" Oct 02 02:47:48 crc kubenswrapper[4775]: I1002 02:47:48.998151 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dfad766-13d7-407a-82fc-92dcc742e8b5" containerName="extract-utilities" Oct 02 02:47:48 crc kubenswrapper[4775]: E1002 02:47:48.998201 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dfad766-13d7-407a-82fc-92dcc742e8b5" containerName="extract-content" Oct 02 02:47:48 crc kubenswrapper[4775]: I1002 02:47:48.998223 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dfad766-13d7-407a-82fc-92dcc742e8b5" containerName="extract-content" Oct 02 02:47:48 crc kubenswrapper[4775]: I1002 02:47:48.998676 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dfad766-13d7-407a-82fc-92dcc742e8b5" containerName="registry-server" Oct 02 02:47:49 crc kubenswrapper[4775]: I1002 02:47:49.001300 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:49 crc kubenswrapper[4775]: I1002 02:47:49.032108 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5v2w8"] Oct 02 02:47:49 crc kubenswrapper[4775]: I1002 02:47:49.158430 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdp2k\" (UniqueName: \"kubernetes.io/projected/56b99c39-a245-4898-88a6-ec8270bb553c-kube-api-access-vdp2k\") pod \"community-operators-5v2w8\" (UID: \"56b99c39-a245-4898-88a6-ec8270bb553c\") " pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:49 crc kubenswrapper[4775]: I1002 02:47:49.158503 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56b99c39-a245-4898-88a6-ec8270bb553c-utilities\") pod \"community-operators-5v2w8\" (UID: \"56b99c39-a245-4898-88a6-ec8270bb553c\") " pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:49 crc kubenswrapper[4775]: I1002 02:47:49.158545 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56b99c39-a245-4898-88a6-ec8270bb553c-catalog-content\") pod \"community-operators-5v2w8\" (UID: \"56b99c39-a245-4898-88a6-ec8270bb553c\") " pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:49 crc kubenswrapper[4775]: I1002 02:47:49.260153 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdp2k\" (UniqueName: \"kubernetes.io/projected/56b99c39-a245-4898-88a6-ec8270bb553c-kube-api-access-vdp2k\") pod \"community-operators-5v2w8\" (UID: \"56b99c39-a245-4898-88a6-ec8270bb553c\") " pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:49 crc kubenswrapper[4775]: I1002 02:47:49.260208 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56b99c39-a245-4898-88a6-ec8270bb553c-utilities\") pod \"community-operators-5v2w8\" (UID: \"56b99c39-a245-4898-88a6-ec8270bb553c\") " pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:49 crc kubenswrapper[4775]: I1002 02:47:49.260232 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56b99c39-a245-4898-88a6-ec8270bb553c-catalog-content\") pod \"community-operators-5v2w8\" (UID: \"56b99c39-a245-4898-88a6-ec8270bb553c\") " pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:49 crc kubenswrapper[4775]: I1002 02:47:49.260661 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56b99c39-a245-4898-88a6-ec8270bb553c-catalog-content\") pod \"community-operators-5v2w8\" (UID: \"56b99c39-a245-4898-88a6-ec8270bb553c\") " pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:49 crc kubenswrapper[4775]: I1002 02:47:49.260799 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56b99c39-a245-4898-88a6-ec8270bb553c-utilities\") pod \"community-operators-5v2w8\" (UID: \"56b99c39-a245-4898-88a6-ec8270bb553c\") " pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:49 crc kubenswrapper[4775]: I1002 02:47:49.293658 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdp2k\" (UniqueName: \"kubernetes.io/projected/56b99c39-a245-4898-88a6-ec8270bb553c-kube-api-access-vdp2k\") pod \"community-operators-5v2w8\" (UID: \"56b99c39-a245-4898-88a6-ec8270bb553c\") " pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:49 crc kubenswrapper[4775]: I1002 02:47:49.330347 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:49 crc kubenswrapper[4775]: I1002 02:47:49.826776 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5v2w8"] Oct 02 02:47:50 crc kubenswrapper[4775]: I1002 02:47:50.706741 4775 generic.go:334] "Generic (PLEG): container finished" podID="56b99c39-a245-4898-88a6-ec8270bb553c" containerID="5e044803cb7fff1671d6e9e8b7204a567221aba71271537a4be81be61df8e7cb" exitCode=0 Oct 02 02:47:50 crc kubenswrapper[4775]: I1002 02:47:50.706783 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5v2w8" event={"ID":"56b99c39-a245-4898-88a6-ec8270bb553c","Type":"ContainerDied","Data":"5e044803cb7fff1671d6e9e8b7204a567221aba71271537a4be81be61df8e7cb"} Oct 02 02:47:50 crc kubenswrapper[4775]: I1002 02:47:50.706808 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5v2w8" event={"ID":"56b99c39-a245-4898-88a6-ec8270bb553c","Type":"ContainerStarted","Data":"65b733f922b32c115c02e1e3594385eaab7fb0c041d2adbebc3da5dd6181a9d3"} Oct 02 02:47:51 crc kubenswrapper[4775]: I1002 02:47:51.719397 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5v2w8" event={"ID":"56b99c39-a245-4898-88a6-ec8270bb553c","Type":"ContainerStarted","Data":"17cbca73f710221b6d027ee2d234f17ac093c86f7df37727b8de4d899af6cbf4"} Oct 02 02:47:52 crc kubenswrapper[4775]: I1002 02:47:52.733215 4775 generic.go:334] "Generic (PLEG): container finished" podID="56b99c39-a245-4898-88a6-ec8270bb553c" containerID="17cbca73f710221b6d027ee2d234f17ac093c86f7df37727b8de4d899af6cbf4" exitCode=0 Oct 02 02:47:52 crc kubenswrapper[4775]: I1002 02:47:52.733292 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5v2w8" event={"ID":"56b99c39-a245-4898-88a6-ec8270bb553c","Type":"ContainerDied","Data":"17cbca73f710221b6d027ee2d234f17ac093c86f7df37727b8de4d899af6cbf4"} Oct 02 02:47:53 crc kubenswrapper[4775]: I1002 02:47:53.746407 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5v2w8" event={"ID":"56b99c39-a245-4898-88a6-ec8270bb553c","Type":"ContainerStarted","Data":"028c3b85fcfe0dfbe0d85ef95877ef16e613c62cb04e4d3b7e469f286df2122e"} Oct 02 02:47:53 crc kubenswrapper[4775]: I1002 02:47:53.793837 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5v2w8" podStartSLOduration=3.360193034 podStartE2EDuration="5.793811146s" podCreationTimestamp="2025-10-02 02:47:48 +0000 UTC" firstStartedPulling="2025-10-02 02:47:50.708901761 +0000 UTC m=+4007.875645841" lastFinishedPulling="2025-10-02 02:47:53.142519903 +0000 UTC m=+4010.309263953" observedRunningTime="2025-10-02 02:47:53.784206926 +0000 UTC m=+4010.950950996" watchObservedRunningTime="2025-10-02 02:47:53.793811146 +0000 UTC m=+4010.960555226" Oct 02 02:47:59 crc kubenswrapper[4775]: I1002 02:47:59.331295 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:59 crc kubenswrapper[4775]: I1002 02:47:59.331900 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:59 crc kubenswrapper[4775]: I1002 02:47:59.397946 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:59 crc kubenswrapper[4775]: I1002 02:47:59.872824 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:47:59 crc kubenswrapper[4775]: I1002 02:47:59.939773 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5v2w8"] Oct 02 02:48:01 crc kubenswrapper[4775]: I1002 02:48:01.814918 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5v2w8" podUID="56b99c39-a245-4898-88a6-ec8270bb553c" containerName="registry-server" containerID="cri-o://028c3b85fcfe0dfbe0d85ef95877ef16e613c62cb04e4d3b7e469f286df2122e" gracePeriod=2 Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.216732 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.367393 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56b99c39-a245-4898-88a6-ec8270bb553c-catalog-content\") pod \"56b99c39-a245-4898-88a6-ec8270bb553c\" (UID: \"56b99c39-a245-4898-88a6-ec8270bb553c\") " Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.367508 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdp2k\" (UniqueName: \"kubernetes.io/projected/56b99c39-a245-4898-88a6-ec8270bb553c-kube-api-access-vdp2k\") pod \"56b99c39-a245-4898-88a6-ec8270bb553c\" (UID: \"56b99c39-a245-4898-88a6-ec8270bb553c\") " Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.367562 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56b99c39-a245-4898-88a6-ec8270bb553c-utilities\") pod \"56b99c39-a245-4898-88a6-ec8270bb553c\" (UID: \"56b99c39-a245-4898-88a6-ec8270bb553c\") " Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.369414 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56b99c39-a245-4898-88a6-ec8270bb553c-utilities" (OuterVolumeSpecName: "utilities") pod "56b99c39-a245-4898-88a6-ec8270bb553c" (UID: "56b99c39-a245-4898-88a6-ec8270bb553c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.375364 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56b99c39-a245-4898-88a6-ec8270bb553c-kube-api-access-vdp2k" (OuterVolumeSpecName: "kube-api-access-vdp2k") pod "56b99c39-a245-4898-88a6-ec8270bb553c" (UID: "56b99c39-a245-4898-88a6-ec8270bb553c"). InnerVolumeSpecName "kube-api-access-vdp2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.466246 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56b99c39-a245-4898-88a6-ec8270bb553c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56b99c39-a245-4898-88a6-ec8270bb553c" (UID: "56b99c39-a245-4898-88a6-ec8270bb553c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.469613 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56b99c39-a245-4898-88a6-ec8270bb553c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.469656 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56b99c39-a245-4898-88a6-ec8270bb553c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.469677 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdp2k\" (UniqueName: \"kubernetes.io/projected/56b99c39-a245-4898-88a6-ec8270bb553c-kube-api-access-vdp2k\") on node \"crc\" DevicePath \"\"" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.827158 4775 generic.go:334] "Generic (PLEG): container finished" podID="56b99c39-a245-4898-88a6-ec8270bb553c" containerID="028c3b85fcfe0dfbe0d85ef95877ef16e613c62cb04e4d3b7e469f286df2122e" exitCode=0 Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.827220 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5v2w8" event={"ID":"56b99c39-a245-4898-88a6-ec8270bb553c","Type":"ContainerDied","Data":"028c3b85fcfe0dfbe0d85ef95877ef16e613c62cb04e4d3b7e469f286df2122e"} Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.827260 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5v2w8" event={"ID":"56b99c39-a245-4898-88a6-ec8270bb553c","Type":"ContainerDied","Data":"65b733f922b32c115c02e1e3594385eaab7fb0c041d2adbebc3da5dd6181a9d3"} Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.827293 4775 scope.go:117] "RemoveContainer" containerID="028c3b85fcfe0dfbe0d85ef95877ef16e613c62cb04e4d3b7e469f286df2122e" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.827459 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5v2w8" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.873627 4775 scope.go:117] "RemoveContainer" containerID="17cbca73f710221b6d027ee2d234f17ac093c86f7df37727b8de4d899af6cbf4" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.884573 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5v2w8"] Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.899737 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5v2w8"] Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.915488 4775 scope.go:117] "RemoveContainer" containerID="5e044803cb7fff1671d6e9e8b7204a567221aba71271537a4be81be61df8e7cb" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.951778 4775 scope.go:117] "RemoveContainer" containerID="028c3b85fcfe0dfbe0d85ef95877ef16e613c62cb04e4d3b7e469f286df2122e" Oct 02 02:48:02 crc kubenswrapper[4775]: E1002 02:48:02.952392 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"028c3b85fcfe0dfbe0d85ef95877ef16e613c62cb04e4d3b7e469f286df2122e\": container with ID starting with 028c3b85fcfe0dfbe0d85ef95877ef16e613c62cb04e4d3b7e469f286df2122e not found: ID does not exist" containerID="028c3b85fcfe0dfbe0d85ef95877ef16e613c62cb04e4d3b7e469f286df2122e" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.952504 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"028c3b85fcfe0dfbe0d85ef95877ef16e613c62cb04e4d3b7e469f286df2122e"} err="failed to get container status \"028c3b85fcfe0dfbe0d85ef95877ef16e613c62cb04e4d3b7e469f286df2122e\": rpc error: code = NotFound desc = could not find container \"028c3b85fcfe0dfbe0d85ef95877ef16e613c62cb04e4d3b7e469f286df2122e\": container with ID starting with 028c3b85fcfe0dfbe0d85ef95877ef16e613c62cb04e4d3b7e469f286df2122e not found: ID does not exist" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.952535 4775 scope.go:117] "RemoveContainer" containerID="17cbca73f710221b6d027ee2d234f17ac093c86f7df37727b8de4d899af6cbf4" Oct 02 02:48:02 crc kubenswrapper[4775]: E1002 02:48:02.952923 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17cbca73f710221b6d027ee2d234f17ac093c86f7df37727b8de4d899af6cbf4\": container with ID starting with 17cbca73f710221b6d027ee2d234f17ac093c86f7df37727b8de4d899af6cbf4 not found: ID does not exist" containerID="17cbca73f710221b6d027ee2d234f17ac093c86f7df37727b8de4d899af6cbf4" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.953033 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17cbca73f710221b6d027ee2d234f17ac093c86f7df37727b8de4d899af6cbf4"} err="failed to get container status \"17cbca73f710221b6d027ee2d234f17ac093c86f7df37727b8de4d899af6cbf4\": rpc error: code = NotFound desc = could not find container \"17cbca73f710221b6d027ee2d234f17ac093c86f7df37727b8de4d899af6cbf4\": container with ID starting with 17cbca73f710221b6d027ee2d234f17ac093c86f7df37727b8de4d899af6cbf4 not found: ID does not exist" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.953073 4775 scope.go:117] "RemoveContainer" containerID="5e044803cb7fff1671d6e9e8b7204a567221aba71271537a4be81be61df8e7cb" Oct 02 02:48:02 crc kubenswrapper[4775]: E1002 02:48:02.953635 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e044803cb7fff1671d6e9e8b7204a567221aba71271537a4be81be61df8e7cb\": container with ID starting with 5e044803cb7fff1671d6e9e8b7204a567221aba71271537a4be81be61df8e7cb not found: ID does not exist" containerID="5e044803cb7fff1671d6e9e8b7204a567221aba71271537a4be81be61df8e7cb" Oct 02 02:48:02 crc kubenswrapper[4775]: I1002 02:48:02.953709 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e044803cb7fff1671d6e9e8b7204a567221aba71271537a4be81be61df8e7cb"} err="failed to get container status \"5e044803cb7fff1671d6e9e8b7204a567221aba71271537a4be81be61df8e7cb\": rpc error: code = NotFound desc = could not find container \"5e044803cb7fff1671d6e9e8b7204a567221aba71271537a4be81be61df8e7cb\": container with ID starting with 5e044803cb7fff1671d6e9e8b7204a567221aba71271537a4be81be61df8e7cb not found: ID does not exist" Oct 02 02:48:03 crc kubenswrapper[4775]: I1002 02:48:03.786682 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56b99c39-a245-4898-88a6-ec8270bb553c" path="/var/lib/kubelet/pods/56b99c39-a245-4898-88a6-ec8270bb553c/volumes" Oct 02 02:48:37 crc kubenswrapper[4775]: I1002 02:48:37.234397 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:48:37 crc kubenswrapper[4775]: I1002 02:48:37.235277 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.293372 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xkgcv"] Oct 02 02:48:42 crc kubenswrapper[4775]: E1002 02:48:42.293806 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56b99c39-a245-4898-88a6-ec8270bb553c" containerName="extract-content" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.293825 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="56b99c39-a245-4898-88a6-ec8270bb553c" containerName="extract-content" Oct 02 02:48:42 crc kubenswrapper[4775]: E1002 02:48:42.293843 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56b99c39-a245-4898-88a6-ec8270bb553c" containerName="registry-server" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.293851 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="56b99c39-a245-4898-88a6-ec8270bb553c" containerName="registry-server" Oct 02 02:48:42 crc kubenswrapper[4775]: E1002 02:48:42.293884 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56b99c39-a245-4898-88a6-ec8270bb553c" containerName="extract-utilities" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.293892 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="56b99c39-a245-4898-88a6-ec8270bb553c" containerName="extract-utilities" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.294328 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="56b99c39-a245-4898-88a6-ec8270bb553c" containerName="registry-server" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.295916 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.307364 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkgcv"] Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.424495 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42af4826-7011-4609-be36-5b70f5980c8e-catalog-content\") pod \"redhat-marketplace-xkgcv\" (UID: \"42af4826-7011-4609-be36-5b70f5980c8e\") " pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.424879 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2btx\" (UniqueName: \"kubernetes.io/projected/42af4826-7011-4609-be36-5b70f5980c8e-kube-api-access-f2btx\") pod \"redhat-marketplace-xkgcv\" (UID: \"42af4826-7011-4609-be36-5b70f5980c8e\") " pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.425088 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42af4826-7011-4609-be36-5b70f5980c8e-utilities\") pod \"redhat-marketplace-xkgcv\" (UID: \"42af4826-7011-4609-be36-5b70f5980c8e\") " pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.526355 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42af4826-7011-4609-be36-5b70f5980c8e-catalog-content\") pod \"redhat-marketplace-xkgcv\" (UID: \"42af4826-7011-4609-be36-5b70f5980c8e\") " pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.526425 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2btx\" (UniqueName: \"kubernetes.io/projected/42af4826-7011-4609-be36-5b70f5980c8e-kube-api-access-f2btx\") pod \"redhat-marketplace-xkgcv\" (UID: \"42af4826-7011-4609-be36-5b70f5980c8e\") " pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.526474 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42af4826-7011-4609-be36-5b70f5980c8e-utilities\") pod \"redhat-marketplace-xkgcv\" (UID: \"42af4826-7011-4609-be36-5b70f5980c8e\") " pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.526848 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42af4826-7011-4609-be36-5b70f5980c8e-catalog-content\") pod \"redhat-marketplace-xkgcv\" (UID: \"42af4826-7011-4609-be36-5b70f5980c8e\") " pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.526992 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42af4826-7011-4609-be36-5b70f5980c8e-utilities\") pod \"redhat-marketplace-xkgcv\" (UID: \"42af4826-7011-4609-be36-5b70f5980c8e\") " pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.556177 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2btx\" (UniqueName: \"kubernetes.io/projected/42af4826-7011-4609-be36-5b70f5980c8e-kube-api-access-f2btx\") pod \"redhat-marketplace-xkgcv\" (UID: \"42af4826-7011-4609-be36-5b70f5980c8e\") " pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:42 crc kubenswrapper[4775]: I1002 02:48:42.628002 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:43 crc kubenswrapper[4775]: I1002 02:48:43.055710 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkgcv"] Oct 02 02:48:43 crc kubenswrapper[4775]: I1002 02:48:43.189912 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkgcv" event={"ID":"42af4826-7011-4609-be36-5b70f5980c8e","Type":"ContainerStarted","Data":"5557dd13b5d62018253600e8ebf78fcb1e854b269074864671e274e1ba1cf586"} Oct 02 02:48:44 crc kubenswrapper[4775]: I1002 02:48:44.200772 4775 generic.go:334] "Generic (PLEG): container finished" podID="42af4826-7011-4609-be36-5b70f5980c8e" containerID="a1202c5b48d5af0f31056ba6295227d4d98e3ce70ac7a6ad4408383600f88e3f" exitCode=0 Oct 02 02:48:44 crc kubenswrapper[4775]: I1002 02:48:44.200867 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkgcv" event={"ID":"42af4826-7011-4609-be36-5b70f5980c8e","Type":"ContainerDied","Data":"a1202c5b48d5af0f31056ba6295227d4d98e3ce70ac7a6ad4408383600f88e3f"} Oct 02 02:48:45 crc kubenswrapper[4775]: I1002 02:48:45.214826 4775 generic.go:334] "Generic (PLEG): container finished" podID="42af4826-7011-4609-be36-5b70f5980c8e" containerID="82fd75f8d8456a77452fbe47696bebc73a07cb7aeb85575ddf38ca2e7aa0412e" exitCode=0 Oct 02 02:48:45 crc kubenswrapper[4775]: I1002 02:48:45.214920 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkgcv" event={"ID":"42af4826-7011-4609-be36-5b70f5980c8e","Type":"ContainerDied","Data":"82fd75f8d8456a77452fbe47696bebc73a07cb7aeb85575ddf38ca2e7aa0412e"} Oct 02 02:48:46 crc kubenswrapper[4775]: I1002 02:48:46.227150 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkgcv" event={"ID":"42af4826-7011-4609-be36-5b70f5980c8e","Type":"ContainerStarted","Data":"ea9af451c86fb6b0f7477787bdce8f859259277c7da6b210bbd46f05a872a498"} Oct 02 02:48:52 crc kubenswrapper[4775]: I1002 02:48:52.628543 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:52 crc kubenswrapper[4775]: I1002 02:48:52.628970 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:52 crc kubenswrapper[4775]: I1002 02:48:52.716351 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:52 crc kubenswrapper[4775]: I1002 02:48:52.755949 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xkgcv" podStartSLOduration=9.330415847 podStartE2EDuration="10.755921773s" podCreationTimestamp="2025-10-02 02:48:42 +0000 UTC" firstStartedPulling="2025-10-02 02:48:44.202759031 +0000 UTC m=+4061.369503101" lastFinishedPulling="2025-10-02 02:48:45.628264957 +0000 UTC m=+4062.795009027" observedRunningTime="2025-10-02 02:48:46.247511605 +0000 UTC m=+4063.414255655" watchObservedRunningTime="2025-10-02 02:48:52.755921773 +0000 UTC m=+4069.922665853" Oct 02 02:48:53 crc kubenswrapper[4775]: I1002 02:48:53.376475 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:53 crc kubenswrapper[4775]: I1002 02:48:53.447350 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkgcv"] Oct 02 02:48:55 crc kubenswrapper[4775]: I1002 02:48:55.318922 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xkgcv" podUID="42af4826-7011-4609-be36-5b70f5980c8e" containerName="registry-server" containerID="cri-o://ea9af451c86fb6b0f7477787bdce8f859259277c7da6b210bbd46f05a872a498" gracePeriod=2 Oct 02 02:48:55 crc kubenswrapper[4775]: I1002 02:48:55.798781 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:55 crc kubenswrapper[4775]: I1002 02:48:55.973791 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2btx\" (UniqueName: \"kubernetes.io/projected/42af4826-7011-4609-be36-5b70f5980c8e-kube-api-access-f2btx\") pod \"42af4826-7011-4609-be36-5b70f5980c8e\" (UID: \"42af4826-7011-4609-be36-5b70f5980c8e\") " Oct 02 02:48:55 crc kubenswrapper[4775]: I1002 02:48:55.973878 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42af4826-7011-4609-be36-5b70f5980c8e-utilities\") pod \"42af4826-7011-4609-be36-5b70f5980c8e\" (UID: \"42af4826-7011-4609-be36-5b70f5980c8e\") " Oct 02 02:48:55 crc kubenswrapper[4775]: I1002 02:48:55.973919 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42af4826-7011-4609-be36-5b70f5980c8e-catalog-content\") pod \"42af4826-7011-4609-be36-5b70f5980c8e\" (UID: \"42af4826-7011-4609-be36-5b70f5980c8e\") " Oct 02 02:48:55 crc kubenswrapper[4775]: I1002 02:48:55.975225 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42af4826-7011-4609-be36-5b70f5980c8e-utilities" (OuterVolumeSpecName: "utilities") pod "42af4826-7011-4609-be36-5b70f5980c8e" (UID: "42af4826-7011-4609-be36-5b70f5980c8e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:48:55 crc kubenswrapper[4775]: I1002 02:48:55.981385 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42af4826-7011-4609-be36-5b70f5980c8e-kube-api-access-f2btx" (OuterVolumeSpecName: "kube-api-access-f2btx") pod "42af4826-7011-4609-be36-5b70f5980c8e" (UID: "42af4826-7011-4609-be36-5b70f5980c8e"). InnerVolumeSpecName "kube-api-access-f2btx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.002839 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42af4826-7011-4609-be36-5b70f5980c8e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42af4826-7011-4609-be36-5b70f5980c8e" (UID: "42af4826-7011-4609-be36-5b70f5980c8e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.075517 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42af4826-7011-4609-be36-5b70f5980c8e-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.075584 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42af4826-7011-4609-be36-5b70f5980c8e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.075605 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2btx\" (UniqueName: \"kubernetes.io/projected/42af4826-7011-4609-be36-5b70f5980c8e-kube-api-access-f2btx\") on node \"crc\" DevicePath \"\"" Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.327563 4775 generic.go:334] "Generic (PLEG): container finished" podID="42af4826-7011-4609-be36-5b70f5980c8e" containerID="ea9af451c86fb6b0f7477787bdce8f859259277c7da6b210bbd46f05a872a498" exitCode=0 Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.327612 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xkgcv" Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.327627 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkgcv" event={"ID":"42af4826-7011-4609-be36-5b70f5980c8e","Type":"ContainerDied","Data":"ea9af451c86fb6b0f7477787bdce8f859259277c7da6b210bbd46f05a872a498"} Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.328088 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xkgcv" event={"ID":"42af4826-7011-4609-be36-5b70f5980c8e","Type":"ContainerDied","Data":"5557dd13b5d62018253600e8ebf78fcb1e854b269074864671e274e1ba1cf586"} Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.328143 4775 scope.go:117] "RemoveContainer" containerID="ea9af451c86fb6b0f7477787bdce8f859259277c7da6b210bbd46f05a872a498" Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.350245 4775 scope.go:117] "RemoveContainer" containerID="82fd75f8d8456a77452fbe47696bebc73a07cb7aeb85575ddf38ca2e7aa0412e" Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.379253 4775 scope.go:117] "RemoveContainer" containerID="a1202c5b48d5af0f31056ba6295227d4d98e3ce70ac7a6ad4408383600f88e3f" Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.381415 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkgcv"] Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.388058 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xkgcv"] Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.435659 4775 scope.go:117] "RemoveContainer" containerID="ea9af451c86fb6b0f7477787bdce8f859259277c7da6b210bbd46f05a872a498" Oct 02 02:48:56 crc kubenswrapper[4775]: E1002 02:48:56.436206 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea9af451c86fb6b0f7477787bdce8f859259277c7da6b210bbd46f05a872a498\": container with ID starting with ea9af451c86fb6b0f7477787bdce8f859259277c7da6b210bbd46f05a872a498 not found: ID does not exist" containerID="ea9af451c86fb6b0f7477787bdce8f859259277c7da6b210bbd46f05a872a498" Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.436245 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea9af451c86fb6b0f7477787bdce8f859259277c7da6b210bbd46f05a872a498"} err="failed to get container status \"ea9af451c86fb6b0f7477787bdce8f859259277c7da6b210bbd46f05a872a498\": rpc error: code = NotFound desc = could not find container \"ea9af451c86fb6b0f7477787bdce8f859259277c7da6b210bbd46f05a872a498\": container with ID starting with ea9af451c86fb6b0f7477787bdce8f859259277c7da6b210bbd46f05a872a498 not found: ID does not exist" Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.436289 4775 scope.go:117] "RemoveContainer" containerID="82fd75f8d8456a77452fbe47696bebc73a07cb7aeb85575ddf38ca2e7aa0412e" Oct 02 02:48:56 crc kubenswrapper[4775]: E1002 02:48:56.436647 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82fd75f8d8456a77452fbe47696bebc73a07cb7aeb85575ddf38ca2e7aa0412e\": container with ID starting with 82fd75f8d8456a77452fbe47696bebc73a07cb7aeb85575ddf38ca2e7aa0412e not found: ID does not exist" containerID="82fd75f8d8456a77452fbe47696bebc73a07cb7aeb85575ddf38ca2e7aa0412e" Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.436693 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82fd75f8d8456a77452fbe47696bebc73a07cb7aeb85575ddf38ca2e7aa0412e"} err="failed to get container status \"82fd75f8d8456a77452fbe47696bebc73a07cb7aeb85575ddf38ca2e7aa0412e\": rpc error: code = NotFound desc = could not find container \"82fd75f8d8456a77452fbe47696bebc73a07cb7aeb85575ddf38ca2e7aa0412e\": container with ID starting with 82fd75f8d8456a77452fbe47696bebc73a07cb7aeb85575ddf38ca2e7aa0412e not found: ID does not exist" Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.436722 4775 scope.go:117] "RemoveContainer" containerID="a1202c5b48d5af0f31056ba6295227d4d98e3ce70ac7a6ad4408383600f88e3f" Oct 02 02:48:56 crc kubenswrapper[4775]: E1002 02:48:56.437039 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1202c5b48d5af0f31056ba6295227d4d98e3ce70ac7a6ad4408383600f88e3f\": container with ID starting with a1202c5b48d5af0f31056ba6295227d4d98e3ce70ac7a6ad4408383600f88e3f not found: ID does not exist" containerID="a1202c5b48d5af0f31056ba6295227d4d98e3ce70ac7a6ad4408383600f88e3f" Oct 02 02:48:56 crc kubenswrapper[4775]: I1002 02:48:56.437074 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1202c5b48d5af0f31056ba6295227d4d98e3ce70ac7a6ad4408383600f88e3f"} err="failed to get container status \"a1202c5b48d5af0f31056ba6295227d4d98e3ce70ac7a6ad4408383600f88e3f\": rpc error: code = NotFound desc = could not find container \"a1202c5b48d5af0f31056ba6295227d4d98e3ce70ac7a6ad4408383600f88e3f\": container with ID starting with a1202c5b48d5af0f31056ba6295227d4d98e3ce70ac7a6ad4408383600f88e3f not found: ID does not exist" Oct 02 02:48:57 crc kubenswrapper[4775]: I1002 02:48:57.777902 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42af4826-7011-4609-be36-5b70f5980c8e" path="/var/lib/kubelet/pods/42af4826-7011-4609-be36-5b70f5980c8e/volumes" Oct 02 02:49:07 crc kubenswrapper[4775]: I1002 02:49:07.233446 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:49:07 crc kubenswrapper[4775]: I1002 02:49:07.234261 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:49:37 crc kubenswrapper[4775]: I1002 02:49:37.233923 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:49:37 crc kubenswrapper[4775]: I1002 02:49:37.234693 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:49:37 crc kubenswrapper[4775]: I1002 02:49:37.234769 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 02:49:37 crc kubenswrapper[4775]: I1002 02:49:37.235614 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"701c2b59237bce746d64de4b4003ac93de1140e75c06122e0289272297fc7b80"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:49:37 crc kubenswrapper[4775]: I1002 02:49:37.235714 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://701c2b59237bce746d64de4b4003ac93de1140e75c06122e0289272297fc7b80" gracePeriod=600 Oct 02 02:49:37 crc kubenswrapper[4775]: I1002 02:49:37.717896 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="701c2b59237bce746d64de4b4003ac93de1140e75c06122e0289272297fc7b80" exitCode=0 Oct 02 02:49:37 crc kubenswrapper[4775]: I1002 02:49:37.717971 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"701c2b59237bce746d64de4b4003ac93de1140e75c06122e0289272297fc7b80"} Oct 02 02:49:37 crc kubenswrapper[4775]: I1002 02:49:37.718478 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c"} Oct 02 02:49:37 crc kubenswrapper[4775]: I1002 02:49:37.718528 4775 scope.go:117] "RemoveContainer" containerID="066de6c28e97df0ef0fab8b86ce79ca7efd13477af747b0226965d78b1ba9c95" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.624474 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jwszs"] Oct 02 02:50:30 crc kubenswrapper[4775]: E1002 02:50:30.625564 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42af4826-7011-4609-be36-5b70f5980c8e" containerName="extract-utilities" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.625588 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="42af4826-7011-4609-be36-5b70f5980c8e" containerName="extract-utilities" Oct 02 02:50:30 crc kubenswrapper[4775]: E1002 02:50:30.625628 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42af4826-7011-4609-be36-5b70f5980c8e" containerName="extract-content" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.625641 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="42af4826-7011-4609-be36-5b70f5980c8e" containerName="extract-content" Oct 02 02:50:30 crc kubenswrapper[4775]: E1002 02:50:30.625661 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42af4826-7011-4609-be36-5b70f5980c8e" containerName="registry-server" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.625674 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="42af4826-7011-4609-be36-5b70f5980c8e" containerName="registry-server" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.625905 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="42af4826-7011-4609-be36-5b70f5980c8e" containerName="registry-server" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.627619 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.652555 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jwszs"] Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.673133 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a30636-4fe0-422b-958f-000c6e109f4b-catalog-content\") pod \"certified-operators-jwszs\" (UID: \"55a30636-4fe0-422b-958f-000c6e109f4b\") " pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.673258 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a30636-4fe0-422b-958f-000c6e109f4b-utilities\") pod \"certified-operators-jwszs\" (UID: \"55a30636-4fe0-422b-958f-000c6e109f4b\") " pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.673303 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh5qq\" (UniqueName: \"kubernetes.io/projected/55a30636-4fe0-422b-958f-000c6e109f4b-kube-api-access-bh5qq\") pod \"certified-operators-jwszs\" (UID: \"55a30636-4fe0-422b-958f-000c6e109f4b\") " pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.775147 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a30636-4fe0-422b-958f-000c6e109f4b-catalog-content\") pod \"certified-operators-jwszs\" (UID: \"55a30636-4fe0-422b-958f-000c6e109f4b\") " pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.775408 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a30636-4fe0-422b-958f-000c6e109f4b-utilities\") pod \"certified-operators-jwszs\" (UID: \"55a30636-4fe0-422b-958f-000c6e109f4b\") " pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.775430 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh5qq\" (UniqueName: \"kubernetes.io/projected/55a30636-4fe0-422b-958f-000c6e109f4b-kube-api-access-bh5qq\") pod \"certified-operators-jwszs\" (UID: \"55a30636-4fe0-422b-958f-000c6e109f4b\") " pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.775903 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a30636-4fe0-422b-958f-000c6e109f4b-utilities\") pod \"certified-operators-jwszs\" (UID: \"55a30636-4fe0-422b-958f-000c6e109f4b\") " pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.776053 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a30636-4fe0-422b-958f-000c6e109f4b-catalog-content\") pod \"certified-operators-jwszs\" (UID: \"55a30636-4fe0-422b-958f-000c6e109f4b\") " pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.803941 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh5qq\" (UniqueName: \"kubernetes.io/projected/55a30636-4fe0-422b-958f-000c6e109f4b-kube-api-access-bh5qq\") pod \"certified-operators-jwszs\" (UID: \"55a30636-4fe0-422b-958f-000c6e109f4b\") " pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:30 crc kubenswrapper[4775]: I1002 02:50:30.959042 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:31 crc kubenswrapper[4775]: I1002 02:50:31.195183 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jwszs"] Oct 02 02:50:31 crc kubenswrapper[4775]: I1002 02:50:31.255637 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jwszs" event={"ID":"55a30636-4fe0-422b-958f-000c6e109f4b","Type":"ContainerStarted","Data":"c8919963046838ddc32f487cee15503caf0d23dba55c74b780296d0650f529e8"} Oct 02 02:50:32 crc kubenswrapper[4775]: I1002 02:50:32.269422 4775 generic.go:334] "Generic (PLEG): container finished" podID="55a30636-4fe0-422b-958f-000c6e109f4b" containerID="b8bbf36593d755278086388108ae81f3eee541f8fb1a847c4afd970a1b4b0de6" exitCode=0 Oct 02 02:50:32 crc kubenswrapper[4775]: I1002 02:50:32.269533 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jwszs" event={"ID":"55a30636-4fe0-422b-958f-000c6e109f4b","Type":"ContainerDied","Data":"b8bbf36593d755278086388108ae81f3eee541f8fb1a847c4afd970a1b4b0de6"} Oct 02 02:50:34 crc kubenswrapper[4775]: I1002 02:50:34.291814 4775 generic.go:334] "Generic (PLEG): container finished" podID="55a30636-4fe0-422b-958f-000c6e109f4b" containerID="e3c5c5b5f2e45c6e6e5f1ad375cc2f9a9c38cb6334334e4177375067d27117b2" exitCode=0 Oct 02 02:50:34 crc kubenswrapper[4775]: I1002 02:50:34.291887 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jwszs" event={"ID":"55a30636-4fe0-422b-958f-000c6e109f4b","Type":"ContainerDied","Data":"e3c5c5b5f2e45c6e6e5f1ad375cc2f9a9c38cb6334334e4177375067d27117b2"} Oct 02 02:50:35 crc kubenswrapper[4775]: I1002 02:50:35.303883 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jwszs" event={"ID":"55a30636-4fe0-422b-958f-000c6e109f4b","Type":"ContainerStarted","Data":"39b54c4f0d3c57986eb4822ebe0aaa3a907085e8ad4807f62e997b6c548fc579"} Oct 02 02:50:35 crc kubenswrapper[4775]: I1002 02:50:35.325725 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jwszs" podStartSLOduration=2.8947995730000002 podStartE2EDuration="5.325690305s" podCreationTimestamp="2025-10-02 02:50:30 +0000 UTC" firstStartedPulling="2025-10-02 02:50:32.272348743 +0000 UTC m=+4169.439092783" lastFinishedPulling="2025-10-02 02:50:34.703239445 +0000 UTC m=+4171.869983515" observedRunningTime="2025-10-02 02:50:35.322639706 +0000 UTC m=+4172.489383756" watchObservedRunningTime="2025-10-02 02:50:35.325690305 +0000 UTC m=+4172.492434355" Oct 02 02:50:40 crc kubenswrapper[4775]: I1002 02:50:40.960055 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:40 crc kubenswrapper[4775]: I1002 02:50:40.960749 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:41 crc kubenswrapper[4775]: I1002 02:50:41.044275 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:41 crc kubenswrapper[4775]: I1002 02:50:41.423136 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:41 crc kubenswrapper[4775]: I1002 02:50:41.492592 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jwszs"] Oct 02 02:50:43 crc kubenswrapper[4775]: I1002 02:50:43.369027 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jwszs" podUID="55a30636-4fe0-422b-958f-000c6e109f4b" containerName="registry-server" containerID="cri-o://39b54c4f0d3c57986eb4822ebe0aaa3a907085e8ad4807f62e997b6c548fc579" gracePeriod=2 Oct 02 02:50:43 crc kubenswrapper[4775]: I1002 02:50:43.832201 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:43 crc kubenswrapper[4775]: I1002 02:50:43.879331 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a30636-4fe0-422b-958f-000c6e109f4b-catalog-content\") pod \"55a30636-4fe0-422b-958f-000c6e109f4b\" (UID: \"55a30636-4fe0-422b-958f-000c6e109f4b\") " Oct 02 02:50:43 crc kubenswrapper[4775]: I1002 02:50:43.879448 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bh5qq\" (UniqueName: \"kubernetes.io/projected/55a30636-4fe0-422b-958f-000c6e109f4b-kube-api-access-bh5qq\") pod \"55a30636-4fe0-422b-958f-000c6e109f4b\" (UID: \"55a30636-4fe0-422b-958f-000c6e109f4b\") " Oct 02 02:50:43 crc kubenswrapper[4775]: I1002 02:50:43.879507 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a30636-4fe0-422b-958f-000c6e109f4b-utilities\") pod \"55a30636-4fe0-422b-958f-000c6e109f4b\" (UID: \"55a30636-4fe0-422b-958f-000c6e109f4b\") " Oct 02 02:50:43 crc kubenswrapper[4775]: I1002 02:50:43.881369 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55a30636-4fe0-422b-958f-000c6e109f4b-utilities" (OuterVolumeSpecName: "utilities") pod "55a30636-4fe0-422b-958f-000c6e109f4b" (UID: "55a30636-4fe0-422b-958f-000c6e109f4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:50:43 crc kubenswrapper[4775]: I1002 02:50:43.890487 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55a30636-4fe0-422b-958f-000c6e109f4b-kube-api-access-bh5qq" (OuterVolumeSpecName: "kube-api-access-bh5qq") pod "55a30636-4fe0-422b-958f-000c6e109f4b" (UID: "55a30636-4fe0-422b-958f-000c6e109f4b"). InnerVolumeSpecName "kube-api-access-bh5qq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:50:43 crc kubenswrapper[4775]: I1002 02:50:43.942375 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55a30636-4fe0-422b-958f-000c6e109f4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55a30636-4fe0-422b-958f-000c6e109f4b" (UID: "55a30636-4fe0-422b-958f-000c6e109f4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:50:43 crc kubenswrapper[4775]: I1002 02:50:43.980892 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55a30636-4fe0-422b-958f-000c6e109f4b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:50:43 crc kubenswrapper[4775]: I1002 02:50:43.980934 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55a30636-4fe0-422b-958f-000c6e109f4b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:50:43 crc kubenswrapper[4775]: I1002 02:50:43.980947 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bh5qq\" (UniqueName: \"kubernetes.io/projected/55a30636-4fe0-422b-958f-000c6e109f4b-kube-api-access-bh5qq\") on node \"crc\" DevicePath \"\"" Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.386163 4775 generic.go:334] "Generic (PLEG): container finished" podID="55a30636-4fe0-422b-958f-000c6e109f4b" containerID="39b54c4f0d3c57986eb4822ebe0aaa3a907085e8ad4807f62e997b6c548fc579" exitCode=0 Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.386227 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jwszs" event={"ID":"55a30636-4fe0-422b-958f-000c6e109f4b","Type":"ContainerDied","Data":"39b54c4f0d3c57986eb4822ebe0aaa3a907085e8ad4807f62e997b6c548fc579"} Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.386276 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jwszs" event={"ID":"55a30636-4fe0-422b-958f-000c6e109f4b","Type":"ContainerDied","Data":"c8919963046838ddc32f487cee15503caf0d23dba55c74b780296d0650f529e8"} Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.386314 4775 scope.go:117] "RemoveContainer" containerID="39b54c4f0d3c57986eb4822ebe0aaa3a907085e8ad4807f62e997b6c548fc579" Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.386339 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jwszs" Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.417947 4775 scope.go:117] "RemoveContainer" containerID="e3c5c5b5f2e45c6e6e5f1ad375cc2f9a9c38cb6334334e4177375067d27117b2" Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.457535 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jwszs"] Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.466379 4775 scope.go:117] "RemoveContainer" containerID="b8bbf36593d755278086388108ae81f3eee541f8fb1a847c4afd970a1b4b0de6" Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.467661 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jwszs"] Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.496709 4775 scope.go:117] "RemoveContainer" containerID="39b54c4f0d3c57986eb4822ebe0aaa3a907085e8ad4807f62e997b6c548fc579" Oct 02 02:50:44 crc kubenswrapper[4775]: E1002 02:50:44.497191 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39b54c4f0d3c57986eb4822ebe0aaa3a907085e8ad4807f62e997b6c548fc579\": container with ID starting with 39b54c4f0d3c57986eb4822ebe0aaa3a907085e8ad4807f62e997b6c548fc579 not found: ID does not exist" containerID="39b54c4f0d3c57986eb4822ebe0aaa3a907085e8ad4807f62e997b6c548fc579" Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.497229 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39b54c4f0d3c57986eb4822ebe0aaa3a907085e8ad4807f62e997b6c548fc579"} err="failed to get container status \"39b54c4f0d3c57986eb4822ebe0aaa3a907085e8ad4807f62e997b6c548fc579\": rpc error: code = NotFound desc = could not find container \"39b54c4f0d3c57986eb4822ebe0aaa3a907085e8ad4807f62e997b6c548fc579\": container with ID starting with 39b54c4f0d3c57986eb4822ebe0aaa3a907085e8ad4807f62e997b6c548fc579 not found: ID does not exist" Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.497258 4775 scope.go:117] "RemoveContainer" containerID="e3c5c5b5f2e45c6e6e5f1ad375cc2f9a9c38cb6334334e4177375067d27117b2" Oct 02 02:50:44 crc kubenswrapper[4775]: E1002 02:50:44.497702 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3c5c5b5f2e45c6e6e5f1ad375cc2f9a9c38cb6334334e4177375067d27117b2\": container with ID starting with e3c5c5b5f2e45c6e6e5f1ad375cc2f9a9c38cb6334334e4177375067d27117b2 not found: ID does not exist" containerID="e3c5c5b5f2e45c6e6e5f1ad375cc2f9a9c38cb6334334e4177375067d27117b2" Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.497758 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3c5c5b5f2e45c6e6e5f1ad375cc2f9a9c38cb6334334e4177375067d27117b2"} err="failed to get container status \"e3c5c5b5f2e45c6e6e5f1ad375cc2f9a9c38cb6334334e4177375067d27117b2\": rpc error: code = NotFound desc = could not find container \"e3c5c5b5f2e45c6e6e5f1ad375cc2f9a9c38cb6334334e4177375067d27117b2\": container with ID starting with e3c5c5b5f2e45c6e6e5f1ad375cc2f9a9c38cb6334334e4177375067d27117b2 not found: ID does not exist" Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.497794 4775 scope.go:117] "RemoveContainer" containerID="b8bbf36593d755278086388108ae81f3eee541f8fb1a847c4afd970a1b4b0de6" Oct 02 02:50:44 crc kubenswrapper[4775]: E1002 02:50:44.498162 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8bbf36593d755278086388108ae81f3eee541f8fb1a847c4afd970a1b4b0de6\": container with ID starting with b8bbf36593d755278086388108ae81f3eee541f8fb1a847c4afd970a1b4b0de6 not found: ID does not exist" containerID="b8bbf36593d755278086388108ae81f3eee541f8fb1a847c4afd970a1b4b0de6" Oct 02 02:50:44 crc kubenswrapper[4775]: I1002 02:50:44.498215 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8bbf36593d755278086388108ae81f3eee541f8fb1a847c4afd970a1b4b0de6"} err="failed to get container status \"b8bbf36593d755278086388108ae81f3eee541f8fb1a847c4afd970a1b4b0de6\": rpc error: code = NotFound desc = could not find container \"b8bbf36593d755278086388108ae81f3eee541f8fb1a847c4afd970a1b4b0de6\": container with ID starting with b8bbf36593d755278086388108ae81f3eee541f8fb1a847c4afd970a1b4b0de6 not found: ID does not exist" Oct 02 02:50:45 crc kubenswrapper[4775]: I1002 02:50:45.783610 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55a30636-4fe0-422b-958f-000c6e109f4b" path="/var/lib/kubelet/pods/55a30636-4fe0-422b-958f-000c6e109f4b/volumes" Oct 02 02:51:37 crc kubenswrapper[4775]: I1002 02:51:37.234126 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:51:37 crc kubenswrapper[4775]: I1002 02:51:37.234729 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:52:07 crc kubenswrapper[4775]: I1002 02:52:07.233526 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:52:07 crc kubenswrapper[4775]: I1002 02:52:07.234571 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:52:37 crc kubenswrapper[4775]: I1002 02:52:37.234225 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 02:52:37 crc kubenswrapper[4775]: I1002 02:52:37.235018 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 02:52:37 crc kubenswrapper[4775]: I1002 02:52:37.235091 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 02:52:37 crc kubenswrapper[4775]: I1002 02:52:37.236102 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 02:52:37 crc kubenswrapper[4775]: I1002 02:52:37.236204 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" gracePeriod=600 Oct 02 02:52:37 crc kubenswrapper[4775]: E1002 02:52:37.360530 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:52:37 crc kubenswrapper[4775]: I1002 02:52:37.403726 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" exitCode=0 Oct 02 02:52:37 crc kubenswrapper[4775]: I1002 02:52:37.403861 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c"} Oct 02 02:52:37 crc kubenswrapper[4775]: I1002 02:52:37.404002 4775 scope.go:117] "RemoveContainer" containerID="701c2b59237bce746d64de4b4003ac93de1140e75c06122e0289272297fc7b80" Oct 02 02:52:37 crc kubenswrapper[4775]: I1002 02:52:37.404555 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:52:37 crc kubenswrapper[4775]: E1002 02:52:37.404915 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:52:52 crc kubenswrapper[4775]: I1002 02:52:52.766270 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:52:52 crc kubenswrapper[4775]: E1002 02:52:52.768063 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:53:05 crc kubenswrapper[4775]: I1002 02:53:05.766470 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:53:05 crc kubenswrapper[4775]: E1002 02:53:05.767332 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:53:16 crc kubenswrapper[4775]: I1002 02:53:16.765694 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:53:16 crc kubenswrapper[4775]: E1002 02:53:16.766631 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:53:31 crc kubenswrapper[4775]: I1002 02:53:31.766354 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:53:31 crc kubenswrapper[4775]: E1002 02:53:31.767373 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:53:42 crc kubenswrapper[4775]: I1002 02:53:42.765182 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:53:42 crc kubenswrapper[4775]: E1002 02:53:42.765839 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:53:53 crc kubenswrapper[4775]: I1002 02:53:53.770733 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:53:53 crc kubenswrapper[4775]: E1002 02:53:53.772003 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:54:04 crc kubenswrapper[4775]: I1002 02:54:04.766080 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:54:04 crc kubenswrapper[4775]: E1002 02:54:04.767144 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:54:16 crc kubenswrapper[4775]: I1002 02:54:16.765516 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:54:16 crc kubenswrapper[4775]: E1002 02:54:16.766471 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:54:27 crc kubenswrapper[4775]: I1002 02:54:27.766099 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:54:27 crc kubenswrapper[4775]: E1002 02:54:27.768310 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:54:39 crc kubenswrapper[4775]: I1002 02:54:39.765374 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:54:39 crc kubenswrapper[4775]: E1002 02:54:39.765890 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:54:51 crc kubenswrapper[4775]: I1002 02:54:51.765440 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:54:51 crc kubenswrapper[4775]: E1002 02:54:51.766540 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:55:03 crc kubenswrapper[4775]: I1002 02:55:03.901396 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-w7vtm"] Oct 02 02:55:03 crc kubenswrapper[4775]: I1002 02:55:03.907470 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-w7vtm"] Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.029974 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-794wf"] Oct 02 02:55:04 crc kubenswrapper[4775]: E1002 02:55:04.030254 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55a30636-4fe0-422b-958f-000c6e109f4b" containerName="registry-server" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.030267 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a30636-4fe0-422b-958f-000c6e109f4b" containerName="registry-server" Oct 02 02:55:04 crc kubenswrapper[4775]: E1002 02:55:04.030283 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55a30636-4fe0-422b-958f-000c6e109f4b" containerName="extract-utilities" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.030289 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a30636-4fe0-422b-958f-000c6e109f4b" containerName="extract-utilities" Oct 02 02:55:04 crc kubenswrapper[4775]: E1002 02:55:04.030308 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55a30636-4fe0-422b-958f-000c6e109f4b" containerName="extract-content" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.030314 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="55a30636-4fe0-422b-958f-000c6e109f4b" containerName="extract-content" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.030507 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="55a30636-4fe0-422b-958f-000c6e109f4b" containerName="registry-server" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.030943 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-794wf" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.033798 4775 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-6dh6b" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.034188 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.036015 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.036054 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.040503 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-794wf"] Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.153493 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/772b543e-1f73-4b53-891f-9e75c81fc39f-crc-storage\") pod \"crc-storage-crc-794wf\" (UID: \"772b543e-1f73-4b53-891f-9e75c81fc39f\") " pod="crc-storage/crc-storage-crc-794wf" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.153777 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-582cm\" (UniqueName: \"kubernetes.io/projected/772b543e-1f73-4b53-891f-9e75c81fc39f-kube-api-access-582cm\") pod \"crc-storage-crc-794wf\" (UID: \"772b543e-1f73-4b53-891f-9e75c81fc39f\") " pod="crc-storage/crc-storage-crc-794wf" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.154009 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/772b543e-1f73-4b53-891f-9e75c81fc39f-node-mnt\") pod \"crc-storage-crc-794wf\" (UID: \"772b543e-1f73-4b53-891f-9e75c81fc39f\") " pod="crc-storage/crc-storage-crc-794wf" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.255763 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/772b543e-1f73-4b53-891f-9e75c81fc39f-crc-storage\") pod \"crc-storage-crc-794wf\" (UID: \"772b543e-1f73-4b53-891f-9e75c81fc39f\") " pod="crc-storage/crc-storage-crc-794wf" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.255919 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-582cm\" (UniqueName: \"kubernetes.io/projected/772b543e-1f73-4b53-891f-9e75c81fc39f-kube-api-access-582cm\") pod \"crc-storage-crc-794wf\" (UID: \"772b543e-1f73-4b53-891f-9e75c81fc39f\") " pod="crc-storage/crc-storage-crc-794wf" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.256042 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/772b543e-1f73-4b53-891f-9e75c81fc39f-node-mnt\") pod \"crc-storage-crc-794wf\" (UID: \"772b543e-1f73-4b53-891f-9e75c81fc39f\") " pod="crc-storage/crc-storage-crc-794wf" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.256420 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/772b543e-1f73-4b53-891f-9e75c81fc39f-node-mnt\") pod \"crc-storage-crc-794wf\" (UID: \"772b543e-1f73-4b53-891f-9e75c81fc39f\") " pod="crc-storage/crc-storage-crc-794wf" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.256643 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/772b543e-1f73-4b53-891f-9e75c81fc39f-crc-storage\") pod \"crc-storage-crc-794wf\" (UID: \"772b543e-1f73-4b53-891f-9e75c81fc39f\") " pod="crc-storage/crc-storage-crc-794wf" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.288833 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-582cm\" (UniqueName: \"kubernetes.io/projected/772b543e-1f73-4b53-891f-9e75c81fc39f-kube-api-access-582cm\") pod \"crc-storage-crc-794wf\" (UID: \"772b543e-1f73-4b53-891f-9e75c81fc39f\") " pod="crc-storage/crc-storage-crc-794wf" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.355692 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-794wf" Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.850351 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-794wf"] Oct 02 02:55:04 crc kubenswrapper[4775]: I1002 02:55:04.857157 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 02:55:05 crc kubenswrapper[4775]: I1002 02:55:05.750136 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-794wf" event={"ID":"772b543e-1f73-4b53-891f-9e75c81fc39f","Type":"ContainerStarted","Data":"693914b1747317796134e52a369fc2ab24a4681cc5644a1182bcb07992968517"} Oct 02 02:55:05 crc kubenswrapper[4775]: I1002 02:55:05.750483 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-794wf" event={"ID":"772b543e-1f73-4b53-891f-9e75c81fc39f","Type":"ContainerStarted","Data":"c70adb89937209943cb1aeb0e66b995655d2cea882921bb8a0aba382df907a5f"} Oct 02 02:55:05 crc kubenswrapper[4775]: I1002 02:55:05.772406 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="crc-storage/crc-storage-crc-794wf" podStartSLOduration=1.148257557 podStartE2EDuration="1.772389835s" podCreationTimestamp="2025-10-02 02:55:04 +0000 UTC" firstStartedPulling="2025-10-02 02:55:04.856873677 +0000 UTC m=+4442.023617727" lastFinishedPulling="2025-10-02 02:55:05.481005955 +0000 UTC m=+4442.647750005" observedRunningTime="2025-10-02 02:55:05.768807892 +0000 UTC m=+4442.935551932" watchObservedRunningTime="2025-10-02 02:55:05.772389835 +0000 UTC m=+4442.939133875" Oct 02 02:55:05 crc kubenswrapper[4775]: I1002 02:55:05.781466 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa5168f2-58f8-4af5-85b3-c8e3b9531ffd" path="/var/lib/kubelet/pods/fa5168f2-58f8-4af5-85b3-c8e3b9531ffd/volumes" Oct 02 02:55:06 crc kubenswrapper[4775]: I1002 02:55:06.761600 4775 generic.go:334] "Generic (PLEG): container finished" podID="772b543e-1f73-4b53-891f-9e75c81fc39f" containerID="693914b1747317796134e52a369fc2ab24a4681cc5644a1182bcb07992968517" exitCode=0 Oct 02 02:55:06 crc kubenswrapper[4775]: I1002 02:55:06.761647 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-794wf" event={"ID":"772b543e-1f73-4b53-891f-9e75c81fc39f","Type":"ContainerDied","Data":"693914b1747317796134e52a369fc2ab24a4681cc5644a1182bcb07992968517"} Oct 02 02:55:06 crc kubenswrapper[4775]: I1002 02:55:06.765667 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:55:06 crc kubenswrapper[4775]: E1002 02:55:06.766186 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:55:08 crc kubenswrapper[4775]: I1002 02:55:08.087426 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-794wf" Oct 02 02:55:08 crc kubenswrapper[4775]: I1002 02:55:08.221595 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-582cm\" (UniqueName: \"kubernetes.io/projected/772b543e-1f73-4b53-891f-9e75c81fc39f-kube-api-access-582cm\") pod \"772b543e-1f73-4b53-891f-9e75c81fc39f\" (UID: \"772b543e-1f73-4b53-891f-9e75c81fc39f\") " Oct 02 02:55:08 crc kubenswrapper[4775]: I1002 02:55:08.221659 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/772b543e-1f73-4b53-891f-9e75c81fc39f-crc-storage\") pod \"772b543e-1f73-4b53-891f-9e75c81fc39f\" (UID: \"772b543e-1f73-4b53-891f-9e75c81fc39f\") " Oct 02 02:55:08 crc kubenswrapper[4775]: I1002 02:55:08.221690 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/772b543e-1f73-4b53-891f-9e75c81fc39f-node-mnt\") pod \"772b543e-1f73-4b53-891f-9e75c81fc39f\" (UID: \"772b543e-1f73-4b53-891f-9e75c81fc39f\") " Oct 02 02:55:08 crc kubenswrapper[4775]: I1002 02:55:08.221898 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/772b543e-1f73-4b53-891f-9e75c81fc39f-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "772b543e-1f73-4b53-891f-9e75c81fc39f" (UID: "772b543e-1f73-4b53-891f-9e75c81fc39f"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:55:08 crc kubenswrapper[4775]: I1002 02:55:08.222207 4775 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/772b543e-1f73-4b53-891f-9e75c81fc39f-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 02 02:55:08 crc kubenswrapper[4775]: I1002 02:55:08.227931 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/772b543e-1f73-4b53-891f-9e75c81fc39f-kube-api-access-582cm" (OuterVolumeSpecName: "kube-api-access-582cm") pod "772b543e-1f73-4b53-891f-9e75c81fc39f" (UID: "772b543e-1f73-4b53-891f-9e75c81fc39f"). InnerVolumeSpecName "kube-api-access-582cm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:55:08 crc kubenswrapper[4775]: I1002 02:55:08.246488 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/772b543e-1f73-4b53-891f-9e75c81fc39f-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "772b543e-1f73-4b53-891f-9e75c81fc39f" (UID: "772b543e-1f73-4b53-891f-9e75c81fc39f"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:55:08 crc kubenswrapper[4775]: I1002 02:55:08.323245 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-582cm\" (UniqueName: \"kubernetes.io/projected/772b543e-1f73-4b53-891f-9e75c81fc39f-kube-api-access-582cm\") on node \"crc\" DevicePath \"\"" Oct 02 02:55:08 crc kubenswrapper[4775]: I1002 02:55:08.323286 4775 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/772b543e-1f73-4b53-891f-9e75c81fc39f-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 02 02:55:08 crc kubenswrapper[4775]: I1002 02:55:08.783455 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-794wf" event={"ID":"772b543e-1f73-4b53-891f-9e75c81fc39f","Type":"ContainerDied","Data":"c70adb89937209943cb1aeb0e66b995655d2cea882921bb8a0aba382df907a5f"} Oct 02 02:55:08 crc kubenswrapper[4775]: I1002 02:55:08.783508 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c70adb89937209943cb1aeb0e66b995655d2cea882921bb8a0aba382df907a5f" Oct 02 02:55:08 crc kubenswrapper[4775]: I1002 02:55:08.783543 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-794wf" Oct 02 02:55:09 crc kubenswrapper[4775]: I1002 02:55:09.982680 4775 scope.go:117] "RemoveContainer" containerID="d7aa23990581129b590ef8697a4348d53eb75c9327128fa2f71f21e590de372a" Oct 02 02:55:09 crc kubenswrapper[4775]: I1002 02:55:09.982760 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-794wf"] Oct 02 02:55:09 crc kubenswrapper[4775]: I1002 02:55:09.993650 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-794wf"] Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.159494 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-h9922"] Oct 02 02:55:10 crc kubenswrapper[4775]: E1002 02:55:10.160044 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="772b543e-1f73-4b53-891f-9e75c81fc39f" containerName="storage" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.160078 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="772b543e-1f73-4b53-891f-9e75c81fc39f" containerName="storage" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.160359 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="772b543e-1f73-4b53-891f-9e75c81fc39f" containerName="storage" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.161146 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-h9922" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.166271 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.166940 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.170852 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-h9922"] Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.171593 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.174518 4775 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-6dh6b" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.254458 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8p7m\" (UniqueName: \"kubernetes.io/projected/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-kube-api-access-m8p7m\") pod \"crc-storage-crc-h9922\" (UID: \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\") " pod="crc-storage/crc-storage-crc-h9922" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.254619 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-node-mnt\") pod \"crc-storage-crc-h9922\" (UID: \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\") " pod="crc-storage/crc-storage-crc-h9922" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.254745 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-crc-storage\") pod \"crc-storage-crc-h9922\" (UID: \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\") " pod="crc-storage/crc-storage-crc-h9922" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.356410 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-crc-storage\") pod \"crc-storage-crc-h9922\" (UID: \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\") " pod="crc-storage/crc-storage-crc-h9922" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.356534 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8p7m\" (UniqueName: \"kubernetes.io/projected/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-kube-api-access-m8p7m\") pod \"crc-storage-crc-h9922\" (UID: \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\") " pod="crc-storage/crc-storage-crc-h9922" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.356650 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-node-mnt\") pod \"crc-storage-crc-h9922\" (UID: \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\") " pod="crc-storage/crc-storage-crc-h9922" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.356994 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-node-mnt\") pod \"crc-storage-crc-h9922\" (UID: \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\") " pod="crc-storage/crc-storage-crc-h9922" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.357878 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-crc-storage\") pod \"crc-storage-crc-h9922\" (UID: \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\") " pod="crc-storage/crc-storage-crc-h9922" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.396145 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8p7m\" (UniqueName: \"kubernetes.io/projected/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-kube-api-access-m8p7m\") pod \"crc-storage-crc-h9922\" (UID: \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\") " pod="crc-storage/crc-storage-crc-h9922" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.498282 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-h9922" Oct 02 02:55:10 crc kubenswrapper[4775]: I1002 02:55:10.996853 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-h9922"] Oct 02 02:55:11 crc kubenswrapper[4775]: I1002 02:55:11.779202 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="772b543e-1f73-4b53-891f-9e75c81fc39f" path="/var/lib/kubelet/pods/772b543e-1f73-4b53-891f-9e75c81fc39f/volumes" Oct 02 02:55:11 crc kubenswrapper[4775]: I1002 02:55:11.811557 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-h9922" event={"ID":"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2","Type":"ContainerStarted","Data":"4372509f99896adc99ce0ea9bc3eb35a652250dfc01dd1b3d1665cd80563930c"} Oct 02 02:55:12 crc kubenswrapper[4775]: I1002 02:55:12.822728 4775 generic.go:334] "Generic (PLEG): container finished" podID="da82f8e5-1e5f-4cf2-914a-f765d3aeccc2" containerID="4d8d6c429bf4dfa59be825c7f35b6baaecd6da5f5da2da1613ba2cfb2bfd161b" exitCode=0 Oct 02 02:55:12 crc kubenswrapper[4775]: I1002 02:55:12.822788 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-h9922" event={"ID":"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2","Type":"ContainerDied","Data":"4d8d6c429bf4dfa59be825c7f35b6baaecd6da5f5da2da1613ba2cfb2bfd161b"} Oct 02 02:55:14 crc kubenswrapper[4775]: I1002 02:55:14.172512 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-h9922" Oct 02 02:55:14 crc kubenswrapper[4775]: I1002 02:55:14.320796 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8p7m\" (UniqueName: \"kubernetes.io/projected/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-kube-api-access-m8p7m\") pod \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\" (UID: \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\") " Oct 02 02:55:14 crc kubenswrapper[4775]: I1002 02:55:14.321315 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-crc-storage\") pod \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\" (UID: \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\") " Oct 02 02:55:14 crc kubenswrapper[4775]: I1002 02:55:14.321616 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-node-mnt\") pod \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\" (UID: \"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2\") " Oct 02 02:55:14 crc kubenswrapper[4775]: I1002 02:55:14.321780 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "da82f8e5-1e5f-4cf2-914a-f765d3aeccc2" (UID: "da82f8e5-1e5f-4cf2-914a-f765d3aeccc2"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 02:55:14 crc kubenswrapper[4775]: I1002 02:55:14.323847 4775 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-node-mnt\") on node \"crc\" DevicePath \"\"" Oct 02 02:55:14 crc kubenswrapper[4775]: I1002 02:55:14.325839 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-kube-api-access-m8p7m" (OuterVolumeSpecName: "kube-api-access-m8p7m") pod "da82f8e5-1e5f-4cf2-914a-f765d3aeccc2" (UID: "da82f8e5-1e5f-4cf2-914a-f765d3aeccc2"). InnerVolumeSpecName "kube-api-access-m8p7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:55:14 crc kubenswrapper[4775]: I1002 02:55:14.351575 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "da82f8e5-1e5f-4cf2-914a-f765d3aeccc2" (UID: "da82f8e5-1e5f-4cf2-914a-f765d3aeccc2"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:55:14 crc kubenswrapper[4775]: I1002 02:55:14.426182 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8p7m\" (UniqueName: \"kubernetes.io/projected/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-kube-api-access-m8p7m\") on node \"crc\" DevicePath \"\"" Oct 02 02:55:14 crc kubenswrapper[4775]: I1002 02:55:14.426251 4775 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/da82f8e5-1e5f-4cf2-914a-f765d3aeccc2-crc-storage\") on node \"crc\" DevicePath \"\"" Oct 02 02:55:14 crc kubenswrapper[4775]: I1002 02:55:14.842643 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-h9922" event={"ID":"da82f8e5-1e5f-4cf2-914a-f765d3aeccc2","Type":"ContainerDied","Data":"4372509f99896adc99ce0ea9bc3eb35a652250dfc01dd1b3d1665cd80563930c"} Oct 02 02:55:14 crc kubenswrapper[4775]: I1002 02:55:14.842983 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4372509f99896adc99ce0ea9bc3eb35a652250dfc01dd1b3d1665cd80563930c" Oct 02 02:55:14 crc kubenswrapper[4775]: I1002 02:55:14.842748 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-h9922" Oct 02 02:55:20 crc kubenswrapper[4775]: I1002 02:55:20.765172 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:55:20 crc kubenswrapper[4775]: E1002 02:55:20.765639 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:55:33 crc kubenswrapper[4775]: I1002 02:55:33.772816 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:55:33 crc kubenswrapper[4775]: E1002 02:55:33.773902 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:55:44 crc kubenswrapper[4775]: I1002 02:55:44.766129 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:55:44 crc kubenswrapper[4775]: E1002 02:55:44.767024 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:55:56 crc kubenswrapper[4775]: I1002 02:55:56.765661 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:55:56 crc kubenswrapper[4775]: E1002 02:55:56.766871 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:56:09 crc kubenswrapper[4775]: I1002 02:56:09.767056 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:56:09 crc kubenswrapper[4775]: E1002 02:56:09.769704 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:56:21 crc kubenswrapper[4775]: I1002 02:56:21.765408 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:56:21 crc kubenswrapper[4775]: E1002 02:56:21.766659 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:56:36 crc kubenswrapper[4775]: I1002 02:56:36.766287 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:56:36 crc kubenswrapper[4775]: E1002 02:56:36.767264 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:56:49 crc kubenswrapper[4775]: I1002 02:56:49.767223 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:56:49 crc kubenswrapper[4775]: E1002 02:56:49.768597 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:57:03 crc kubenswrapper[4775]: I1002 02:57:03.775225 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:57:03 crc kubenswrapper[4775]: E1002 02:57:03.780732 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:57:16 crc kubenswrapper[4775]: I1002 02:57:16.765511 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:57:16 crc kubenswrapper[4775]: E1002 02:57:16.766355 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:57:31 crc kubenswrapper[4775]: I1002 02:57:31.765921 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:57:31 crc kubenswrapper[4775]: E1002 02:57:31.767181 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 02:57:46 crc kubenswrapper[4775]: I1002 02:57:46.765945 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 02:57:47 crc kubenswrapper[4775]: I1002 02:57:47.269510 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"36c9b00ccca04be89dfec3790031865696944363d02afbb5b6006911151d9160"} Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.168973 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-7vfq9"] Oct 02 02:58:21 crc kubenswrapper[4775]: E1002 02:58:21.169692 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da82f8e5-1e5f-4cf2-914a-f765d3aeccc2" containerName="storage" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.169717 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="da82f8e5-1e5f-4cf2-914a-f765d3aeccc2" containerName="storage" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.169839 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="da82f8e5-1e5f-4cf2-914a-f765d3aeccc2" containerName="storage" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.170506 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.173472 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.173570 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.181527 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-7vfq9"] Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.182091 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-7xx9q" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.182356 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.182499 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.331161 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-config\") pod \"dnsmasq-dns-5d7b5456f5-7vfq9\" (UID: \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\") " pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.331283 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-7vfq9\" (UID: \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\") " pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.331326 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2wvt\" (UniqueName: \"kubernetes.io/projected/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-kube-api-access-p2wvt\") pod \"dnsmasq-dns-5d7b5456f5-7vfq9\" (UID: \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\") " pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.420998 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-g5fwj"] Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.422113 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.432494 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-config\") pod \"dnsmasq-dns-5d7b5456f5-7vfq9\" (UID: \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\") " pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.432860 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-7vfq9\" (UID: \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\") " pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.432928 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2wvt\" (UniqueName: \"kubernetes.io/projected/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-kube-api-access-p2wvt\") pod \"dnsmasq-dns-5d7b5456f5-7vfq9\" (UID: \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\") " pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.433562 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-config\") pod \"dnsmasq-dns-5d7b5456f5-7vfq9\" (UID: \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\") " pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.434223 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-7vfq9\" (UID: \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\") " pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.435803 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-g5fwj"] Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.479911 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2wvt\" (UniqueName: \"kubernetes.io/projected/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-kube-api-access-p2wvt\") pod \"dnsmasq-dns-5d7b5456f5-7vfq9\" (UID: \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\") " pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.496571 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.534008 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-config\") pod \"dnsmasq-dns-98ddfc8f-g5fwj\" (UID: \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\") " pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.534251 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lscll\" (UniqueName: \"kubernetes.io/projected/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-kube-api-access-lscll\") pod \"dnsmasq-dns-98ddfc8f-g5fwj\" (UID: \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\") " pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.534316 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-g5fwj\" (UID: \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\") " pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.635793 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-config\") pod \"dnsmasq-dns-98ddfc8f-g5fwj\" (UID: \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\") " pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.636149 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lscll\" (UniqueName: \"kubernetes.io/projected/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-kube-api-access-lscll\") pod \"dnsmasq-dns-98ddfc8f-g5fwj\" (UID: \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\") " pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.636201 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-g5fwj\" (UID: \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\") " pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.637101 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-g5fwj\" (UID: \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\") " pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.637807 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-config\") pod \"dnsmasq-dns-98ddfc8f-g5fwj\" (UID: \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\") " pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.658136 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lscll\" (UniqueName: \"kubernetes.io/projected/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-kube-api-access-lscll\") pod \"dnsmasq-dns-98ddfc8f-g5fwj\" (UID: \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\") " pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.739071 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.932055 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-7vfq9"] Oct 02 02:58:21 crc kubenswrapper[4775]: I1002 02:58:21.962453 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-g5fwj"] Oct 02 02:58:21 crc kubenswrapper[4775]: W1002 02:58:21.964611 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb9dbb225_b6b9_4b52_92c4_574fcedb4ff6.slice/crio-356ab411980f5673a0f10ad3fa3dfadc62e9f16ece00f8e5234e9bcaa2eb3d7d WatchSource:0}: Error finding container 356ab411980f5673a0f10ad3fa3dfadc62e9f16ece00f8e5234e9bcaa2eb3d7d: Status 404 returned error can't find the container with id 356ab411980f5673a0f10ad3fa3dfadc62e9f16ece00f8e5234e9bcaa2eb3d7d Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.311044 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.312195 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.314750 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.315030 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.315261 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.315438 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.316356 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-g8wph" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.333734 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.447670 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.448078 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/99fc6db3-c1af-4901-a31f-d53db4848812-server-conf\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.448124 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngj6w\" (UniqueName: \"kubernetes.io/projected/99fc6db3-c1af-4901-a31f-d53db4848812-kube-api-access-ngj6w\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.448158 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/99fc6db3-c1af-4901-a31f-d53db4848812-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.448286 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.448429 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.448668 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/99fc6db3-c1af-4901-a31f-d53db4848812-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.448721 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/99fc6db3-c1af-4901-a31f-d53db4848812-pod-info\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.448755 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.549748 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/99fc6db3-c1af-4901-a31f-d53db4848812-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.549815 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.549849 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/99fc6db3-c1af-4901-a31f-d53db4848812-pod-info\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.549996 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.550085 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/99fc6db3-c1af-4901-a31f-d53db4848812-server-conf\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.550123 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngj6w\" (UniqueName: \"kubernetes.io/projected/99fc6db3-c1af-4901-a31f-d53db4848812-kube-api-access-ngj6w\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.550160 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/99fc6db3-c1af-4901-a31f-d53db4848812-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.550211 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.550239 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.551286 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.552385 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.553205 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/99fc6db3-c1af-4901-a31f-d53db4848812-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.553530 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/99fc6db3-c1af-4901-a31f-d53db4848812-server-conf\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.555740 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.556049 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.556113 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0e979186b38ce1a2c20c88de979de968f05e62b573b4fa37c206cee176706195/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.558284 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/99fc6db3-c1af-4901-a31f-d53db4848812-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.560125 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/99fc6db3-c1af-4901-a31f-d53db4848812-pod-info\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.590131 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngj6w\" (UniqueName: \"kubernetes.io/projected/99fc6db3-c1af-4901-a31f-d53db4848812-kube-api-access-ngj6w\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.594798 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.599177 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.604667 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-ntv97" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.605006 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.605228 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.605375 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.605425 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.612376 4775 generic.go:334] "Generic (PLEG): container finished" podID="b9dbb225-b6b9-4b52-92c4-574fcedb4ff6" containerID="5f3b845076fb6f0405bb72690422c287579fc450b5efe471886361e0f96947b5" exitCode=0 Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.612468 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" event={"ID":"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6","Type":"ContainerDied","Data":"5f3b845076fb6f0405bb72690422c287579fc450b5efe471886361e0f96947b5"} Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.612507 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" event={"ID":"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6","Type":"ContainerStarted","Data":"356ab411980f5673a0f10ad3fa3dfadc62e9f16ece00f8e5234e9bcaa2eb3d7d"} Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.618234 4775 generic.go:334] "Generic (PLEG): container finished" podID="0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef" containerID="78a3e4cc0705bb7981b53c729086628309a8020f57e5fa2e751a6dafbdc8d178" exitCode=0 Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.618319 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" event={"ID":"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef","Type":"ContainerDied","Data":"78a3e4cc0705bb7981b53c729086628309a8020f57e5fa2e751a6dafbdc8d178"} Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.621251 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" event={"ID":"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef","Type":"ContainerStarted","Data":"f78d46228c32de1a11edea3a22666ad8821663f9332a436148291e92ca1b0d2f"} Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.626585 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\") pod \"rabbitmq-server-0\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " pod="openstack/rabbitmq-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.634799 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.755892 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1a70685d-850e-4d83-b161-782b2593cbab\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a70685d-850e-4d83-b161-782b2593cbab\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.756102 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/420b6a74-82ea-4527-8fbf-693134e967ab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.756158 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s47gv\" (UniqueName: \"kubernetes.io/projected/420b6a74-82ea-4527-8fbf-693134e967ab-kube-api-access-s47gv\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.756213 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.756248 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/420b6a74-82ea-4527-8fbf-693134e967ab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.756316 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/420b6a74-82ea-4527-8fbf-693134e967ab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.756345 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/420b6a74-82ea-4527-8fbf-693134e967ab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.756397 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.756460 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.857326 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/420b6a74-82ea-4527-8fbf-693134e967ab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.857372 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s47gv\" (UniqueName: \"kubernetes.io/projected/420b6a74-82ea-4527-8fbf-693134e967ab-kube-api-access-s47gv\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.857402 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.857418 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/420b6a74-82ea-4527-8fbf-693134e967ab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.857448 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/420b6a74-82ea-4527-8fbf-693134e967ab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.857463 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/420b6a74-82ea-4527-8fbf-693134e967ab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.857486 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.857506 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.857541 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1a70685d-850e-4d83-b161-782b2593cbab\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a70685d-850e-4d83-b161-782b2593cbab\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.859259 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/420b6a74-82ea-4527-8fbf-693134e967ab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.859429 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.859682 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.859868 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/420b6a74-82ea-4527-8fbf-693134e967ab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.862310 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.862362 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1a70685d-850e-4d83-b161-782b2593cbab\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a70685d-850e-4d83-b161-782b2593cbab\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bdce4809b2d1fad6e3aff8ca06261a732fcca42e6e6fdf3a8e553419523103d5/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.862872 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.863612 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/420b6a74-82ea-4527-8fbf-693134e967ab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.865518 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/420b6a74-82ea-4527-8fbf-693134e967ab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.872541 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s47gv\" (UniqueName: \"kubernetes.io/projected/420b6a74-82ea-4527-8fbf-693134e967ab-kube-api-access-s47gv\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.892139 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1a70685d-850e-4d83-b161-782b2593cbab\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a70685d-850e-4d83-b161-782b2593cbab\") pod \"rabbitmq-cell1-server-0\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:22 crc kubenswrapper[4775]: I1002 02:58:22.929537 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 02:58:23 crc kubenswrapper[4775]: I1002 02:58:23.028495 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:58:23 crc kubenswrapper[4775]: I1002 02:58:23.483619 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:58:23 crc kubenswrapper[4775]: I1002 02:58:23.529875 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:58:23 crc kubenswrapper[4775]: W1002 02:58:23.534859 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod420b6a74_82ea_4527_8fbf_693134e967ab.slice/crio-705141c0cf2b492625a39adc3baea4ddc0f60a004f3a46edbdcfe0f7f763e211 WatchSource:0}: Error finding container 705141c0cf2b492625a39adc3baea4ddc0f60a004f3a46edbdcfe0f7f763e211: Status 404 returned error can't find the container with id 705141c0cf2b492625a39adc3baea4ddc0f60a004f3a46edbdcfe0f7f763e211 Oct 02 02:58:23 crc kubenswrapper[4775]: I1002 02:58:23.630708 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"420b6a74-82ea-4527-8fbf-693134e967ab","Type":"ContainerStarted","Data":"705141c0cf2b492625a39adc3baea4ddc0f60a004f3a46edbdcfe0f7f763e211"} Oct 02 02:58:23 crc kubenswrapper[4775]: I1002 02:58:23.633477 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" event={"ID":"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6","Type":"ContainerStarted","Data":"3f2a54ccd14db0e2fcd7b6e000fbcc776364ccce4a442cebabaac8e73f10c44a"} Oct 02 02:58:23 crc kubenswrapper[4775]: I1002 02:58:23.633565 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:58:23 crc kubenswrapper[4775]: I1002 02:58:23.635482 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"99fc6db3-c1af-4901-a31f-d53db4848812","Type":"ContainerStarted","Data":"de2df915a7ee207a294778926a8e5b83ad7b7e33bf4bf49eaf2f9ec9641d6ad3"} Oct 02 02:58:23 crc kubenswrapper[4775]: I1002 02:58:23.638335 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" event={"ID":"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef","Type":"ContainerStarted","Data":"981c9827bafae4aa8be96569396f560bd9620461e3d7176e9262cf62b47a325b"} Oct 02 02:58:23 crc kubenswrapper[4775]: I1002 02:58:23.639573 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:23 crc kubenswrapper[4775]: I1002 02:58:23.658054 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" podStartSLOduration=2.657328498 podStartE2EDuration="2.657328498s" podCreationTimestamp="2025-10-02 02:58:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:58:23.653145059 +0000 UTC m=+4640.819889109" watchObservedRunningTime="2025-10-02 02:58:23.657328498 +0000 UTC m=+4640.824072538" Oct 02 02:58:23 crc kubenswrapper[4775]: I1002 02:58:23.672820 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" podStartSLOduration=2.672802102 podStartE2EDuration="2.672802102s" podCreationTimestamp="2025-10-02 02:58:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:58:23.667945285 +0000 UTC m=+4640.834689365" watchObservedRunningTime="2025-10-02 02:58:23.672802102 +0000 UTC m=+4640.839546132" Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.043693 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.044851 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.046657 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.047593 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-s6bg8" Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.066894 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.183135 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5wvz\" (UniqueName: \"kubernetes.io/projected/e7850678-5334-43e5-9262-d1442e4959b1-kube-api-access-b5wvz\") pod \"memcached-0\" (UID: \"e7850678-5334-43e5-9262-d1442e4959b1\") " pod="openstack/memcached-0" Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.183336 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e7850678-5334-43e5-9262-d1442e4959b1-kolla-config\") pod \"memcached-0\" (UID: \"e7850678-5334-43e5-9262-d1442e4959b1\") " pod="openstack/memcached-0" Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.183460 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7850678-5334-43e5-9262-d1442e4959b1-config-data\") pod \"memcached-0\" (UID: \"e7850678-5334-43e5-9262-d1442e4959b1\") " pod="openstack/memcached-0" Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.284811 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7850678-5334-43e5-9262-d1442e4959b1-config-data\") pod \"memcached-0\" (UID: \"e7850678-5334-43e5-9262-d1442e4959b1\") " pod="openstack/memcached-0" Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.284907 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5wvz\" (UniqueName: \"kubernetes.io/projected/e7850678-5334-43e5-9262-d1442e4959b1-kube-api-access-b5wvz\") pod \"memcached-0\" (UID: \"e7850678-5334-43e5-9262-d1442e4959b1\") " pod="openstack/memcached-0" Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.284957 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e7850678-5334-43e5-9262-d1442e4959b1-kolla-config\") pod \"memcached-0\" (UID: \"e7850678-5334-43e5-9262-d1442e4959b1\") " pod="openstack/memcached-0" Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.285686 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e7850678-5334-43e5-9262-d1442e4959b1-kolla-config\") pod \"memcached-0\" (UID: \"e7850678-5334-43e5-9262-d1442e4959b1\") " pod="openstack/memcached-0" Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.286231 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e7850678-5334-43e5-9262-d1442e4959b1-config-data\") pod \"memcached-0\" (UID: \"e7850678-5334-43e5-9262-d1442e4959b1\") " pod="openstack/memcached-0" Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.383284 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5wvz\" (UniqueName: \"kubernetes.io/projected/e7850678-5334-43e5-9262-d1442e4959b1-kube-api-access-b5wvz\") pod \"memcached-0\" (UID: \"e7850678-5334-43e5-9262-d1442e4959b1\") " pod="openstack/memcached-0" Oct 02 02:58:24 crc kubenswrapper[4775]: I1002 02:58:24.659023 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.382658 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.384316 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.386791 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.386798 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.391517 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.391692 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-djl8p" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.392221 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.395712 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.397808 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.434506 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.443064 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.445804 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.446574 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.447319 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-j8rq8" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.447710 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.462119 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.506190 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-kolla-config\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.506532 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/63eca845-367d-47d2-a693-60b6b52252d3-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.506565 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/63eca845-367d-47d2-a693-60b6b52252d3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.506585 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.506609 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2a4a6381-8bbe-43b6-b120-ddc9d4c0c221\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2a4a6381-8bbe-43b6-b120-ddc9d4c0c221\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.506700 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.507411 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-700db44b-c993-4dc6-80ff-b28149777b47\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-700db44b-c993-4dc6-80ff-b28149777b47\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.507535 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63eca845-367d-47d2-a693-60b6b52252d3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.507653 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-secrets\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.507704 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-config-data-default\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.507747 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.507812 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/63eca845-367d-47d2-a693-60b6b52252d3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.507842 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63eca845-367d-47d2-a693-60b6b52252d3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.507903 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/63eca845-367d-47d2-a693-60b6b52252d3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.508010 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zqtm\" (UniqueName: \"kubernetes.io/projected/63eca845-367d-47d2-a693-60b6b52252d3-kube-api-access-9zqtm\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.508034 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/63eca845-367d-47d2-a693-60b6b52252d3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.508187 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-474ln\" (UniqueName: \"kubernetes.io/projected/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-kube-api-access-474ln\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.508206 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.577455 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 02 02:58:25 crc kubenswrapper[4775]: W1002 02:58:25.582073 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7850678_5334_43e5_9262_d1442e4959b1.slice/crio-809fe2bb564f40c8483984c67f4b2971d15be294459c2a44b6d3ea96ce9f7abf WatchSource:0}: Error finding container 809fe2bb564f40c8483984c67f4b2971d15be294459c2a44b6d3ea96ce9f7abf: Status 404 returned error can't find the container with id 809fe2bb564f40c8483984c67f4b2971d15be294459c2a44b6d3ea96ce9f7abf Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.609640 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zqtm\" (UniqueName: \"kubernetes.io/projected/63eca845-367d-47d2-a693-60b6b52252d3-kube-api-access-9zqtm\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.609719 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/63eca845-367d-47d2-a693-60b6b52252d3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.610779 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-474ln\" (UniqueName: \"kubernetes.io/projected/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-kube-api-access-474ln\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.610811 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.610867 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-kolla-config\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.610896 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/63eca845-367d-47d2-a693-60b6b52252d3-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.610922 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/63eca845-367d-47d2-a693-60b6b52252d3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.610981 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.611015 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2a4a6381-8bbe-43b6-b120-ddc9d4c0c221\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2a4a6381-8bbe-43b6-b120-ddc9d4c0c221\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.611050 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.611075 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-700db44b-c993-4dc6-80ff-b28149777b47\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-700db44b-c993-4dc6-80ff-b28149777b47\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.611112 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63eca845-367d-47d2-a693-60b6b52252d3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.611137 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-secrets\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.611162 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-config-data-default\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.611213 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.611246 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/63eca845-367d-47d2-a693-60b6b52252d3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.611270 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63eca845-367d-47d2-a693-60b6b52252d3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.611328 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/63eca845-367d-47d2-a693-60b6b52252d3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.611497 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.612187 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/63eca845-367d-47d2-a693-60b6b52252d3-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.612472 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/63eca845-367d-47d2-a693-60b6b52252d3-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.612620 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-kolla-config\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.613160 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/63eca845-367d-47d2-a693-60b6b52252d3-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.614222 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-config-data-default\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.614728 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.616365 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/63eca845-367d-47d2-a693-60b6b52252d3-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.617870 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.617903 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2a4a6381-8bbe-43b6-b120-ddc9d4c0c221\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2a4a6381-8bbe-43b6-b120-ddc9d4c0c221\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/efeff42a1344acc514270eed07885511058fa039c58fff7199ca1baf6c008b56/globalmount\"" pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.618313 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.618352 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-700db44b-c993-4dc6-80ff-b28149777b47\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-700db44b-c993-4dc6-80ff-b28149777b47\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0d4f283a96939d7f697ca3a441e78946454fa2587d06c490ec99dd5e413e0c8c/globalmount\"" pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.618495 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.619084 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-secrets\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.620448 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.620501 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/63eca845-367d-47d2-a693-60b6b52252d3-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.625571 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/63eca845-367d-47d2-a693-60b6b52252d3-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.627704 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/63eca845-367d-47d2-a693-60b6b52252d3-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.627854 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-474ln\" (UniqueName: \"kubernetes.io/projected/f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608-kube-api-access-474ln\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.630769 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zqtm\" (UniqueName: \"kubernetes.io/projected/63eca845-367d-47d2-a693-60b6b52252d3-kube-api-access-9zqtm\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.649342 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-700db44b-c993-4dc6-80ff-b28149777b47\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-700db44b-c993-4dc6-80ff-b28149777b47\") pod \"openstack-galera-0\" (UID: \"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608\") " pod="openstack/openstack-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.655085 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2a4a6381-8bbe-43b6-b120-ddc9d4c0c221\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2a4a6381-8bbe-43b6-b120-ddc9d4c0c221\") pod \"openstack-cell1-galera-0\" (UID: \"63eca845-367d-47d2-a693-60b6b52252d3\") " pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.657358 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e7850678-5334-43e5-9262-d1442e4959b1","Type":"ContainerStarted","Data":"809fe2bb564f40c8483984c67f4b2971d15be294459c2a44b6d3ea96ce9f7abf"} Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.660549 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"99fc6db3-c1af-4901-a31f-d53db4848812","Type":"ContainerStarted","Data":"cc2d12bb162e34b77ebb9006e65d254bcda59cb3d50eeb1685aee09c5b728139"} Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.666868 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"420b6a74-82ea-4527-8fbf-693134e967ab","Type":"ContainerStarted","Data":"2d14d3219f2f7c57493d64680a568054547aa6b9df059c06f9b2f67a5d92295c"} Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.714113 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:25 crc kubenswrapper[4775]: I1002 02:58:25.768553 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 02 02:58:26 crc kubenswrapper[4775]: W1002 02:58:26.209894 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63eca845_367d_47d2_a693_60b6b52252d3.slice/crio-19c07a0656369cded5a30a93c4a86d02b39ee069e3c00ea4d40b42fbe86e2972 WatchSource:0}: Error finding container 19c07a0656369cded5a30a93c4a86d02b39ee069e3c00ea4d40b42fbe86e2972: Status 404 returned error can't find the container with id 19c07a0656369cded5a30a93c4a86d02b39ee069e3c00ea4d40b42fbe86e2972 Oct 02 02:58:26 crc kubenswrapper[4775]: I1002 02:58:26.211628 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 02 02:58:26 crc kubenswrapper[4775]: I1002 02:58:26.262911 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 02 02:58:26 crc kubenswrapper[4775]: W1002 02:58:26.276396 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf87e89d2_f8ad_4ce8_bbc3_13a3bc1d6608.slice/crio-ec2211d27923757f2c2f630e9aaca2ed2204efa340cdcc482a1bed550817416d WatchSource:0}: Error finding container ec2211d27923757f2c2f630e9aaca2ed2204efa340cdcc482a1bed550817416d: Status 404 returned error can't find the container with id ec2211d27923757f2c2f630e9aaca2ed2204efa340cdcc482a1bed550817416d Oct 02 02:58:26 crc kubenswrapper[4775]: I1002 02:58:26.683786 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"e7850678-5334-43e5-9262-d1442e4959b1","Type":"ContainerStarted","Data":"2ffec3552931d7988ef088bcf856b691442d6522910c45bd135e519bd7f50275"} Oct 02 02:58:26 crc kubenswrapper[4775]: I1002 02:58:26.685577 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 02 02:58:26 crc kubenswrapper[4775]: I1002 02:58:26.689557 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"63eca845-367d-47d2-a693-60b6b52252d3","Type":"ContainerStarted","Data":"cbf2ab1582cd4a6a907cc6c482c40d78f891344cdea56ad7cce3e50e207017c2"} Oct 02 02:58:26 crc kubenswrapper[4775]: I1002 02:58:26.689633 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"63eca845-367d-47d2-a693-60b6b52252d3","Type":"ContainerStarted","Data":"19c07a0656369cded5a30a93c4a86d02b39ee069e3c00ea4d40b42fbe86e2972"} Oct 02 02:58:26 crc kubenswrapper[4775]: I1002 02:58:26.693809 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608","Type":"ContainerStarted","Data":"f49ab3e61d2218717266367c26c8f64f7ab3de6c985f6e6853803416a2b0a909"} Oct 02 02:58:26 crc kubenswrapper[4775]: I1002 02:58:26.694101 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608","Type":"ContainerStarted","Data":"ec2211d27923757f2c2f630e9aaca2ed2204efa340cdcc482a1bed550817416d"} Oct 02 02:58:26 crc kubenswrapper[4775]: I1002 02:58:26.717576 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.717537604 podStartE2EDuration="2.717537604s" podCreationTimestamp="2025-10-02 02:58:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:58:26.707708288 +0000 UTC m=+4643.874452398" watchObservedRunningTime="2025-10-02 02:58:26.717537604 +0000 UTC m=+4643.884281684" Oct 02 02:58:30 crc kubenswrapper[4775]: I1002 02:58:30.735382 4775 generic.go:334] "Generic (PLEG): container finished" podID="f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608" containerID="f49ab3e61d2218717266367c26c8f64f7ab3de6c985f6e6853803416a2b0a909" exitCode=0 Oct 02 02:58:30 crc kubenswrapper[4775]: I1002 02:58:30.735497 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608","Type":"ContainerDied","Data":"f49ab3e61d2218717266367c26c8f64f7ab3de6c985f6e6853803416a2b0a909"} Oct 02 02:58:30 crc kubenswrapper[4775]: I1002 02:58:30.741181 4775 generic.go:334] "Generic (PLEG): container finished" podID="63eca845-367d-47d2-a693-60b6b52252d3" containerID="cbf2ab1582cd4a6a907cc6c482c40d78f891344cdea56ad7cce3e50e207017c2" exitCode=0 Oct 02 02:58:30 crc kubenswrapper[4775]: I1002 02:58:30.741287 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"63eca845-367d-47d2-a693-60b6b52252d3","Type":"ContainerDied","Data":"cbf2ab1582cd4a6a907cc6c482c40d78f891344cdea56ad7cce3e50e207017c2"} Oct 02 02:58:31 crc kubenswrapper[4775]: I1002 02:58:31.499245 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:31 crc kubenswrapper[4775]: I1002 02:58:31.740817 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:58:31 crc kubenswrapper[4775]: I1002 02:58:31.749369 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"63eca845-367d-47d2-a693-60b6b52252d3","Type":"ContainerStarted","Data":"3aaa3e7dd3b7f33c57e17ea4cf40886583596fddf40acabbfcbc13c7bbacb5e2"} Oct 02 02:58:31 crc kubenswrapper[4775]: I1002 02:58:31.751766 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608","Type":"ContainerStarted","Data":"c320d90f33b812c2b6d864e97021a24195d9f7f3c47e4364399d5e847cb713bb"} Oct 02 02:58:31 crc kubenswrapper[4775]: I1002 02:58:31.785157 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-7vfq9"] Oct 02 02:58:31 crc kubenswrapper[4775]: I1002 02:58:31.785447 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" podUID="0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef" containerName="dnsmasq-dns" containerID="cri-o://981c9827bafae4aa8be96569396f560bd9620461e3d7176e9262cf62b47a325b" gracePeriod=10 Oct 02 02:58:31 crc kubenswrapper[4775]: I1002 02:58:31.794847 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=7.794832409 podStartE2EDuration="7.794832409s" podCreationTimestamp="2025-10-02 02:58:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:58:31.79410035 +0000 UTC m=+4648.960844390" watchObservedRunningTime="2025-10-02 02:58:31.794832409 +0000 UTC m=+4648.961576449" Oct 02 02:58:31 crc kubenswrapper[4775]: I1002 02:58:31.826558 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.826534256 podStartE2EDuration="7.826534256s" podCreationTimestamp="2025-10-02 02:58:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:58:31.821150485 +0000 UTC m=+4648.987894525" watchObservedRunningTime="2025-10-02 02:58:31.826534256 +0000 UTC m=+4648.993278316" Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.258385 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.331093 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2wvt\" (UniqueName: \"kubernetes.io/projected/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-kube-api-access-p2wvt\") pod \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\" (UID: \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\") " Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.331466 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-dns-svc\") pod \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\" (UID: \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\") " Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.331566 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-config\") pod \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\" (UID: \"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef\") " Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.335588 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-kube-api-access-p2wvt" (OuterVolumeSpecName: "kube-api-access-p2wvt") pod "0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef" (UID: "0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef"). InnerVolumeSpecName "kube-api-access-p2wvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.365031 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-config" (OuterVolumeSpecName: "config") pod "0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef" (UID: "0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.371010 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef" (UID: "0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.433528 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.433577 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.433589 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2wvt\" (UniqueName: \"kubernetes.io/projected/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef-kube-api-access-p2wvt\") on node \"crc\" DevicePath \"\"" Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.762445 4775 generic.go:334] "Generic (PLEG): container finished" podID="0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef" containerID="981c9827bafae4aa8be96569396f560bd9620461e3d7176e9262cf62b47a325b" exitCode=0 Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.762504 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" event={"ID":"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef","Type":"ContainerDied","Data":"981c9827bafae4aa8be96569396f560bd9620461e3d7176e9262cf62b47a325b"} Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.762544 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" event={"ID":"0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef","Type":"ContainerDied","Data":"f78d46228c32de1a11edea3a22666ad8821663f9332a436148291e92ca1b0d2f"} Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.762574 4775 scope.go:117] "RemoveContainer" containerID="981c9827bafae4aa8be96569396f560bd9620461e3d7176e9262cf62b47a325b" Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.762914 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-7vfq9" Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.799114 4775 scope.go:117] "RemoveContainer" containerID="78a3e4cc0705bb7981b53c729086628309a8020f57e5fa2e751a6dafbdc8d178" Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.814301 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-7vfq9"] Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.845838 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-7vfq9"] Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.858976 4775 scope.go:117] "RemoveContainer" containerID="981c9827bafae4aa8be96569396f560bd9620461e3d7176e9262cf62b47a325b" Oct 02 02:58:32 crc kubenswrapper[4775]: E1002 02:58:32.859612 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"981c9827bafae4aa8be96569396f560bd9620461e3d7176e9262cf62b47a325b\": container with ID starting with 981c9827bafae4aa8be96569396f560bd9620461e3d7176e9262cf62b47a325b not found: ID does not exist" containerID="981c9827bafae4aa8be96569396f560bd9620461e3d7176e9262cf62b47a325b" Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.859663 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"981c9827bafae4aa8be96569396f560bd9620461e3d7176e9262cf62b47a325b"} err="failed to get container status \"981c9827bafae4aa8be96569396f560bd9620461e3d7176e9262cf62b47a325b\": rpc error: code = NotFound desc = could not find container \"981c9827bafae4aa8be96569396f560bd9620461e3d7176e9262cf62b47a325b\": container with ID starting with 981c9827bafae4aa8be96569396f560bd9620461e3d7176e9262cf62b47a325b not found: ID does not exist" Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.859697 4775 scope.go:117] "RemoveContainer" containerID="78a3e4cc0705bb7981b53c729086628309a8020f57e5fa2e751a6dafbdc8d178" Oct 02 02:58:32 crc kubenswrapper[4775]: E1002 02:58:32.860455 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78a3e4cc0705bb7981b53c729086628309a8020f57e5fa2e751a6dafbdc8d178\": container with ID starting with 78a3e4cc0705bb7981b53c729086628309a8020f57e5fa2e751a6dafbdc8d178 not found: ID does not exist" containerID="78a3e4cc0705bb7981b53c729086628309a8020f57e5fa2e751a6dafbdc8d178" Oct 02 02:58:32 crc kubenswrapper[4775]: I1002 02:58:32.860492 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78a3e4cc0705bb7981b53c729086628309a8020f57e5fa2e751a6dafbdc8d178"} err="failed to get container status \"78a3e4cc0705bb7981b53c729086628309a8020f57e5fa2e751a6dafbdc8d178\": rpc error: code = NotFound desc = could not find container \"78a3e4cc0705bb7981b53c729086628309a8020f57e5fa2e751a6dafbdc8d178\": container with ID starting with 78a3e4cc0705bb7981b53c729086628309a8020f57e5fa2e751a6dafbdc8d178 not found: ID does not exist" Oct 02 02:58:33 crc kubenswrapper[4775]: I1002 02:58:33.780383 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef" path="/var/lib/kubelet/pods/0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef/volumes" Oct 02 02:58:34 crc kubenswrapper[4775]: I1002 02:58:34.661256 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 02 02:58:35 crc kubenswrapper[4775]: I1002 02:58:35.714571 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:35 crc kubenswrapper[4775]: I1002 02:58:35.714681 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:35 crc kubenswrapper[4775]: I1002 02:58:35.784514 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 02 02:58:35 crc kubenswrapper[4775]: I1002 02:58:35.784648 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:35 crc kubenswrapper[4775]: I1002 02:58:35.784665 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 02 02:58:35 crc kubenswrapper[4775]: I1002 02:58:35.843877 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 02 02:58:37 crc kubenswrapper[4775]: I1002 02:58:37.826582 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 02 02:58:37 crc kubenswrapper[4775]: I1002 02:58:37.892842 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 02 02:58:58 crc kubenswrapper[4775]: I1002 02:58:58.013772 4775 generic.go:334] "Generic (PLEG): container finished" podID="99fc6db3-c1af-4901-a31f-d53db4848812" containerID="cc2d12bb162e34b77ebb9006e65d254bcda59cb3d50eeb1685aee09c5b728139" exitCode=0 Oct 02 02:58:58 crc kubenswrapper[4775]: I1002 02:58:58.013836 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"99fc6db3-c1af-4901-a31f-d53db4848812","Type":"ContainerDied","Data":"cc2d12bb162e34b77ebb9006e65d254bcda59cb3d50eeb1685aee09c5b728139"} Oct 02 02:58:59 crc kubenswrapper[4775]: I1002 02:58:59.028289 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"99fc6db3-c1af-4901-a31f-d53db4848812","Type":"ContainerStarted","Data":"1e2fb2ef0f0422cba22de4086de226d27399f67395284849c554b529cc58338e"} Oct 02 02:58:59 crc kubenswrapper[4775]: I1002 02:58:59.029236 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 02:58:59 crc kubenswrapper[4775]: I1002 02:58:59.031112 4775 generic.go:334] "Generic (PLEG): container finished" podID="420b6a74-82ea-4527-8fbf-693134e967ab" containerID="2d14d3219f2f7c57493d64680a568054547aa6b9df059c06f9b2f67a5d92295c" exitCode=0 Oct 02 02:58:59 crc kubenswrapper[4775]: I1002 02:58:59.031371 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"420b6a74-82ea-4527-8fbf-693134e967ab","Type":"ContainerDied","Data":"2d14d3219f2f7c57493d64680a568054547aa6b9df059c06f9b2f67a5d92295c"} Oct 02 02:58:59 crc kubenswrapper[4775]: I1002 02:58:59.094946 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.094906714 podStartE2EDuration="38.094906714s" podCreationTimestamp="2025-10-02 02:58:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:58:59.084085362 +0000 UTC m=+4676.250829432" watchObservedRunningTime="2025-10-02 02:58:59.094906714 +0000 UTC m=+4676.261650804" Oct 02 02:59:00 crc kubenswrapper[4775]: I1002 02:59:00.042896 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"420b6a74-82ea-4527-8fbf-693134e967ab","Type":"ContainerStarted","Data":"0f3d565e93f6e9320d5737fe7ad35765dde06839dd9fd60e104e75a7a3889596"} Oct 02 02:59:00 crc kubenswrapper[4775]: I1002 02:59:00.075562 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=39.07553666 podStartE2EDuration="39.07553666s" podCreationTimestamp="2025-10-02 02:58:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:59:00.067043199 +0000 UTC m=+4677.233787309" watchObservedRunningTime="2025-10-02 02:59:00.07553666 +0000 UTC m=+4677.242280740" Oct 02 02:59:03 crc kubenswrapper[4775]: I1002 02:59:03.029160 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.415021 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8twmf"] Oct 02 02:59:04 crc kubenswrapper[4775]: E1002 02:59:04.415357 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef" containerName="dnsmasq-dns" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.415371 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef" containerName="dnsmasq-dns" Oct 02 02:59:04 crc kubenswrapper[4775]: E1002 02:59:04.415387 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef" containerName="init" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.415395 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef" containerName="init" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.415586 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d31d2b2-f4b7-48e3-89b9-4addfd9b80ef" containerName="dnsmasq-dns" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.416895 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.435212 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8twmf"] Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.557205 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh9kv\" (UniqueName: \"kubernetes.io/projected/79bd73b6-a318-44be-b594-f7612c922130-kube-api-access-kh9kv\") pod \"redhat-marketplace-8twmf\" (UID: \"79bd73b6-a318-44be-b594-f7612c922130\") " pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.557324 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79bd73b6-a318-44be-b594-f7612c922130-catalog-content\") pod \"redhat-marketplace-8twmf\" (UID: \"79bd73b6-a318-44be-b594-f7612c922130\") " pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.557370 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79bd73b6-a318-44be-b594-f7612c922130-utilities\") pod \"redhat-marketplace-8twmf\" (UID: \"79bd73b6-a318-44be-b594-f7612c922130\") " pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.658424 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh9kv\" (UniqueName: \"kubernetes.io/projected/79bd73b6-a318-44be-b594-f7612c922130-kube-api-access-kh9kv\") pod \"redhat-marketplace-8twmf\" (UID: \"79bd73b6-a318-44be-b594-f7612c922130\") " pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.658540 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79bd73b6-a318-44be-b594-f7612c922130-catalog-content\") pod \"redhat-marketplace-8twmf\" (UID: \"79bd73b6-a318-44be-b594-f7612c922130\") " pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.658604 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79bd73b6-a318-44be-b594-f7612c922130-utilities\") pod \"redhat-marketplace-8twmf\" (UID: \"79bd73b6-a318-44be-b594-f7612c922130\") " pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.659075 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79bd73b6-a318-44be-b594-f7612c922130-catalog-content\") pod \"redhat-marketplace-8twmf\" (UID: \"79bd73b6-a318-44be-b594-f7612c922130\") " pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.659320 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79bd73b6-a318-44be-b594-f7612c922130-utilities\") pod \"redhat-marketplace-8twmf\" (UID: \"79bd73b6-a318-44be-b594-f7612c922130\") " pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.687785 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh9kv\" (UniqueName: \"kubernetes.io/projected/79bd73b6-a318-44be-b594-f7612c922130-kube-api-access-kh9kv\") pod \"redhat-marketplace-8twmf\" (UID: \"79bd73b6-a318-44be-b594-f7612c922130\") " pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:04 crc kubenswrapper[4775]: I1002 02:59:04.736428 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:05 crc kubenswrapper[4775]: I1002 02:59:05.036379 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8twmf"] Oct 02 02:59:05 crc kubenswrapper[4775]: W1002 02:59:05.039696 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod79bd73b6_a318_44be_b594_f7612c922130.slice/crio-fb667ce6eaab116246893e805417238abb7dd73f3c220b2e2a5368e3fba06a7f WatchSource:0}: Error finding container fb667ce6eaab116246893e805417238abb7dd73f3c220b2e2a5368e3fba06a7f: Status 404 returned error can't find the container with id fb667ce6eaab116246893e805417238abb7dd73f3c220b2e2a5368e3fba06a7f Oct 02 02:59:05 crc kubenswrapper[4775]: I1002 02:59:05.084024 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8twmf" event={"ID":"79bd73b6-a318-44be-b594-f7612c922130","Type":"ContainerStarted","Data":"fb667ce6eaab116246893e805417238abb7dd73f3c220b2e2a5368e3fba06a7f"} Oct 02 02:59:06 crc kubenswrapper[4775]: I1002 02:59:06.096221 4775 generic.go:334] "Generic (PLEG): container finished" podID="79bd73b6-a318-44be-b594-f7612c922130" containerID="9c4584d726023652de3086a8c954f3a05493b08cc85cabd3a804edcaa3070a26" exitCode=0 Oct 02 02:59:06 crc kubenswrapper[4775]: I1002 02:59:06.096278 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8twmf" event={"ID":"79bd73b6-a318-44be-b594-f7612c922130","Type":"ContainerDied","Data":"9c4584d726023652de3086a8c954f3a05493b08cc85cabd3a804edcaa3070a26"} Oct 02 02:59:08 crc kubenswrapper[4775]: I1002 02:59:08.115128 4775 generic.go:334] "Generic (PLEG): container finished" podID="79bd73b6-a318-44be-b594-f7612c922130" containerID="c86f8502415644d56b4e2accd32a20398aab4a66475b87f1fad3a65e144c004c" exitCode=0 Oct 02 02:59:08 crc kubenswrapper[4775]: I1002 02:59:08.115230 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8twmf" event={"ID":"79bd73b6-a318-44be-b594-f7612c922130","Type":"ContainerDied","Data":"c86f8502415644d56b4e2accd32a20398aab4a66475b87f1fad3a65e144c004c"} Oct 02 02:59:10 crc kubenswrapper[4775]: I1002 02:59:10.137950 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8twmf" event={"ID":"79bd73b6-a318-44be-b594-f7612c922130","Type":"ContainerStarted","Data":"75ce6b00ef0a18213c69edfd9e881773154610536800feadd36f3891cb3d1208"} Oct 02 02:59:10 crc kubenswrapper[4775]: I1002 02:59:10.163635 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8twmf" podStartSLOduration=3.508963168 podStartE2EDuration="6.163617416s" podCreationTimestamp="2025-10-02 02:59:04 +0000 UTC" firstStartedPulling="2025-10-02 02:59:06.097893776 +0000 UTC m=+4683.264637826" lastFinishedPulling="2025-10-02 02:59:08.752547994 +0000 UTC m=+4685.919292074" observedRunningTime="2025-10-02 02:59:10.161574073 +0000 UTC m=+4687.328318153" watchObservedRunningTime="2025-10-02 02:59:10.163617416 +0000 UTC m=+4687.330361466" Oct 02 02:59:12 crc kubenswrapper[4775]: I1002 02:59:12.933288 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 02:59:13 crc kubenswrapper[4775]: I1002 02:59:13.031282 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:14 crc kubenswrapper[4775]: I1002 02:59:14.737392 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:14 crc kubenswrapper[4775]: I1002 02:59:14.737439 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:14 crc kubenswrapper[4775]: I1002 02:59:14.818827 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:15 crc kubenswrapper[4775]: I1002 02:59:15.248510 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:15 crc kubenswrapper[4775]: I1002 02:59:15.311258 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8twmf"] Oct 02 02:59:17 crc kubenswrapper[4775]: I1002 02:59:17.230263 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8twmf" podUID="79bd73b6-a318-44be-b594-f7612c922130" containerName="registry-server" containerID="cri-o://75ce6b00ef0a18213c69edfd9e881773154610536800feadd36f3891cb3d1208" gracePeriod=2 Oct 02 02:59:17 crc kubenswrapper[4775]: I1002 02:59:17.655505 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:17 crc kubenswrapper[4775]: I1002 02:59:17.829517 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kh9kv\" (UniqueName: \"kubernetes.io/projected/79bd73b6-a318-44be-b594-f7612c922130-kube-api-access-kh9kv\") pod \"79bd73b6-a318-44be-b594-f7612c922130\" (UID: \"79bd73b6-a318-44be-b594-f7612c922130\") " Oct 02 02:59:17 crc kubenswrapper[4775]: I1002 02:59:17.829626 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79bd73b6-a318-44be-b594-f7612c922130-catalog-content\") pod \"79bd73b6-a318-44be-b594-f7612c922130\" (UID: \"79bd73b6-a318-44be-b594-f7612c922130\") " Oct 02 02:59:17 crc kubenswrapper[4775]: I1002 02:59:17.831005 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79bd73b6-a318-44be-b594-f7612c922130-utilities\") pod \"79bd73b6-a318-44be-b594-f7612c922130\" (UID: \"79bd73b6-a318-44be-b594-f7612c922130\") " Oct 02 02:59:17 crc kubenswrapper[4775]: I1002 02:59:17.831947 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79bd73b6-a318-44be-b594-f7612c922130-utilities" (OuterVolumeSpecName: "utilities") pod "79bd73b6-a318-44be-b594-f7612c922130" (UID: "79bd73b6-a318-44be-b594-f7612c922130"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:59:17 crc kubenswrapper[4775]: I1002 02:59:17.842226 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79bd73b6-a318-44be-b594-f7612c922130-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79bd73b6-a318-44be-b594-f7612c922130" (UID: "79bd73b6-a318-44be-b594-f7612c922130"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:59:17 crc kubenswrapper[4775]: I1002 02:59:17.843139 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79bd73b6-a318-44be-b594-f7612c922130-kube-api-access-kh9kv" (OuterVolumeSpecName: "kube-api-access-kh9kv") pod "79bd73b6-a318-44be-b594-f7612c922130" (UID: "79bd73b6-a318-44be-b594-f7612c922130"). InnerVolumeSpecName "kube-api-access-kh9kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:59:17 crc kubenswrapper[4775]: I1002 02:59:17.932811 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79bd73b6-a318-44be-b594-f7612c922130-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:17 crc kubenswrapper[4775]: I1002 02:59:17.932849 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kh9kv\" (UniqueName: \"kubernetes.io/projected/79bd73b6-a318-44be-b594-f7612c922130-kube-api-access-kh9kv\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:17 crc kubenswrapper[4775]: I1002 02:59:17.932867 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79bd73b6-a318-44be-b594-f7612c922130-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.243601 4775 generic.go:334] "Generic (PLEG): container finished" podID="79bd73b6-a318-44be-b594-f7612c922130" containerID="75ce6b00ef0a18213c69edfd9e881773154610536800feadd36f3891cb3d1208" exitCode=0 Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.243647 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8twmf" event={"ID":"79bd73b6-a318-44be-b594-f7612c922130","Type":"ContainerDied","Data":"75ce6b00ef0a18213c69edfd9e881773154610536800feadd36f3891cb3d1208"} Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.243679 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8twmf" event={"ID":"79bd73b6-a318-44be-b594-f7612c922130","Type":"ContainerDied","Data":"fb667ce6eaab116246893e805417238abb7dd73f3c220b2e2a5368e3fba06a7f"} Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.243701 4775 scope.go:117] "RemoveContainer" containerID="75ce6b00ef0a18213c69edfd9e881773154610536800feadd36f3891cb3d1208" Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.243728 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8twmf" Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.279339 4775 scope.go:117] "RemoveContainer" containerID="c86f8502415644d56b4e2accd32a20398aab4a66475b87f1fad3a65e144c004c" Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.290095 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8twmf"] Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.305361 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8twmf"] Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.322593 4775 scope.go:117] "RemoveContainer" containerID="9c4584d726023652de3086a8c954f3a05493b08cc85cabd3a804edcaa3070a26" Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.352458 4775 scope.go:117] "RemoveContainer" containerID="75ce6b00ef0a18213c69edfd9e881773154610536800feadd36f3891cb3d1208" Oct 02 02:59:18 crc kubenswrapper[4775]: E1002 02:59:18.353139 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75ce6b00ef0a18213c69edfd9e881773154610536800feadd36f3891cb3d1208\": container with ID starting with 75ce6b00ef0a18213c69edfd9e881773154610536800feadd36f3891cb3d1208 not found: ID does not exist" containerID="75ce6b00ef0a18213c69edfd9e881773154610536800feadd36f3891cb3d1208" Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.353228 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75ce6b00ef0a18213c69edfd9e881773154610536800feadd36f3891cb3d1208"} err="failed to get container status \"75ce6b00ef0a18213c69edfd9e881773154610536800feadd36f3891cb3d1208\": rpc error: code = NotFound desc = could not find container \"75ce6b00ef0a18213c69edfd9e881773154610536800feadd36f3891cb3d1208\": container with ID starting with 75ce6b00ef0a18213c69edfd9e881773154610536800feadd36f3891cb3d1208 not found: ID does not exist" Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.353264 4775 scope.go:117] "RemoveContainer" containerID="c86f8502415644d56b4e2accd32a20398aab4a66475b87f1fad3a65e144c004c" Oct 02 02:59:18 crc kubenswrapper[4775]: E1002 02:59:18.353657 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c86f8502415644d56b4e2accd32a20398aab4a66475b87f1fad3a65e144c004c\": container with ID starting with c86f8502415644d56b4e2accd32a20398aab4a66475b87f1fad3a65e144c004c not found: ID does not exist" containerID="c86f8502415644d56b4e2accd32a20398aab4a66475b87f1fad3a65e144c004c" Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.353717 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c86f8502415644d56b4e2accd32a20398aab4a66475b87f1fad3a65e144c004c"} err="failed to get container status \"c86f8502415644d56b4e2accd32a20398aab4a66475b87f1fad3a65e144c004c\": rpc error: code = NotFound desc = could not find container \"c86f8502415644d56b4e2accd32a20398aab4a66475b87f1fad3a65e144c004c\": container with ID starting with c86f8502415644d56b4e2accd32a20398aab4a66475b87f1fad3a65e144c004c not found: ID does not exist" Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.353764 4775 scope.go:117] "RemoveContainer" containerID="9c4584d726023652de3086a8c954f3a05493b08cc85cabd3a804edcaa3070a26" Oct 02 02:59:18 crc kubenswrapper[4775]: E1002 02:59:18.354191 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c4584d726023652de3086a8c954f3a05493b08cc85cabd3a804edcaa3070a26\": container with ID starting with 9c4584d726023652de3086a8c954f3a05493b08cc85cabd3a804edcaa3070a26 not found: ID does not exist" containerID="9c4584d726023652de3086a8c954f3a05493b08cc85cabd3a804edcaa3070a26" Oct 02 02:59:18 crc kubenswrapper[4775]: I1002 02:59:18.354239 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c4584d726023652de3086a8c954f3a05493b08cc85cabd3a804edcaa3070a26"} err="failed to get container status \"9c4584d726023652de3086a8c954f3a05493b08cc85cabd3a804edcaa3070a26\": rpc error: code = NotFound desc = could not find container \"9c4584d726023652de3086a8c954f3a05493b08cc85cabd3a804edcaa3070a26\": container with ID starting with 9c4584d726023652de3086a8c954f3a05493b08cc85cabd3a804edcaa3070a26 not found: ID does not exist" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.605947 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-bgzg5"] Oct 02 02:59:19 crc kubenswrapper[4775]: E1002 02:59:19.606275 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79bd73b6-a318-44be-b594-f7612c922130" containerName="registry-server" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.606286 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="79bd73b6-a318-44be-b594-f7612c922130" containerName="registry-server" Oct 02 02:59:19 crc kubenswrapper[4775]: E1002 02:59:19.606314 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79bd73b6-a318-44be-b594-f7612c922130" containerName="extract-content" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.606320 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="79bd73b6-a318-44be-b594-f7612c922130" containerName="extract-content" Oct 02 02:59:19 crc kubenswrapper[4775]: E1002 02:59:19.606338 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79bd73b6-a318-44be-b594-f7612c922130" containerName="extract-utilities" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.606344 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="79bd73b6-a318-44be-b594-f7612c922130" containerName="extract-utilities" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.606486 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="79bd73b6-a318-44be-b594-f7612c922130" containerName="registry-server" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.607229 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.668141 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-bgzg5"] Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.763990 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/186f8050-b541-4d3f-b79a-8113264792d8-config\") pod \"dnsmasq-dns-5b7946d7b9-bgzg5\" (UID: \"186f8050-b541-4d3f-b79a-8113264792d8\") " pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.764083 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xcxd\" (UniqueName: \"kubernetes.io/projected/186f8050-b541-4d3f-b79a-8113264792d8-kube-api-access-7xcxd\") pod \"dnsmasq-dns-5b7946d7b9-bgzg5\" (UID: \"186f8050-b541-4d3f-b79a-8113264792d8\") " pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.764112 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/186f8050-b541-4d3f-b79a-8113264792d8-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-bgzg5\" (UID: \"186f8050-b541-4d3f-b79a-8113264792d8\") " pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.774359 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79bd73b6-a318-44be-b594-f7612c922130" path="/var/lib/kubelet/pods/79bd73b6-a318-44be-b594-f7612c922130/volumes" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.865229 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xcxd\" (UniqueName: \"kubernetes.io/projected/186f8050-b541-4d3f-b79a-8113264792d8-kube-api-access-7xcxd\") pod \"dnsmasq-dns-5b7946d7b9-bgzg5\" (UID: \"186f8050-b541-4d3f-b79a-8113264792d8\") " pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.865287 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/186f8050-b541-4d3f-b79a-8113264792d8-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-bgzg5\" (UID: \"186f8050-b541-4d3f-b79a-8113264792d8\") " pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.865481 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/186f8050-b541-4d3f-b79a-8113264792d8-config\") pod \"dnsmasq-dns-5b7946d7b9-bgzg5\" (UID: \"186f8050-b541-4d3f-b79a-8113264792d8\") " pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.866530 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/186f8050-b541-4d3f-b79a-8113264792d8-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-bgzg5\" (UID: \"186f8050-b541-4d3f-b79a-8113264792d8\") " pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.867773 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/186f8050-b541-4d3f-b79a-8113264792d8-config\") pod \"dnsmasq-dns-5b7946d7b9-bgzg5\" (UID: \"186f8050-b541-4d3f-b79a-8113264792d8\") " pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.886334 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xcxd\" (UniqueName: \"kubernetes.io/projected/186f8050-b541-4d3f-b79a-8113264792d8-kube-api-access-7xcxd\") pod \"dnsmasq-dns-5b7946d7b9-bgzg5\" (UID: \"186f8050-b541-4d3f-b79a-8113264792d8\") " pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 02:59:19 crc kubenswrapper[4775]: I1002 02:59:19.932664 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 02:59:20 crc kubenswrapper[4775]: I1002 02:59:20.221906 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-bgzg5"] Oct 02 02:59:20 crc kubenswrapper[4775]: I1002 02:59:20.276575 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" event={"ID":"186f8050-b541-4d3f-b79a-8113264792d8","Type":"ContainerStarted","Data":"00150e8dc885934dc8db44b73e8b1931b217bd4ef1c97b7a636762bc9920d340"} Oct 02 02:59:20 crc kubenswrapper[4775]: I1002 02:59:20.419262 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:59:21 crc kubenswrapper[4775]: I1002 02:59:21.122551 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:59:21 crc kubenswrapper[4775]: I1002 02:59:21.284670 4775 generic.go:334] "Generic (PLEG): container finished" podID="186f8050-b541-4d3f-b79a-8113264792d8" containerID="9a670f7be5f11680f800b7815672b4f6d8fac5de283a0f31aed8b01effed334d" exitCode=0 Oct 02 02:59:21 crc kubenswrapper[4775]: I1002 02:59:21.284716 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" event={"ID":"186f8050-b541-4d3f-b79a-8113264792d8","Type":"ContainerDied","Data":"9a670f7be5f11680f800b7815672b4f6d8fac5de283a0f31aed8b01effed334d"} Oct 02 02:59:22 crc kubenswrapper[4775]: I1002 02:59:22.293791 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" event={"ID":"186f8050-b541-4d3f-b79a-8113264792d8","Type":"ContainerStarted","Data":"7a70bee8fc23fff70a91faa6041da6b3f16970a6325dc89ce6826d7b9f587102"} Oct 02 02:59:22 crc kubenswrapper[4775]: I1002 02:59:22.294766 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 02:59:22 crc kubenswrapper[4775]: I1002 02:59:22.315682 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" podStartSLOduration=3.315663915 podStartE2EDuration="3.315663915s" podCreationTimestamp="2025-10-02 02:59:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 02:59:22.313622552 +0000 UTC m=+4699.480366592" watchObservedRunningTime="2025-10-02 02:59:22.315663915 +0000 UTC m=+4699.482407955" Oct 02 02:59:23 crc kubenswrapper[4775]: I1002 02:59:23.063624 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="99fc6db3-c1af-4901-a31f-d53db4848812" containerName="rabbitmq" containerID="cri-o://1e2fb2ef0f0422cba22de4086de226d27399f67395284849c554b529cc58338e" gracePeriod=604798 Oct 02 02:59:23 crc kubenswrapper[4775]: I1002 02:59:23.481614 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="420b6a74-82ea-4527-8fbf-693134e967ab" containerName="rabbitmq" containerID="cri-o://0f3d565e93f6e9320d5737fe7ad35765dde06839dd9fd60e104e75a7a3889596" gracePeriod=604798 Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.371588 4775 generic.go:334] "Generic (PLEG): container finished" podID="99fc6db3-c1af-4901-a31f-d53db4848812" containerID="1e2fb2ef0f0422cba22de4086de226d27399f67395284849c554b529cc58338e" exitCode=0 Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.371687 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"99fc6db3-c1af-4901-a31f-d53db4848812","Type":"ContainerDied","Data":"1e2fb2ef0f0422cba22de4086de226d27399f67395284849c554b529cc58338e"} Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.834018 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.925630 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/99fc6db3-c1af-4901-a31f-d53db4848812-pod-info\") pod \"99fc6db3-c1af-4901-a31f-d53db4848812\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.925678 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/99fc6db3-c1af-4901-a31f-d53db4848812-erlang-cookie-secret\") pod \"99fc6db3-c1af-4901-a31f-d53db4848812\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.925723 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngj6w\" (UniqueName: \"kubernetes.io/projected/99fc6db3-c1af-4901-a31f-d53db4848812-kube-api-access-ngj6w\") pod \"99fc6db3-c1af-4901-a31f-d53db4848812\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.925906 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\") pod \"99fc6db3-c1af-4901-a31f-d53db4848812\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.925936 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-confd\") pod \"99fc6db3-c1af-4901-a31f-d53db4848812\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.925980 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/99fc6db3-c1af-4901-a31f-d53db4848812-server-conf\") pod \"99fc6db3-c1af-4901-a31f-d53db4848812\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.925999 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/99fc6db3-c1af-4901-a31f-d53db4848812-plugins-conf\") pod \"99fc6db3-c1af-4901-a31f-d53db4848812\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.926047 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-erlang-cookie\") pod \"99fc6db3-c1af-4901-a31f-d53db4848812\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.926064 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-plugins\") pod \"99fc6db3-c1af-4901-a31f-d53db4848812\" (UID: \"99fc6db3-c1af-4901-a31f-d53db4848812\") " Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.926997 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "99fc6db3-c1af-4901-a31f-d53db4848812" (UID: "99fc6db3-c1af-4901-a31f-d53db4848812"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.929385 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99fc6db3-c1af-4901-a31f-d53db4848812-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "99fc6db3-c1af-4901-a31f-d53db4848812" (UID: "99fc6db3-c1af-4901-a31f-d53db4848812"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.929522 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "99fc6db3-c1af-4901-a31f-d53db4848812" (UID: "99fc6db3-c1af-4901-a31f-d53db4848812"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.932350 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99fc6db3-c1af-4901-a31f-d53db4848812-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "99fc6db3-c1af-4901-a31f-d53db4848812" (UID: "99fc6db3-c1af-4901-a31f-d53db4848812"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.933901 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/99fc6db3-c1af-4901-a31f-d53db4848812-pod-info" (OuterVolumeSpecName: "pod-info") pod "99fc6db3-c1af-4901-a31f-d53db4848812" (UID: "99fc6db3-c1af-4901-a31f-d53db4848812"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.934947 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.937213 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99fc6db3-c1af-4901-a31f-d53db4848812-kube-api-access-ngj6w" (OuterVolumeSpecName: "kube-api-access-ngj6w") pod "99fc6db3-c1af-4901-a31f-d53db4848812" (UID: "99fc6db3-c1af-4901-a31f-d53db4848812"). InnerVolumeSpecName "kube-api-access-ngj6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.946674 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04" (OuterVolumeSpecName: "persistence") pod "99fc6db3-c1af-4901-a31f-d53db4848812" (UID: "99fc6db3-c1af-4901-a31f-d53db4848812"). InnerVolumeSpecName "pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 02:59:29 crc kubenswrapper[4775]: I1002 02:59:29.949829 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99fc6db3-c1af-4901-a31f-d53db4848812-server-conf" (OuterVolumeSpecName: "server-conf") pod "99fc6db3-c1af-4901-a31f-d53db4848812" (UID: "99fc6db3-c1af-4901-a31f-d53db4848812"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.021302 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "99fc6db3-c1af-4901-a31f-d53db4848812" (UID: "99fc6db3-c1af-4901-a31f-d53db4848812"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.036710 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\") on node \"crc\" " Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.036758 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.036770 4775 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/99fc6db3-c1af-4901-a31f-d53db4848812-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.036780 4775 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/99fc6db3-c1af-4901-a31f-d53db4848812-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.036790 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.036803 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/99fc6db3-c1af-4901-a31f-d53db4848812-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.036814 4775 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/99fc6db3-c1af-4901-a31f-d53db4848812-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.036822 4775 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/99fc6db3-c1af-4901-a31f-d53db4848812-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.036834 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngj6w\" (UniqueName: \"kubernetes.io/projected/99fc6db3-c1af-4901-a31f-d53db4848812-kube-api-access-ngj6w\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.060621 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-g5fwj"] Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.061144 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" podUID="b9dbb225-b6b9-4b52-92c4-574fcedb4ff6" containerName="dnsmasq-dns" containerID="cri-o://3f2a54ccd14db0e2fcd7b6e000fbcc776364ccce4a442cebabaac8e73f10c44a" gracePeriod=10 Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.065493 4775 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.065654 4775 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04") on node "crc" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.067809 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.138673 4775 reconciler_common.go:293] "Volume detached for volume \"pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.240244 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/420b6a74-82ea-4527-8fbf-693134e967ab-server-conf\") pod \"420b6a74-82ea-4527-8fbf-693134e967ab\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.240326 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/420b6a74-82ea-4527-8fbf-693134e967ab-plugins-conf\") pod \"420b6a74-82ea-4527-8fbf-693134e967ab\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.240390 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-erlang-cookie\") pod \"420b6a74-82ea-4527-8fbf-693134e967ab\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.240719 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a70685d-850e-4d83-b161-782b2593cbab\") pod \"420b6a74-82ea-4527-8fbf-693134e967ab\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.240748 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-plugins\") pod \"420b6a74-82ea-4527-8fbf-693134e967ab\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.240795 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/420b6a74-82ea-4527-8fbf-693134e967ab-pod-info\") pod \"420b6a74-82ea-4527-8fbf-693134e967ab\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.240817 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/420b6a74-82ea-4527-8fbf-693134e967ab-erlang-cookie-secret\") pod \"420b6a74-82ea-4527-8fbf-693134e967ab\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.240842 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-confd\") pod \"420b6a74-82ea-4527-8fbf-693134e967ab\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.240839 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/420b6a74-82ea-4527-8fbf-693134e967ab-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "420b6a74-82ea-4527-8fbf-693134e967ab" (UID: "420b6a74-82ea-4527-8fbf-693134e967ab"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.240892 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s47gv\" (UniqueName: \"kubernetes.io/projected/420b6a74-82ea-4527-8fbf-693134e967ab-kube-api-access-s47gv\") pod \"420b6a74-82ea-4527-8fbf-693134e967ab\" (UID: \"420b6a74-82ea-4527-8fbf-693134e967ab\") " Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.241256 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "420b6a74-82ea-4527-8fbf-693134e967ab" (UID: "420b6a74-82ea-4527-8fbf-693134e967ab"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.241549 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "420b6a74-82ea-4527-8fbf-693134e967ab" (UID: "420b6a74-82ea-4527-8fbf-693134e967ab"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.241574 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.241737 4775 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/420b6a74-82ea-4527-8fbf-693134e967ab-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.244840 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/420b6a74-82ea-4527-8fbf-693134e967ab-kube-api-access-s47gv" (OuterVolumeSpecName: "kube-api-access-s47gv") pod "420b6a74-82ea-4527-8fbf-693134e967ab" (UID: "420b6a74-82ea-4527-8fbf-693134e967ab"). InnerVolumeSpecName "kube-api-access-s47gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.246072 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/420b6a74-82ea-4527-8fbf-693134e967ab-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "420b6a74-82ea-4527-8fbf-693134e967ab" (UID: "420b6a74-82ea-4527-8fbf-693134e967ab"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.246890 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/420b6a74-82ea-4527-8fbf-693134e967ab-pod-info" (OuterVolumeSpecName: "pod-info") pod "420b6a74-82ea-4527-8fbf-693134e967ab" (UID: "420b6a74-82ea-4527-8fbf-693134e967ab"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.256024 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a70685d-850e-4d83-b161-782b2593cbab" (OuterVolumeSpecName: "persistence") pod "420b6a74-82ea-4527-8fbf-693134e967ab" (UID: "420b6a74-82ea-4527-8fbf-693134e967ab"). InnerVolumeSpecName "pvc-1a70685d-850e-4d83-b161-782b2593cbab". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.262247 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/420b6a74-82ea-4527-8fbf-693134e967ab-server-conf" (OuterVolumeSpecName: "server-conf") pod "420b6a74-82ea-4527-8fbf-693134e967ab" (UID: "420b6a74-82ea-4527-8fbf-693134e967ab"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.316643 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "420b6a74-82ea-4527-8fbf-693134e967ab" (UID: "420b6a74-82ea-4527-8fbf-693134e967ab"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.344900 4775 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-1a70685d-850e-4d83-b161-782b2593cbab\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a70685d-850e-4d83-b161-782b2593cbab\") on node \"crc\" " Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.344969 4775 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/420b6a74-82ea-4527-8fbf-693134e967ab-pod-info\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.345009 4775 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/420b6a74-82ea-4527-8fbf-693134e967ab-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.345024 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.345036 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s47gv\" (UniqueName: \"kubernetes.io/projected/420b6a74-82ea-4527-8fbf-693134e967ab-kube-api-access-s47gv\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.345047 4775 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/420b6a74-82ea-4527-8fbf-693134e967ab-server-conf\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.345057 4775 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/420b6a74-82ea-4527-8fbf-693134e967ab-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.374245 4775 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.374387 4775 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-1a70685d-850e-4d83-b161-782b2593cbab" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a70685d-850e-4d83-b161-782b2593cbab") on node "crc" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.388191 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"99fc6db3-c1af-4901-a31f-d53db4848812","Type":"ContainerDied","Data":"de2df915a7ee207a294778926a8e5b83ad7b7e33bf4bf49eaf2f9ec9641d6ad3"} Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.388241 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.388261 4775 scope.go:117] "RemoveContainer" containerID="1e2fb2ef0f0422cba22de4086de226d27399f67395284849c554b529cc58338e" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.391853 4775 generic.go:334] "Generic (PLEG): container finished" podID="420b6a74-82ea-4527-8fbf-693134e967ab" containerID="0f3d565e93f6e9320d5737fe7ad35765dde06839dd9fd60e104e75a7a3889596" exitCode=0 Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.391919 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"420b6a74-82ea-4527-8fbf-693134e967ab","Type":"ContainerDied","Data":"0f3d565e93f6e9320d5737fe7ad35765dde06839dd9fd60e104e75a7a3889596"} Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.391942 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"420b6a74-82ea-4527-8fbf-693134e967ab","Type":"ContainerDied","Data":"705141c0cf2b492625a39adc3baea4ddc0f60a004f3a46edbdcfe0f7f763e211"} Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.392095 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.398720 4775 generic.go:334] "Generic (PLEG): container finished" podID="b9dbb225-b6b9-4b52-92c4-574fcedb4ff6" containerID="3f2a54ccd14db0e2fcd7b6e000fbcc776364ccce4a442cebabaac8e73f10c44a" exitCode=0 Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.398760 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" event={"ID":"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6","Type":"ContainerDied","Data":"3f2a54ccd14db0e2fcd7b6e000fbcc776364ccce4a442cebabaac8e73f10c44a"} Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.425231 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.442007 4775 scope.go:117] "RemoveContainer" containerID="cc2d12bb162e34b77ebb9006e65d254bcda59cb3d50eeb1685aee09c5b728139" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.447696 4775 reconciler_common.go:293] "Volume detached for volume \"pvc-1a70685d-850e-4d83-b161-782b2593cbab\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a70685d-850e-4d83-b161-782b2593cbab\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.458430 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.466235 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.475293 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.484385 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:59:30 crc kubenswrapper[4775]: E1002 02:59:30.484799 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="420b6a74-82ea-4527-8fbf-693134e967ab" containerName="rabbitmq" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.484820 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="420b6a74-82ea-4527-8fbf-693134e967ab" containerName="rabbitmq" Oct 02 02:59:30 crc kubenswrapper[4775]: E1002 02:59:30.484832 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99fc6db3-c1af-4901-a31f-d53db4848812" containerName="setup-container" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.484839 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="99fc6db3-c1af-4901-a31f-d53db4848812" containerName="setup-container" Oct 02 02:59:30 crc kubenswrapper[4775]: E1002 02:59:30.484864 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="420b6a74-82ea-4527-8fbf-693134e967ab" containerName="setup-container" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.484870 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="420b6a74-82ea-4527-8fbf-693134e967ab" containerName="setup-container" Oct 02 02:59:30 crc kubenswrapper[4775]: E1002 02:59:30.484883 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99fc6db3-c1af-4901-a31f-d53db4848812" containerName="rabbitmq" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.484889 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="99fc6db3-c1af-4901-a31f-d53db4848812" containerName="rabbitmq" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.485058 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="99fc6db3-c1af-4901-a31f-d53db4848812" containerName="rabbitmq" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.485075 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="420b6a74-82ea-4527-8fbf-693134e967ab" containerName="rabbitmq" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.485878 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.487649 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.487827 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.488114 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-g8wph" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.488611 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.491861 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.494367 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.512914 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.514118 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.515726 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.516661 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.516660 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.516782 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-ntv97" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.517250 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.520856 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.532095 4775 scope.go:117] "RemoveContainer" containerID="0f3d565e93f6e9320d5737fe7ad35765dde06839dd9fd60e104e75a7a3889596" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.547580 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.553852 4775 scope.go:117] "RemoveContainer" containerID="2d14d3219f2f7c57493d64680a568054547aa6b9df059c06f9b2f67a5d92295c" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.596219 4775 scope.go:117] "RemoveContainer" containerID="0f3d565e93f6e9320d5737fe7ad35765dde06839dd9fd60e104e75a7a3889596" Oct 02 02:59:30 crc kubenswrapper[4775]: E1002 02:59:30.596704 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f3d565e93f6e9320d5737fe7ad35765dde06839dd9fd60e104e75a7a3889596\": container with ID starting with 0f3d565e93f6e9320d5737fe7ad35765dde06839dd9fd60e104e75a7a3889596 not found: ID does not exist" containerID="0f3d565e93f6e9320d5737fe7ad35765dde06839dd9fd60e104e75a7a3889596" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.596734 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f3d565e93f6e9320d5737fe7ad35765dde06839dd9fd60e104e75a7a3889596"} err="failed to get container status \"0f3d565e93f6e9320d5737fe7ad35765dde06839dd9fd60e104e75a7a3889596\": rpc error: code = NotFound desc = could not find container \"0f3d565e93f6e9320d5737fe7ad35765dde06839dd9fd60e104e75a7a3889596\": container with ID starting with 0f3d565e93f6e9320d5737fe7ad35765dde06839dd9fd60e104e75a7a3889596 not found: ID does not exist" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.596758 4775 scope.go:117] "RemoveContainer" containerID="2d14d3219f2f7c57493d64680a568054547aa6b9df059c06f9b2f67a5d92295c" Oct 02 02:59:30 crc kubenswrapper[4775]: E1002 02:59:30.597188 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d14d3219f2f7c57493d64680a568054547aa6b9df059c06f9b2f67a5d92295c\": container with ID starting with 2d14d3219f2f7c57493d64680a568054547aa6b9df059c06f9b2f67a5d92295c not found: ID does not exist" containerID="2d14d3219f2f7c57493d64680a568054547aa6b9df059c06f9b2f67a5d92295c" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.597208 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d14d3219f2f7c57493d64680a568054547aa6b9df059c06f9b2f67a5d92295c"} err="failed to get container status \"2d14d3219f2f7c57493d64680a568054547aa6b9df059c06f9b2f67a5d92295c\": rpc error: code = NotFound desc = could not find container \"2d14d3219f2f7c57493d64680a568054547aa6b9df059c06f9b2f67a5d92295c\": container with ID starting with 2d14d3219f2f7c57493d64680a568054547aa6b9df059c06f9b2f67a5d92295c not found: ID does not exist" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.649642 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-dns-svc\") pod \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\" (UID: \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\") " Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.649685 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lscll\" (UniqueName: \"kubernetes.io/projected/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-kube-api-access-lscll\") pod \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\" (UID: \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\") " Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.649839 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-config\") pod \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\" (UID: \"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6\") " Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650011 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/88fede09-aa59-48c1-9957-4ce95fcb95d9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650039 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bf4059c4-4a05-4598-bf74-3e1584829b7b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650054 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghkbm\" (UniqueName: \"kubernetes.io/projected/88fede09-aa59-48c1-9957-4ce95fcb95d9-kube-api-access-ghkbm\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650077 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1a70685d-850e-4d83-b161-782b2593cbab\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a70685d-850e-4d83-b161-782b2593cbab\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650095 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bf4059c4-4a05-4598-bf74-3e1584829b7b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650111 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bf4059c4-4a05-4598-bf74-3e1584829b7b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650128 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/88fede09-aa59-48c1-9957-4ce95fcb95d9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650356 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/88fede09-aa59-48c1-9957-4ce95fcb95d9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650465 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650489 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98wbs\" (UniqueName: \"kubernetes.io/projected/bf4059c4-4a05-4598-bf74-3e1584829b7b-kube-api-access-98wbs\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650539 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/88fede09-aa59-48c1-9957-4ce95fcb95d9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650566 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bf4059c4-4a05-4598-bf74-3e1584829b7b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650591 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bf4059c4-4a05-4598-bf74-3e1584829b7b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650611 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bf4059c4-4a05-4598-bf74-3e1584829b7b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650628 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/88fede09-aa59-48c1-9957-4ce95fcb95d9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650676 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/88fede09-aa59-48c1-9957-4ce95fcb95d9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650815 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bf4059c4-4a05-4598-bf74-3e1584829b7b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.650875 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/88fede09-aa59-48c1-9957-4ce95fcb95d9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.653373 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-kube-api-access-lscll" (OuterVolumeSpecName: "kube-api-access-lscll") pod "b9dbb225-b6b9-4b52-92c4-574fcedb4ff6" (UID: "b9dbb225-b6b9-4b52-92c4-574fcedb4ff6"). InnerVolumeSpecName "kube-api-access-lscll". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.682792 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b9dbb225-b6b9-4b52-92c4-574fcedb4ff6" (UID: "b9dbb225-b6b9-4b52-92c4-574fcedb4ff6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.684176 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-config" (OuterVolumeSpecName: "config") pod "b9dbb225-b6b9-4b52-92c4-574fcedb4ff6" (UID: "b9dbb225-b6b9-4b52-92c4-574fcedb4ff6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752634 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bf4059c4-4a05-4598-bf74-3e1584829b7b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752682 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/88fede09-aa59-48c1-9957-4ce95fcb95d9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752716 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/88fede09-aa59-48c1-9957-4ce95fcb95d9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752753 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752771 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98wbs\" (UniqueName: \"kubernetes.io/projected/bf4059c4-4a05-4598-bf74-3e1584829b7b-kube-api-access-98wbs\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752796 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/88fede09-aa59-48c1-9957-4ce95fcb95d9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752814 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bf4059c4-4a05-4598-bf74-3e1584829b7b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752829 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bf4059c4-4a05-4598-bf74-3e1584829b7b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752847 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/88fede09-aa59-48c1-9957-4ce95fcb95d9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752864 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bf4059c4-4a05-4598-bf74-3e1584829b7b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752886 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/88fede09-aa59-48c1-9957-4ce95fcb95d9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752919 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bf4059c4-4a05-4598-bf74-3e1584829b7b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752944 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/88fede09-aa59-48c1-9957-4ce95fcb95d9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752982 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/88fede09-aa59-48c1-9957-4ce95fcb95d9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.752998 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bf4059c4-4a05-4598-bf74-3e1584829b7b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.753013 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghkbm\" (UniqueName: \"kubernetes.io/projected/88fede09-aa59-48c1-9957-4ce95fcb95d9-kube-api-access-ghkbm\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.753034 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1a70685d-850e-4d83-b161-782b2593cbab\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a70685d-850e-4d83-b161-782b2593cbab\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.753049 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bf4059c4-4a05-4598-bf74-3e1584829b7b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.753088 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-config\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.753098 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.753107 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lscll\" (UniqueName: \"kubernetes.io/projected/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6-kube-api-access-lscll\") on node \"crc\" DevicePath \"\"" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.753471 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bf4059c4-4a05-4598-bf74-3e1584829b7b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.753865 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bf4059c4-4a05-4598-bf74-3e1584829b7b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.754210 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bf4059c4-4a05-4598-bf74-3e1584829b7b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.754359 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/88fede09-aa59-48c1-9957-4ce95fcb95d9-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.754610 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bf4059c4-4a05-4598-bf74-3e1584829b7b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.755141 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/88fede09-aa59-48c1-9957-4ce95fcb95d9-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.755217 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/88fede09-aa59-48c1-9957-4ce95fcb95d9-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.755413 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/88fede09-aa59-48c1-9957-4ce95fcb95d9-server-conf\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.756356 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.756382 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1a70685d-850e-4d83-b161-782b2593cbab\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a70685d-850e-4d83-b161-782b2593cbab\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/bdce4809b2d1fad6e3aff8ca06261a732fcca42e6e6fdf3a8e553419523103d5/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.758205 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bf4059c4-4a05-4598-bf74-3e1584829b7b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.759508 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/88fede09-aa59-48c1-9957-4ce95fcb95d9-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.759805 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/88fede09-aa59-48c1-9957-4ce95fcb95d9-pod-info\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.760099 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.761019 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bf4059c4-4a05-4598-bf74-3e1584829b7b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.761114 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bf4059c4-4a05-4598-bf74-3e1584829b7b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.762083 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0e979186b38ce1a2c20c88de979de968f05e62b573b4fa37c206cee176706195/globalmount\"" pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.771095 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98wbs\" (UniqueName: \"kubernetes.io/projected/bf4059c4-4a05-4598-bf74-3e1584829b7b-kube-api-access-98wbs\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.773440 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/88fede09-aa59-48c1-9957-4ce95fcb95d9-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.783280 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghkbm\" (UniqueName: \"kubernetes.io/projected/88fede09-aa59-48c1-9957-4ce95fcb95d9-kube-api-access-ghkbm\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.792135 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ad5e2dc4-4834-44bc-90b6-d5f1317c5a04\") pod \"rabbitmq-server-0\" (UID: \"88fede09-aa59-48c1-9957-4ce95fcb95d9\") " pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.800987 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1a70685d-850e-4d83-b161-782b2593cbab\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1a70685d-850e-4d83-b161-782b2593cbab\") pod \"rabbitmq-cell1-server-0\" (UID: \"bf4059c4-4a05-4598-bf74-3e1584829b7b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.864297 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 02 02:59:30 crc kubenswrapper[4775]: I1002 02:59:30.884480 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 02 02:59:31 crc kubenswrapper[4775]: I1002 02:59:31.281338 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 02 02:59:31 crc kubenswrapper[4775]: I1002 02:59:31.327450 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 02 02:59:31 crc kubenswrapper[4775]: I1002 02:59:31.411235 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bf4059c4-4a05-4598-bf74-3e1584829b7b","Type":"ContainerStarted","Data":"c359bcbc6ffa72f11ee3fe2d6f5c586b9c00e3c1859b60fd5c78470aa398ef8d"} Oct 02 02:59:31 crc kubenswrapper[4775]: I1002 02:59:31.413923 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"88fede09-aa59-48c1-9957-4ce95fcb95d9","Type":"ContainerStarted","Data":"a07b33db6371147058f4ec23abc57f5415e0173c4569754d5f2f471db784f5ad"} Oct 02 02:59:31 crc kubenswrapper[4775]: I1002 02:59:31.419598 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" event={"ID":"b9dbb225-b6b9-4b52-92c4-574fcedb4ff6","Type":"ContainerDied","Data":"356ab411980f5673a0f10ad3fa3dfadc62e9f16ece00f8e5234e9bcaa2eb3d7d"} Oct 02 02:59:31 crc kubenswrapper[4775]: I1002 02:59:31.419641 4775 scope.go:117] "RemoveContainer" containerID="3f2a54ccd14db0e2fcd7b6e000fbcc776364ccce4a442cebabaac8e73f10c44a" Oct 02 02:59:31 crc kubenswrapper[4775]: I1002 02:59:31.419718 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-g5fwj" Oct 02 02:59:31 crc kubenswrapper[4775]: I1002 02:59:31.683997 4775 scope.go:117] "RemoveContainer" containerID="5f3b845076fb6f0405bb72690422c287579fc450b5efe471886361e0f96947b5" Oct 02 02:59:31 crc kubenswrapper[4775]: I1002 02:59:31.717157 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-g5fwj"] Oct 02 02:59:31 crc kubenswrapper[4775]: I1002 02:59:31.729877 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-g5fwj"] Oct 02 02:59:31 crc kubenswrapper[4775]: I1002 02:59:31.780776 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="420b6a74-82ea-4527-8fbf-693134e967ab" path="/var/lib/kubelet/pods/420b6a74-82ea-4527-8fbf-693134e967ab/volumes" Oct 02 02:59:31 crc kubenswrapper[4775]: I1002 02:59:31.783281 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99fc6db3-c1af-4901-a31f-d53db4848812" path="/var/lib/kubelet/pods/99fc6db3-c1af-4901-a31f-d53db4848812/volumes" Oct 02 02:59:31 crc kubenswrapper[4775]: I1002 02:59:31.784146 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9dbb225-b6b9-4b52-92c4-574fcedb4ff6" path="/var/lib/kubelet/pods/b9dbb225-b6b9-4b52-92c4-574fcedb4ff6/volumes" Oct 02 02:59:34 crc kubenswrapper[4775]: I1002 02:59:34.452696 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bf4059c4-4a05-4598-bf74-3e1584829b7b","Type":"ContainerStarted","Data":"c77e9cec5bae8d43e752452aa0b0f90b3336553397c45e22c4a0be048a4cc3f6"} Oct 02 02:59:34 crc kubenswrapper[4775]: I1002 02:59:34.455573 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"88fede09-aa59-48c1-9957-4ce95fcb95d9","Type":"ContainerStarted","Data":"5a758b5ebfa3198ab6b0e016453205754fa6ce6a651f37662b5804ef1bbde6da"} Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.165397 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l"] Oct 02 03:00:00 crc kubenswrapper[4775]: E1002 03:00:00.166460 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9dbb225-b6b9-4b52-92c4-574fcedb4ff6" containerName="init" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.166481 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9dbb225-b6b9-4b52-92c4-574fcedb4ff6" containerName="init" Oct 02 03:00:00 crc kubenswrapper[4775]: E1002 03:00:00.166512 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9dbb225-b6b9-4b52-92c4-574fcedb4ff6" containerName="dnsmasq-dns" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.166526 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9dbb225-b6b9-4b52-92c4-574fcedb4ff6" containerName="dnsmasq-dns" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.166838 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9dbb225-b6b9-4b52-92c4-574fcedb4ff6" containerName="dnsmasq-dns" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.167680 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.170423 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.171019 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.180663 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l"] Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.349656 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50401de0-705a-4de5-b211-56688c040651-config-volume\") pod \"collect-profiles-29322900-bft5l\" (UID: \"50401de0-705a-4de5-b211-56688c040651\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.349817 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50401de0-705a-4de5-b211-56688c040651-secret-volume\") pod \"collect-profiles-29322900-bft5l\" (UID: \"50401de0-705a-4de5-b211-56688c040651\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.349864 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km5sp\" (UniqueName: \"kubernetes.io/projected/50401de0-705a-4de5-b211-56688c040651-kube-api-access-km5sp\") pod \"collect-profiles-29322900-bft5l\" (UID: \"50401de0-705a-4de5-b211-56688c040651\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.451311 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50401de0-705a-4de5-b211-56688c040651-secret-volume\") pod \"collect-profiles-29322900-bft5l\" (UID: \"50401de0-705a-4de5-b211-56688c040651\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.451374 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km5sp\" (UniqueName: \"kubernetes.io/projected/50401de0-705a-4de5-b211-56688c040651-kube-api-access-km5sp\") pod \"collect-profiles-29322900-bft5l\" (UID: \"50401de0-705a-4de5-b211-56688c040651\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.451484 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50401de0-705a-4de5-b211-56688c040651-config-volume\") pod \"collect-profiles-29322900-bft5l\" (UID: \"50401de0-705a-4de5-b211-56688c040651\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.452521 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50401de0-705a-4de5-b211-56688c040651-config-volume\") pod \"collect-profiles-29322900-bft5l\" (UID: \"50401de0-705a-4de5-b211-56688c040651\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.460192 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50401de0-705a-4de5-b211-56688c040651-secret-volume\") pod \"collect-profiles-29322900-bft5l\" (UID: \"50401de0-705a-4de5-b211-56688c040651\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.471232 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km5sp\" (UniqueName: \"kubernetes.io/projected/50401de0-705a-4de5-b211-56688c040651-kube-api-access-km5sp\") pod \"collect-profiles-29322900-bft5l\" (UID: \"50401de0-705a-4de5-b211-56688c040651\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.502563 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" Oct 02 03:00:00 crc kubenswrapper[4775]: I1002 03:00:00.992393 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l"] Oct 02 03:00:01 crc kubenswrapper[4775]: I1002 03:00:01.719563 4775 generic.go:334] "Generic (PLEG): container finished" podID="50401de0-705a-4de5-b211-56688c040651" containerID="3fcb369bc37736ee508ccc3df41fe39721387a332a1c8dffcd55e2d473ade333" exitCode=0 Oct 02 03:00:01 crc kubenswrapper[4775]: I1002 03:00:01.719607 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" event={"ID":"50401de0-705a-4de5-b211-56688c040651","Type":"ContainerDied","Data":"3fcb369bc37736ee508ccc3df41fe39721387a332a1c8dffcd55e2d473ade333"} Oct 02 03:00:01 crc kubenswrapper[4775]: I1002 03:00:01.719839 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" event={"ID":"50401de0-705a-4de5-b211-56688c040651","Type":"ContainerStarted","Data":"f261b1959cbeda7ee02d32fcf6f9cad64fa2ae275fbdb5e0975e1353679f5d16"} Oct 02 03:00:03 crc kubenswrapper[4775]: I1002 03:00:03.098052 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" Oct 02 03:00:03 crc kubenswrapper[4775]: I1002 03:00:03.193358 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km5sp\" (UniqueName: \"kubernetes.io/projected/50401de0-705a-4de5-b211-56688c040651-kube-api-access-km5sp\") pod \"50401de0-705a-4de5-b211-56688c040651\" (UID: \"50401de0-705a-4de5-b211-56688c040651\") " Oct 02 03:00:03 crc kubenswrapper[4775]: I1002 03:00:03.193580 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50401de0-705a-4de5-b211-56688c040651-config-volume\") pod \"50401de0-705a-4de5-b211-56688c040651\" (UID: \"50401de0-705a-4de5-b211-56688c040651\") " Oct 02 03:00:03 crc kubenswrapper[4775]: I1002 03:00:03.193783 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50401de0-705a-4de5-b211-56688c040651-secret-volume\") pod \"50401de0-705a-4de5-b211-56688c040651\" (UID: \"50401de0-705a-4de5-b211-56688c040651\") " Oct 02 03:00:03 crc kubenswrapper[4775]: I1002 03:00:03.194657 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/50401de0-705a-4de5-b211-56688c040651-config-volume" (OuterVolumeSpecName: "config-volume") pod "50401de0-705a-4de5-b211-56688c040651" (UID: "50401de0-705a-4de5-b211-56688c040651"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:00:03 crc kubenswrapper[4775]: I1002 03:00:03.203833 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50401de0-705a-4de5-b211-56688c040651-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "50401de0-705a-4de5-b211-56688c040651" (UID: "50401de0-705a-4de5-b211-56688c040651"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:00:03 crc kubenswrapper[4775]: I1002 03:00:03.204092 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50401de0-705a-4de5-b211-56688c040651-kube-api-access-km5sp" (OuterVolumeSpecName: "kube-api-access-km5sp") pod "50401de0-705a-4de5-b211-56688c040651" (UID: "50401de0-705a-4de5-b211-56688c040651"). InnerVolumeSpecName "kube-api-access-km5sp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:00:03 crc kubenswrapper[4775]: I1002 03:00:03.296462 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km5sp\" (UniqueName: \"kubernetes.io/projected/50401de0-705a-4de5-b211-56688c040651-kube-api-access-km5sp\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:03 crc kubenswrapper[4775]: I1002 03:00:03.296541 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/50401de0-705a-4de5-b211-56688c040651-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:03 crc kubenswrapper[4775]: I1002 03:00:03.296560 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/50401de0-705a-4de5-b211-56688c040651-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:03 crc kubenswrapper[4775]: I1002 03:00:03.743732 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" event={"ID":"50401de0-705a-4de5-b211-56688c040651","Type":"ContainerDied","Data":"f261b1959cbeda7ee02d32fcf6f9cad64fa2ae275fbdb5e0975e1353679f5d16"} Oct 02 03:00:03 crc kubenswrapper[4775]: I1002 03:00:03.743788 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f261b1959cbeda7ee02d32fcf6f9cad64fa2ae275fbdb5e0975e1353679f5d16" Oct 02 03:00:03 crc kubenswrapper[4775]: I1002 03:00:03.743848 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l" Oct 02 03:00:04 crc kubenswrapper[4775]: I1002 03:00:04.174384 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8"] Oct 02 03:00:04 crc kubenswrapper[4775]: I1002 03:00:04.181882 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322855-nxwc8"] Oct 02 03:00:05 crc kubenswrapper[4775]: I1002 03:00:05.782634 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0add0bd8-0890-44ee-baf8-0ff973140d3d" path="/var/lib/kubelet/pods/0add0bd8-0890-44ee-baf8-0ff973140d3d/volumes" Oct 02 03:00:06 crc kubenswrapper[4775]: I1002 03:00:06.780319 4775 generic.go:334] "Generic (PLEG): container finished" podID="bf4059c4-4a05-4598-bf74-3e1584829b7b" containerID="c77e9cec5bae8d43e752452aa0b0f90b3336553397c45e22c4a0be048a4cc3f6" exitCode=0 Oct 02 03:00:06 crc kubenswrapper[4775]: I1002 03:00:06.780418 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bf4059c4-4a05-4598-bf74-3e1584829b7b","Type":"ContainerDied","Data":"c77e9cec5bae8d43e752452aa0b0f90b3336553397c45e22c4a0be048a4cc3f6"} Oct 02 03:00:06 crc kubenswrapper[4775]: I1002 03:00:06.786412 4775 generic.go:334] "Generic (PLEG): container finished" podID="88fede09-aa59-48c1-9957-4ce95fcb95d9" containerID="5a758b5ebfa3198ab6b0e016453205754fa6ce6a651f37662b5804ef1bbde6da" exitCode=0 Oct 02 03:00:06 crc kubenswrapper[4775]: I1002 03:00:06.787159 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"88fede09-aa59-48c1-9957-4ce95fcb95d9","Type":"ContainerDied","Data":"5a758b5ebfa3198ab6b0e016453205754fa6ce6a651f37662b5804ef1bbde6da"} Oct 02 03:00:07 crc kubenswrapper[4775]: I1002 03:00:07.233615 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:00:07 crc kubenswrapper[4775]: I1002 03:00:07.234032 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:00:07 crc kubenswrapper[4775]: E1002 03:00:07.628511 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50401de0_705a_4de5_b211_56688c040651.slice\": RecentStats: unable to find data in memory cache]" Oct 02 03:00:07 crc kubenswrapper[4775]: I1002 03:00:07.797255 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"bf4059c4-4a05-4598-bf74-3e1584829b7b","Type":"ContainerStarted","Data":"155cff3b736da65e14fcffb1444f9d19814d222f0d6976eb6821986a83226e93"} Oct 02 03:00:07 crc kubenswrapper[4775]: I1002 03:00:07.797486 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 02 03:00:07 crc kubenswrapper[4775]: I1002 03:00:07.800437 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"88fede09-aa59-48c1-9957-4ce95fcb95d9","Type":"ContainerStarted","Data":"75a52bb133a1bf94cfda0a0ae38713051bac8bc03f0c0ab85cf8fc2a3c46595a"} Oct 02 03:00:07 crc kubenswrapper[4775]: I1002 03:00:07.800667 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 02 03:00:07 crc kubenswrapper[4775]: I1002 03:00:07.817533 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.817513758 podStartE2EDuration="37.817513758s" podCreationTimestamp="2025-10-02 02:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:00:07.815914316 +0000 UTC m=+4744.982658376" watchObservedRunningTime="2025-10-02 03:00:07.817513758 +0000 UTC m=+4744.984257808" Oct 02 03:00:07 crc kubenswrapper[4775]: I1002 03:00:07.839100 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.839081551 podStartE2EDuration="37.839081551s" podCreationTimestamp="2025-10-02 02:59:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:00:07.834708697 +0000 UTC m=+4745.001452737" watchObservedRunningTime="2025-10-02 03:00:07.839081551 +0000 UTC m=+4745.005825611" Oct 02 03:00:10 crc kubenswrapper[4775]: I1002 03:00:10.142980 4775 scope.go:117] "RemoveContainer" containerID="a84d3b997d8bde10f020c0eb5e84422a2906a19876e9708c1eeebb6b847d644c" Oct 02 03:00:17 crc kubenswrapper[4775]: E1002 03:00:17.896232 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50401de0_705a_4de5_b211_56688c040651.slice\": RecentStats: unable to find data in memory cache]" Oct 02 03:00:20 crc kubenswrapper[4775]: I1002 03:00:20.868310 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 02 03:00:20 crc kubenswrapper[4775]: I1002 03:00:20.888524 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 02 03:00:28 crc kubenswrapper[4775]: E1002 03:00:28.211085 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50401de0_705a_4de5_b211_56688c040651.slice\": RecentStats: unable to find data in memory cache]" Oct 02 03:00:30 crc kubenswrapper[4775]: I1002 03:00:30.041538 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Oct 02 03:00:30 crc kubenswrapper[4775]: E1002 03:00:30.042428 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50401de0-705a-4de5-b211-56688c040651" containerName="collect-profiles" Oct 02 03:00:30 crc kubenswrapper[4775]: I1002 03:00:30.042458 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="50401de0-705a-4de5-b211-56688c040651" containerName="collect-profiles" Oct 02 03:00:30 crc kubenswrapper[4775]: I1002 03:00:30.042773 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="50401de0-705a-4de5-b211-56688c040651" containerName="collect-profiles" Oct 02 03:00:30 crc kubenswrapper[4775]: I1002 03:00:30.043703 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 02 03:00:30 crc kubenswrapper[4775]: I1002 03:00:30.047367 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-tnzbn" Oct 02 03:00:30 crc kubenswrapper[4775]: I1002 03:00:30.055806 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 02 03:00:30 crc kubenswrapper[4775]: I1002 03:00:30.070424 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2ng7\" (UniqueName: \"kubernetes.io/projected/cb6a4264-cc90-4214-b111-beb74367dfbd-kube-api-access-p2ng7\") pod \"mariadb-client-1-default\" (UID: \"cb6a4264-cc90-4214-b111-beb74367dfbd\") " pod="openstack/mariadb-client-1-default" Oct 02 03:00:30 crc kubenswrapper[4775]: I1002 03:00:30.172235 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2ng7\" (UniqueName: \"kubernetes.io/projected/cb6a4264-cc90-4214-b111-beb74367dfbd-kube-api-access-p2ng7\") pod \"mariadb-client-1-default\" (UID: \"cb6a4264-cc90-4214-b111-beb74367dfbd\") " pod="openstack/mariadb-client-1-default" Oct 02 03:00:30 crc kubenswrapper[4775]: I1002 03:00:30.211099 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2ng7\" (UniqueName: \"kubernetes.io/projected/cb6a4264-cc90-4214-b111-beb74367dfbd-kube-api-access-p2ng7\") pod \"mariadb-client-1-default\" (UID: \"cb6a4264-cc90-4214-b111-beb74367dfbd\") " pod="openstack/mariadb-client-1-default" Oct 02 03:00:30 crc kubenswrapper[4775]: I1002 03:00:30.363682 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 02 03:00:30 crc kubenswrapper[4775]: I1002 03:00:30.773308 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 02 03:00:31 crc kubenswrapper[4775]: I1002 03:00:31.018819 4775 generic.go:334] "Generic (PLEG): container finished" podID="cb6a4264-cc90-4214-b111-beb74367dfbd" containerID="98f3f05d32c954cc623282251bab7200b1197c5d21755e5900ba47c85b6b16c0" exitCode=0 Oct 02 03:00:31 crc kubenswrapper[4775]: I1002 03:00:31.018877 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"cb6a4264-cc90-4214-b111-beb74367dfbd","Type":"ContainerDied","Data":"98f3f05d32c954cc623282251bab7200b1197c5d21755e5900ba47c85b6b16c0"} Oct 02 03:00:31 crc kubenswrapper[4775]: I1002 03:00:31.018914 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"cb6a4264-cc90-4214-b111-beb74367dfbd","Type":"ContainerStarted","Data":"3f9b70b06682a234b887d2ad51a2120e1130608c38a2d94efd48461f56bfe593"} Oct 02 03:00:32 crc kubenswrapper[4775]: I1002 03:00:32.551693 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 02 03:00:32 crc kubenswrapper[4775]: I1002 03:00:32.585739 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_cb6a4264-cc90-4214-b111-beb74367dfbd/mariadb-client-1-default/0.log" Oct 02 03:00:32 crc kubenswrapper[4775]: I1002 03:00:32.608764 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2ng7\" (UniqueName: \"kubernetes.io/projected/cb6a4264-cc90-4214-b111-beb74367dfbd-kube-api-access-p2ng7\") pod \"cb6a4264-cc90-4214-b111-beb74367dfbd\" (UID: \"cb6a4264-cc90-4214-b111-beb74367dfbd\") " Oct 02 03:00:32 crc kubenswrapper[4775]: I1002 03:00:32.616053 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 02 03:00:32 crc kubenswrapper[4775]: I1002 03:00:32.616670 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb6a4264-cc90-4214-b111-beb74367dfbd-kube-api-access-p2ng7" (OuterVolumeSpecName: "kube-api-access-p2ng7") pod "cb6a4264-cc90-4214-b111-beb74367dfbd" (UID: "cb6a4264-cc90-4214-b111-beb74367dfbd"). InnerVolumeSpecName "kube-api-access-p2ng7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:00:32 crc kubenswrapper[4775]: I1002 03:00:32.624156 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Oct 02 03:00:32 crc kubenswrapper[4775]: I1002 03:00:32.710489 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2ng7\" (UniqueName: \"kubernetes.io/projected/cb6a4264-cc90-4214-b111-beb74367dfbd-kube-api-access-p2ng7\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:33 crc kubenswrapper[4775]: I1002 03:00:33.043775 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f9b70b06682a234b887d2ad51a2120e1130608c38a2d94efd48461f56bfe593" Oct 02 03:00:33 crc kubenswrapper[4775]: I1002 03:00:33.043884 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Oct 02 03:00:33 crc kubenswrapper[4775]: I1002 03:00:33.133795 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Oct 02 03:00:33 crc kubenswrapper[4775]: E1002 03:00:33.134200 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb6a4264-cc90-4214-b111-beb74367dfbd" containerName="mariadb-client-1-default" Oct 02 03:00:33 crc kubenswrapper[4775]: I1002 03:00:33.134216 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb6a4264-cc90-4214-b111-beb74367dfbd" containerName="mariadb-client-1-default" Oct 02 03:00:33 crc kubenswrapper[4775]: I1002 03:00:33.134385 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb6a4264-cc90-4214-b111-beb74367dfbd" containerName="mariadb-client-1-default" Oct 02 03:00:33 crc kubenswrapper[4775]: I1002 03:00:33.134942 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 02 03:00:33 crc kubenswrapper[4775]: I1002 03:00:33.137351 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-tnzbn" Oct 02 03:00:33 crc kubenswrapper[4775]: I1002 03:00:33.147285 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 02 03:00:33 crc kubenswrapper[4775]: I1002 03:00:33.220815 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bbcl\" (UniqueName: \"kubernetes.io/projected/0089837e-94bd-4547-84af-eba2701b5c47-kube-api-access-2bbcl\") pod \"mariadb-client-2-default\" (UID: \"0089837e-94bd-4547-84af-eba2701b5c47\") " pod="openstack/mariadb-client-2-default" Oct 02 03:00:33 crc kubenswrapper[4775]: I1002 03:00:33.322321 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bbcl\" (UniqueName: \"kubernetes.io/projected/0089837e-94bd-4547-84af-eba2701b5c47-kube-api-access-2bbcl\") pod \"mariadb-client-2-default\" (UID: \"0089837e-94bd-4547-84af-eba2701b5c47\") " pod="openstack/mariadb-client-2-default" Oct 02 03:00:33 crc kubenswrapper[4775]: I1002 03:00:33.353255 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bbcl\" (UniqueName: \"kubernetes.io/projected/0089837e-94bd-4547-84af-eba2701b5c47-kube-api-access-2bbcl\") pod \"mariadb-client-2-default\" (UID: \"0089837e-94bd-4547-84af-eba2701b5c47\") " pod="openstack/mariadb-client-2-default" Oct 02 03:00:33 crc kubenswrapper[4775]: I1002 03:00:33.463442 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 02 03:00:33 crc kubenswrapper[4775]: I1002 03:00:33.782597 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb6a4264-cc90-4214-b111-beb74367dfbd" path="/var/lib/kubelet/pods/cb6a4264-cc90-4214-b111-beb74367dfbd/volumes" Oct 02 03:00:34 crc kubenswrapper[4775]: I1002 03:00:34.115556 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 02 03:00:34 crc kubenswrapper[4775]: W1002 03:00:34.125253 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0089837e_94bd_4547_84af_eba2701b5c47.slice/crio-cb794f10223592253e53413ce4076ddfdb5fc0165bccfb837c0bbc257983d764 WatchSource:0}: Error finding container cb794f10223592253e53413ce4076ddfdb5fc0165bccfb837c0bbc257983d764: Status 404 returned error can't find the container with id cb794f10223592253e53413ce4076ddfdb5fc0165bccfb837c0bbc257983d764 Oct 02 03:00:35 crc kubenswrapper[4775]: I1002 03:00:35.063418 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"0089837e-94bd-4547-84af-eba2701b5c47","Type":"ContainerStarted","Data":"2a225463593bf5f379339e9642c5e4bd93bfc749740d97e7c0ff663fc62eed76"} Oct 02 03:00:35 crc kubenswrapper[4775]: I1002 03:00:35.063894 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"0089837e-94bd-4547-84af-eba2701b5c47","Type":"ContainerStarted","Data":"cb794f10223592253e53413ce4076ddfdb5fc0165bccfb837c0bbc257983d764"} Oct 02 03:00:35 crc kubenswrapper[4775]: I1002 03:00:35.213299 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2-default_0089837e-94bd-4547-84af-eba2701b5c47/mariadb-client-2-default/0.log" Oct 02 03:00:36 crc kubenswrapper[4775]: I1002 03:00:36.073653 4775 generic.go:334] "Generic (PLEG): container finished" podID="0089837e-94bd-4547-84af-eba2701b5c47" containerID="2a225463593bf5f379339e9642c5e4bd93bfc749740d97e7c0ff663fc62eed76" exitCode=0 Oct 02 03:00:36 crc kubenswrapper[4775]: I1002 03:00:36.073791 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"0089837e-94bd-4547-84af-eba2701b5c47","Type":"ContainerDied","Data":"2a225463593bf5f379339e9642c5e4bd93bfc749740d97e7c0ff663fc62eed76"} Oct 02 03:00:37 crc kubenswrapper[4775]: I1002 03:00:37.235086 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:00:37 crc kubenswrapper[4775]: I1002 03:00:37.235173 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.293513 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.327486 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bbcl\" (UniqueName: \"kubernetes.io/projected/0089837e-94bd-4547-84af-eba2701b5c47-kube-api-access-2bbcl\") pod \"0089837e-94bd-4547-84af-eba2701b5c47\" (UID: \"0089837e-94bd-4547-84af-eba2701b5c47\") " Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.334337 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.337339 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0089837e-94bd-4547-84af-eba2701b5c47-kube-api-access-2bbcl" (OuterVolumeSpecName: "kube-api-access-2bbcl") pod "0089837e-94bd-4547-84af-eba2701b5c47" (UID: "0089837e-94bd-4547-84af-eba2701b5c47"). InnerVolumeSpecName "kube-api-access-2bbcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.341606 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.429030 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bbcl\" (UniqueName: \"kubernetes.io/projected/0089837e-94bd-4547-84af-eba2701b5c47-kube-api-access-2bbcl\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:38 crc kubenswrapper[4775]: E1002 03:00:38.444783 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50401de0_705a_4de5_b211_56688c040651.slice\": RecentStats: unable to find data in memory cache]" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.507459 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2jkl4"] Oct 02 03:00:38 crc kubenswrapper[4775]: E1002 03:00:38.509416 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0089837e-94bd-4547-84af-eba2701b5c47" containerName="mariadb-client-2-default" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.509464 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0089837e-94bd-4547-84af-eba2701b5c47" containerName="mariadb-client-2-default" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.509714 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0089837e-94bd-4547-84af-eba2701b5c47" containerName="mariadb-client-2-default" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.511133 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.523086 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2jkl4"] Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.529925 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d322d4f2-73f4-40dd-a66e-7564e032f72b-catalog-content\") pod \"certified-operators-2jkl4\" (UID: \"d322d4f2-73f4-40dd-a66e-7564e032f72b\") " pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.530019 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d322d4f2-73f4-40dd-a66e-7564e032f72b-utilities\") pod \"certified-operators-2jkl4\" (UID: \"d322d4f2-73f4-40dd-a66e-7564e032f72b\") " pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.530390 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz498\" (UniqueName: \"kubernetes.io/projected/d322d4f2-73f4-40dd-a66e-7564e032f72b-kube-api-access-zz498\") pod \"certified-operators-2jkl4\" (UID: \"d322d4f2-73f4-40dd-a66e-7564e032f72b\") " pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.632453 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz498\" (UniqueName: \"kubernetes.io/projected/d322d4f2-73f4-40dd-a66e-7564e032f72b-kube-api-access-zz498\") pod \"certified-operators-2jkl4\" (UID: \"d322d4f2-73f4-40dd-a66e-7564e032f72b\") " pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.632829 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d322d4f2-73f4-40dd-a66e-7564e032f72b-catalog-content\") pod \"certified-operators-2jkl4\" (UID: \"d322d4f2-73f4-40dd-a66e-7564e032f72b\") " pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.632855 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d322d4f2-73f4-40dd-a66e-7564e032f72b-utilities\") pod \"certified-operators-2jkl4\" (UID: \"d322d4f2-73f4-40dd-a66e-7564e032f72b\") " pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.633366 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d322d4f2-73f4-40dd-a66e-7564e032f72b-utilities\") pod \"certified-operators-2jkl4\" (UID: \"d322d4f2-73f4-40dd-a66e-7564e032f72b\") " pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.633560 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d322d4f2-73f4-40dd-a66e-7564e032f72b-catalog-content\") pod \"certified-operators-2jkl4\" (UID: \"d322d4f2-73f4-40dd-a66e-7564e032f72b\") " pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.654093 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz498\" (UniqueName: \"kubernetes.io/projected/d322d4f2-73f4-40dd-a66e-7564e032f72b-kube-api-access-zz498\") pod \"certified-operators-2jkl4\" (UID: \"d322d4f2-73f4-40dd-a66e-7564e032f72b\") " pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.835556 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.953373 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.957078 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 02 03:00:38 crc kubenswrapper[4775]: I1002 03:00:38.964968 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.140101 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mshvr\" (UniqueName: \"kubernetes.io/projected/24c57c9c-fea1-438c-93b9-f1f69eb0de7f-kube-api-access-mshvr\") pod \"mariadb-client-1\" (UID: \"24c57c9c-fea1-438c-93b9-f1f69eb0de7f\") " pod="openstack/mariadb-client-1" Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.142632 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb794f10223592253e53413ce4076ddfdb5fc0165bccfb837c0bbc257983d764" Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.142700 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.241751 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mshvr\" (UniqueName: \"kubernetes.io/projected/24c57c9c-fea1-438c-93b9-f1f69eb0de7f-kube-api-access-mshvr\") pod \"mariadb-client-1\" (UID: \"24c57c9c-fea1-438c-93b9-f1f69eb0de7f\") " pod="openstack/mariadb-client-1" Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.264926 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mshvr\" (UniqueName: \"kubernetes.io/projected/24c57c9c-fea1-438c-93b9-f1f69eb0de7f-kube-api-access-mshvr\") pod \"mariadb-client-1\" (UID: \"24c57c9c-fea1-438c-93b9-f1f69eb0de7f\") " pod="openstack/mariadb-client-1" Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.288805 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.353112 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2jkl4"] Oct 02 03:00:39 crc kubenswrapper[4775]: W1002 03:00:39.357776 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd322d4f2_73f4_40dd_a66e_7564e032f72b.slice/crio-efee1b8ead72bf304788f37114d55415358250f1e0ec295b9ee83a84fe330c38 WatchSource:0}: Error finding container efee1b8ead72bf304788f37114d55415358250f1e0ec295b9ee83a84fe330c38: Status 404 returned error can't find the container with id efee1b8ead72bf304788f37114d55415358250f1e0ec295b9ee83a84fe330c38 Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.606708 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Oct 02 03:00:39 crc kubenswrapper[4775]: W1002 03:00:39.634353 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24c57c9c_fea1_438c_93b9_f1f69eb0de7f.slice/crio-c299a44477e0f32b2205237347c33bce771893636dc97b7693089f70d9aa1ac8 WatchSource:0}: Error finding container c299a44477e0f32b2205237347c33bce771893636dc97b7693089f70d9aa1ac8: Status 404 returned error can't find the container with id c299a44477e0f32b2205237347c33bce771893636dc97b7693089f70d9aa1ac8 Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.782349 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0089837e-94bd-4547-84af-eba2701b5c47" path="/var/lib/kubelet/pods/0089837e-94bd-4547-84af-eba2701b5c47/volumes" Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.911561 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2jlt5"] Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.917993 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.933641 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2jlt5"] Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.962814 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3be6e12-351e-47db-9d9e-d96a9e143ef6-utilities\") pod \"redhat-operators-2jlt5\" (UID: \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\") " pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.963074 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xklj\" (UniqueName: \"kubernetes.io/projected/d3be6e12-351e-47db-9d9e-d96a9e143ef6-kube-api-access-4xklj\") pod \"redhat-operators-2jlt5\" (UID: \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\") " pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:39 crc kubenswrapper[4775]: I1002 03:00:39.963177 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3be6e12-351e-47db-9d9e-d96a9e143ef6-catalog-content\") pod \"redhat-operators-2jlt5\" (UID: \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\") " pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.064494 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3be6e12-351e-47db-9d9e-d96a9e143ef6-catalog-content\") pod \"redhat-operators-2jlt5\" (UID: \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\") " pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.065086 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3be6e12-351e-47db-9d9e-d96a9e143ef6-utilities\") pod \"redhat-operators-2jlt5\" (UID: \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\") " pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.065256 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3be6e12-351e-47db-9d9e-d96a9e143ef6-catalog-content\") pod \"redhat-operators-2jlt5\" (UID: \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\") " pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.065283 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xklj\" (UniqueName: \"kubernetes.io/projected/d3be6e12-351e-47db-9d9e-d96a9e143ef6-kube-api-access-4xklj\") pod \"redhat-operators-2jlt5\" (UID: \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\") " pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.065478 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3be6e12-351e-47db-9d9e-d96a9e143ef6-utilities\") pod \"redhat-operators-2jlt5\" (UID: \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\") " pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.090362 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xklj\" (UniqueName: \"kubernetes.io/projected/d3be6e12-351e-47db-9d9e-d96a9e143ef6-kube-api-access-4xklj\") pod \"redhat-operators-2jlt5\" (UID: \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\") " pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.150926 4775 generic.go:334] "Generic (PLEG): container finished" podID="24c57c9c-fea1-438c-93b9-f1f69eb0de7f" containerID="a78579927e4fa41fb61c66f826c632d566377772e79a9f0768fba7d188f74e45" exitCode=0 Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.150985 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"24c57c9c-fea1-438c-93b9-f1f69eb0de7f","Type":"ContainerDied","Data":"a78579927e4fa41fb61c66f826c632d566377772e79a9f0768fba7d188f74e45"} Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.151028 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"24c57c9c-fea1-438c-93b9-f1f69eb0de7f","Type":"ContainerStarted","Data":"c299a44477e0f32b2205237347c33bce771893636dc97b7693089f70d9aa1ac8"} Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.152710 4775 generic.go:334] "Generic (PLEG): container finished" podID="d322d4f2-73f4-40dd-a66e-7564e032f72b" containerID="369cc334c79e0621d07ce7bded2932fb04244a920c2671e9b5db1d96bb9a16e3" exitCode=0 Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.152743 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2jkl4" event={"ID":"d322d4f2-73f4-40dd-a66e-7564e032f72b","Type":"ContainerDied","Data":"369cc334c79e0621d07ce7bded2932fb04244a920c2671e9b5db1d96bb9a16e3"} Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.152763 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2jkl4" event={"ID":"d322d4f2-73f4-40dd-a66e-7564e032f72b","Type":"ContainerStarted","Data":"efee1b8ead72bf304788f37114d55415358250f1e0ec295b9ee83a84fe330c38"} Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.154687 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.239873 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.654211 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2jlt5"] Oct 02 03:00:40 crc kubenswrapper[4775]: W1002 03:00:40.656673 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3be6e12_351e_47db_9d9e_d96a9e143ef6.slice/crio-7f68ff38019800d53c87ea46b2a27c322e26048bca2799bda56d4d7de2195433 WatchSource:0}: Error finding container 7f68ff38019800d53c87ea46b2a27c322e26048bca2799bda56d4d7de2195433: Status 404 returned error can't find the container with id 7f68ff38019800d53c87ea46b2a27c322e26048bca2799bda56d4d7de2195433 Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.898097 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cxrjw"] Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.899771 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.909627 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cxrjw"] Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.984036 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwmm2\" (UniqueName: \"kubernetes.io/projected/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-kube-api-access-fwmm2\") pod \"community-operators-cxrjw\" (UID: \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\") " pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.984166 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-utilities\") pod \"community-operators-cxrjw\" (UID: \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\") " pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:40 crc kubenswrapper[4775]: I1002 03:00:40.984351 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-catalog-content\") pod \"community-operators-cxrjw\" (UID: \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\") " pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.085628 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-utilities\") pod \"community-operators-cxrjw\" (UID: \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\") " pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.085704 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-catalog-content\") pod \"community-operators-cxrjw\" (UID: \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\") " pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.085739 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwmm2\" (UniqueName: \"kubernetes.io/projected/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-kube-api-access-fwmm2\") pod \"community-operators-cxrjw\" (UID: \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\") " pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.086328 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-catalog-content\") pod \"community-operators-cxrjw\" (UID: \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\") " pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.086406 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-utilities\") pod \"community-operators-cxrjw\" (UID: \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\") " pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.104341 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwmm2\" (UniqueName: \"kubernetes.io/projected/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-kube-api-access-fwmm2\") pod \"community-operators-cxrjw\" (UID: \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\") " pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.161867 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2jkl4" event={"ID":"d322d4f2-73f4-40dd-a66e-7564e032f72b","Type":"ContainerStarted","Data":"fad2bc6324d014ce2d0b23509b04be74261b40e138fd6a2e46369347cb669b7f"} Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.163629 4775 generic.go:334] "Generic (PLEG): container finished" podID="d3be6e12-351e-47db-9d9e-d96a9e143ef6" containerID="d5a3a09d39017a660f1c9fb9f1103ed73e2baf0aa929614d7a97d0f316599d53" exitCode=0 Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.163679 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2jlt5" event={"ID":"d3be6e12-351e-47db-9d9e-d96a9e143ef6","Type":"ContainerDied","Data":"d5a3a09d39017a660f1c9fb9f1103ed73e2baf0aa929614d7a97d0f316599d53"} Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.163706 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2jlt5" event={"ID":"d3be6e12-351e-47db-9d9e-d96a9e143ef6","Type":"ContainerStarted","Data":"7f68ff38019800d53c87ea46b2a27c322e26048bca2799bda56d4d7de2195433"} Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.216878 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.586056 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cxrjw"] Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.596364 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.612096 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_24c57c9c-fea1-438c-93b9-f1f69eb0de7f/mariadb-client-1/0.log" Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.638270 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.642590 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.693609 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mshvr\" (UniqueName: \"kubernetes.io/projected/24c57c9c-fea1-438c-93b9-f1f69eb0de7f-kube-api-access-mshvr\") pod \"24c57c9c-fea1-438c-93b9-f1f69eb0de7f\" (UID: \"24c57c9c-fea1-438c-93b9-f1f69eb0de7f\") " Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.698761 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24c57c9c-fea1-438c-93b9-f1f69eb0de7f-kube-api-access-mshvr" (OuterVolumeSpecName: "kube-api-access-mshvr") pod "24c57c9c-fea1-438c-93b9-f1f69eb0de7f" (UID: "24c57c9c-fea1-438c-93b9-f1f69eb0de7f"). InnerVolumeSpecName "kube-api-access-mshvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.774283 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24c57c9c-fea1-438c-93b9-f1f69eb0de7f" path="/var/lib/kubelet/pods/24c57c9c-fea1-438c-93b9-f1f69eb0de7f/volumes" Oct 02 03:00:41 crc kubenswrapper[4775]: I1002 03:00:41.795328 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mshvr\" (UniqueName: \"kubernetes.io/projected/24c57c9c-fea1-438c-93b9-f1f69eb0de7f-kube-api-access-mshvr\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.181196 4775 generic.go:334] "Generic (PLEG): container finished" podID="d322d4f2-73f4-40dd-a66e-7564e032f72b" containerID="fad2bc6324d014ce2d0b23509b04be74261b40e138fd6a2e46369347cb669b7f" exitCode=0 Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.181331 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2jkl4" event={"ID":"d322d4f2-73f4-40dd-a66e-7564e032f72b","Type":"ContainerDied","Data":"fad2bc6324d014ce2d0b23509b04be74261b40e138fd6a2e46369347cb669b7f"} Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.193886 4775 scope.go:117] "RemoveContainer" containerID="a78579927e4fa41fb61c66f826c632d566377772e79a9f0768fba7d188f74e45" Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.194045 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.198554 4775 generic.go:334] "Generic (PLEG): container finished" podID="4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" containerID="e28b87cf4dd9f71a00129377eb8ece7028d69ce260c4fde229a09eb66fc60e50" exitCode=0 Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.198597 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxrjw" event={"ID":"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1","Type":"ContainerDied","Data":"e28b87cf4dd9f71a00129377eb8ece7028d69ce260c4fde229a09eb66fc60e50"} Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.198623 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxrjw" event={"ID":"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1","Type":"ContainerStarted","Data":"1ad34586dfbd8dd2377af67d64e519fc0648674b6eda06e2d267f856dd84d6fd"} Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.227741 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Oct 02 03:00:42 crc kubenswrapper[4775]: E1002 03:00:42.228150 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24c57c9c-fea1-438c-93b9-f1f69eb0de7f" containerName="mariadb-client-1" Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.228171 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="24c57c9c-fea1-438c-93b9-f1f69eb0de7f" containerName="mariadb-client-1" Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.228389 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="24c57c9c-fea1-438c-93b9-f1f69eb0de7f" containerName="mariadb-client-1" Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.238792 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.238909 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.241792 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-tnzbn" Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.304414 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zkvw\" (UniqueName: \"kubernetes.io/projected/4a215510-e147-48ea-bf82-5c6fa838c2ca-kube-api-access-6zkvw\") pod \"mariadb-client-4-default\" (UID: \"4a215510-e147-48ea-bf82-5c6fa838c2ca\") " pod="openstack/mariadb-client-4-default" Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.407920 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zkvw\" (UniqueName: \"kubernetes.io/projected/4a215510-e147-48ea-bf82-5c6fa838c2ca-kube-api-access-6zkvw\") pod \"mariadb-client-4-default\" (UID: \"4a215510-e147-48ea-bf82-5c6fa838c2ca\") " pod="openstack/mariadb-client-4-default" Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.440018 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zkvw\" (UniqueName: \"kubernetes.io/projected/4a215510-e147-48ea-bf82-5c6fa838c2ca-kube-api-access-6zkvw\") pod \"mariadb-client-4-default\" (UID: \"4a215510-e147-48ea-bf82-5c6fa838c2ca\") " pod="openstack/mariadb-client-4-default" Oct 02 03:00:42 crc kubenswrapper[4775]: I1002 03:00:42.654506 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 02 03:00:43 crc kubenswrapper[4775]: I1002 03:00:43.208555 4775 generic.go:334] "Generic (PLEG): container finished" podID="d3be6e12-351e-47db-9d9e-d96a9e143ef6" containerID="529995ec6d32298785ad9d318d0d7794b310e251aaa410bebbebc4ead9b2b8c1" exitCode=0 Oct 02 03:00:43 crc kubenswrapper[4775]: I1002 03:00:43.208617 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2jlt5" event={"ID":"d3be6e12-351e-47db-9d9e-d96a9e143ef6","Type":"ContainerDied","Data":"529995ec6d32298785ad9d318d0d7794b310e251aaa410bebbebc4ead9b2b8c1"} Oct 02 03:00:43 crc kubenswrapper[4775]: I1002 03:00:43.214908 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxrjw" event={"ID":"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1","Type":"ContainerStarted","Data":"c5666dce1fb7c056d1a7387853a4e48e4e3527aff89f3478eb34c59f806a717f"} Oct 02 03:00:43 crc kubenswrapper[4775]: I1002 03:00:43.224103 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2jkl4" event={"ID":"d322d4f2-73f4-40dd-a66e-7564e032f72b","Type":"ContainerStarted","Data":"1b5206f9745d26c88b7308321a3826fc8f589314f6f280977a90d726c96ea4f3"} Oct 02 03:00:43 crc kubenswrapper[4775]: I1002 03:00:43.279178 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2jkl4" podStartSLOduration=2.717880961 podStartE2EDuration="5.279163194s" podCreationTimestamp="2025-10-02 03:00:38 +0000 UTC" firstStartedPulling="2025-10-02 03:00:40.154414904 +0000 UTC m=+4777.321158954" lastFinishedPulling="2025-10-02 03:00:42.715697137 +0000 UTC m=+4779.882441187" observedRunningTime="2025-10-02 03:00:43.267477639 +0000 UTC m=+4780.434221689" watchObservedRunningTime="2025-10-02 03:00:43.279163194 +0000 UTC m=+4780.445907224" Oct 02 03:00:43 crc kubenswrapper[4775]: I1002 03:00:43.284383 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 02 03:00:44 crc kubenswrapper[4775]: I1002 03:00:44.236549 4775 generic.go:334] "Generic (PLEG): container finished" podID="4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" containerID="c5666dce1fb7c056d1a7387853a4e48e4e3527aff89f3478eb34c59f806a717f" exitCode=0 Oct 02 03:00:44 crc kubenswrapper[4775]: I1002 03:00:44.236600 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxrjw" event={"ID":"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1","Type":"ContainerDied","Data":"c5666dce1fb7c056d1a7387853a4e48e4e3527aff89f3478eb34c59f806a717f"} Oct 02 03:00:44 crc kubenswrapper[4775]: I1002 03:00:44.239104 4775 generic.go:334] "Generic (PLEG): container finished" podID="4a215510-e147-48ea-bf82-5c6fa838c2ca" containerID="b3f3e8c29ca9a4418cbddaab1020b56e9ee21fef6673169945063d30f1568f86" exitCode=0 Oct 02 03:00:44 crc kubenswrapper[4775]: I1002 03:00:44.239205 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"4a215510-e147-48ea-bf82-5c6fa838c2ca","Type":"ContainerDied","Data":"b3f3e8c29ca9a4418cbddaab1020b56e9ee21fef6673169945063d30f1568f86"} Oct 02 03:00:44 crc kubenswrapper[4775]: I1002 03:00:44.239297 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"4a215510-e147-48ea-bf82-5c6fa838c2ca","Type":"ContainerStarted","Data":"2099a61b381fd80c2d2122eff24eeb7c053f668a5aa615ee2bea915ef3fcdf73"} Oct 02 03:00:44 crc kubenswrapper[4775]: I1002 03:00:44.245224 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2jlt5" event={"ID":"d3be6e12-351e-47db-9d9e-d96a9e143ef6","Type":"ContainerStarted","Data":"4a40e3511f0b9abfa80345136e6fa07d2ba77e34f51aa8fc2f899d1e7930db98"} Oct 02 03:00:44 crc kubenswrapper[4775]: I1002 03:00:44.278270 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2jlt5" podStartSLOduration=2.836669509 podStartE2EDuration="5.278244781s" podCreationTimestamp="2025-10-02 03:00:39 +0000 UTC" firstStartedPulling="2025-10-02 03:00:41.164866698 +0000 UTC m=+4778.331610738" lastFinishedPulling="2025-10-02 03:00:43.60644197 +0000 UTC m=+4780.773186010" observedRunningTime="2025-10-02 03:00:44.273376274 +0000 UTC m=+4781.440120314" watchObservedRunningTime="2025-10-02 03:00:44.278244781 +0000 UTC m=+4781.444988851" Oct 02 03:00:45 crc kubenswrapper[4775]: I1002 03:00:45.257645 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxrjw" event={"ID":"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1","Type":"ContainerStarted","Data":"e747dd8cb98aceb9b2d1d0df2a4086059c90a2a0295d8461ebcfc1dcd9896bc4"} Oct 02 03:00:45 crc kubenswrapper[4775]: I1002 03:00:45.278182 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cxrjw" podStartSLOduration=2.703505769 podStartE2EDuration="5.278163841s" podCreationTimestamp="2025-10-02 03:00:40 +0000 UTC" firstStartedPulling="2025-10-02 03:00:42.203992201 +0000 UTC m=+4779.370736281" lastFinishedPulling="2025-10-02 03:00:44.778650303 +0000 UTC m=+4781.945394353" observedRunningTime="2025-10-02 03:00:45.274949397 +0000 UTC m=+4782.441693457" watchObservedRunningTime="2025-10-02 03:00:45.278163841 +0000 UTC m=+4782.444907881" Oct 02 03:00:45 crc kubenswrapper[4775]: I1002 03:00:45.646481 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 02 03:00:45 crc kubenswrapper[4775]: I1002 03:00:45.666835 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_4a215510-e147-48ea-bf82-5c6fa838c2ca/mariadb-client-4-default/0.log" Oct 02 03:00:45 crc kubenswrapper[4775]: I1002 03:00:45.700499 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 02 03:00:45 crc kubenswrapper[4775]: I1002 03:00:45.706755 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Oct 02 03:00:45 crc kubenswrapper[4775]: I1002 03:00:45.794722 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zkvw\" (UniqueName: \"kubernetes.io/projected/4a215510-e147-48ea-bf82-5c6fa838c2ca-kube-api-access-6zkvw\") pod \"4a215510-e147-48ea-bf82-5c6fa838c2ca\" (UID: \"4a215510-e147-48ea-bf82-5c6fa838c2ca\") " Oct 02 03:00:45 crc kubenswrapper[4775]: I1002 03:00:45.803772 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a215510-e147-48ea-bf82-5c6fa838c2ca-kube-api-access-6zkvw" (OuterVolumeSpecName: "kube-api-access-6zkvw") pod "4a215510-e147-48ea-bf82-5c6fa838c2ca" (UID: "4a215510-e147-48ea-bf82-5c6fa838c2ca"). InnerVolumeSpecName "kube-api-access-6zkvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:00:45 crc kubenswrapper[4775]: I1002 03:00:45.896631 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zkvw\" (UniqueName: \"kubernetes.io/projected/4a215510-e147-48ea-bf82-5c6fa838c2ca-kube-api-access-6zkvw\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:46 crc kubenswrapper[4775]: I1002 03:00:46.268508 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Oct 02 03:00:46 crc kubenswrapper[4775]: I1002 03:00:46.268508 4775 scope.go:117] "RemoveContainer" containerID="b3f3e8c29ca9a4418cbddaab1020b56e9ee21fef6673169945063d30f1568f86" Oct 02 03:00:47 crc kubenswrapper[4775]: I1002 03:00:47.779598 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a215510-e147-48ea-bf82-5c6fa838c2ca" path="/var/lib/kubelet/pods/4a215510-e147-48ea-bf82-5c6fa838c2ca/volumes" Oct 02 03:00:48 crc kubenswrapper[4775]: E1002 03:00:48.676439 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50401de0_705a_4de5_b211_56688c040651.slice\": RecentStats: unable to find data in memory cache]" Oct 02 03:00:48 crc kubenswrapper[4775]: I1002 03:00:48.836853 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:48 crc kubenswrapper[4775]: I1002 03:00:48.837712 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:48 crc kubenswrapper[4775]: I1002 03:00:48.915727 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:49 crc kubenswrapper[4775]: I1002 03:00:49.390717 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.053262 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Oct 02 03:00:50 crc kubenswrapper[4775]: E1002 03:00:50.054131 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a215510-e147-48ea-bf82-5c6fa838c2ca" containerName="mariadb-client-4-default" Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.054157 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a215510-e147-48ea-bf82-5c6fa838c2ca" containerName="mariadb-client-4-default" Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.054453 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a215510-e147-48ea-bf82-5c6fa838c2ca" containerName="mariadb-client-4-default" Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.055290 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.058595 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-tnzbn" Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.061117 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.164524 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgf89\" (UniqueName: \"kubernetes.io/projected/4d617192-6d46-4775-a34c-2fae311c2dff-kube-api-access-qgf89\") pod \"mariadb-client-5-default\" (UID: \"4d617192-6d46-4775-a34c-2fae311c2dff\") " pod="openstack/mariadb-client-5-default" Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.240738 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.240805 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.266480 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgf89\" (UniqueName: \"kubernetes.io/projected/4d617192-6d46-4775-a34c-2fae311c2dff-kube-api-access-qgf89\") pod \"mariadb-client-5-default\" (UID: \"4d617192-6d46-4775-a34c-2fae311c2dff\") " pod="openstack/mariadb-client-5-default" Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.299794 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.306002 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgf89\" (UniqueName: \"kubernetes.io/projected/4d617192-6d46-4775-a34c-2fae311c2dff-kube-api-access-qgf89\") pod \"mariadb-client-5-default\" (UID: \"4d617192-6d46-4775-a34c-2fae311c2dff\") " pod="openstack/mariadb-client-5-default" Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.372907 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.382485 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.782687 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 02 03:00:50 crc kubenswrapper[4775]: I1002 03:00:50.911444 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2jkl4"] Oct 02 03:00:51 crc kubenswrapper[4775]: I1002 03:00:51.217914 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:51 crc kubenswrapper[4775]: I1002 03:00:51.218082 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:51 crc kubenswrapper[4775]: I1002 03:00:51.298076 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:51 crc kubenswrapper[4775]: I1002 03:00:51.333365 4775 generic.go:334] "Generic (PLEG): container finished" podID="4d617192-6d46-4775-a34c-2fae311c2dff" containerID="af0ce6373af58eec3da363e858ca3f553944585860a600e6b0297212ab9b3264" exitCode=0 Oct 02 03:00:51 crc kubenswrapper[4775]: I1002 03:00:51.333465 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"4d617192-6d46-4775-a34c-2fae311c2dff","Type":"ContainerDied","Data":"af0ce6373af58eec3da363e858ca3f553944585860a600e6b0297212ab9b3264"} Oct 02 03:00:51 crc kubenswrapper[4775]: I1002 03:00:51.333549 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"4d617192-6d46-4775-a34c-2fae311c2dff","Type":"ContainerStarted","Data":"92d73e96e9f1a8d7becc9a318d6b5a331728710fd5846af1340be1ee1cf9bad3"} Oct 02 03:00:51 crc kubenswrapper[4775]: I1002 03:00:51.334154 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2jkl4" podUID="d322d4f2-73f4-40dd-a66e-7564e032f72b" containerName="registry-server" containerID="cri-o://1b5206f9745d26c88b7308321a3826fc8f589314f6f280977a90d726c96ea4f3" gracePeriod=2 Oct 02 03:00:51 crc kubenswrapper[4775]: I1002 03:00:51.413941 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.344007 4775 generic.go:334] "Generic (PLEG): container finished" podID="d322d4f2-73f4-40dd-a66e-7564e032f72b" containerID="1b5206f9745d26c88b7308321a3826fc8f589314f6f280977a90d726c96ea4f3" exitCode=0 Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.344091 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2jkl4" event={"ID":"d322d4f2-73f4-40dd-a66e-7564e032f72b","Type":"ContainerDied","Data":"1b5206f9745d26c88b7308321a3826fc8f589314f6f280977a90d726c96ea4f3"} Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.344168 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2jkl4" event={"ID":"d322d4f2-73f4-40dd-a66e-7564e032f72b","Type":"ContainerDied","Data":"efee1b8ead72bf304788f37114d55415358250f1e0ec295b9ee83a84fe330c38"} Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.344196 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efee1b8ead72bf304788f37114d55415358250f1e0ec295b9ee83a84fe330c38" Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.357556 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.508643 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d322d4f2-73f4-40dd-a66e-7564e032f72b-utilities\") pod \"d322d4f2-73f4-40dd-a66e-7564e032f72b\" (UID: \"d322d4f2-73f4-40dd-a66e-7564e032f72b\") " Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.509196 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d322d4f2-73f4-40dd-a66e-7564e032f72b-catalog-content\") pod \"d322d4f2-73f4-40dd-a66e-7564e032f72b\" (UID: \"d322d4f2-73f4-40dd-a66e-7564e032f72b\") " Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.509285 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz498\" (UniqueName: \"kubernetes.io/projected/d322d4f2-73f4-40dd-a66e-7564e032f72b-kube-api-access-zz498\") pod \"d322d4f2-73f4-40dd-a66e-7564e032f72b\" (UID: \"d322d4f2-73f4-40dd-a66e-7564e032f72b\") " Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.510097 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d322d4f2-73f4-40dd-a66e-7564e032f72b-utilities" (OuterVolumeSpecName: "utilities") pod "d322d4f2-73f4-40dd-a66e-7564e032f72b" (UID: "d322d4f2-73f4-40dd-a66e-7564e032f72b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.527625 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d322d4f2-73f4-40dd-a66e-7564e032f72b-kube-api-access-zz498" (OuterVolumeSpecName: "kube-api-access-zz498") pod "d322d4f2-73f4-40dd-a66e-7564e032f72b" (UID: "d322d4f2-73f4-40dd-a66e-7564e032f72b"). InnerVolumeSpecName "kube-api-access-zz498". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.585336 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d322d4f2-73f4-40dd-a66e-7564e032f72b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d322d4f2-73f4-40dd-a66e-7564e032f72b" (UID: "d322d4f2-73f4-40dd-a66e-7564e032f72b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.611367 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d322d4f2-73f4-40dd-a66e-7564e032f72b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.611403 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d322d4f2-73f4-40dd-a66e-7564e032f72b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.611419 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz498\" (UniqueName: \"kubernetes.io/projected/d322d4f2-73f4-40dd-a66e-7564e032f72b-kube-api-access-zz498\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.694655 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2jlt5"] Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.694907 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2jlt5" podUID="d3be6e12-351e-47db-9d9e-d96a9e143ef6" containerName="registry-server" containerID="cri-o://4a40e3511f0b9abfa80345136e6fa07d2ba77e34f51aa8fc2f899d1e7930db98" gracePeriod=2 Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.917615 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.942265 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_4d617192-6d46-4775-a34c-2fae311c2dff/mariadb-client-5-default/0.log" Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.979801 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 02 03:00:52 crc kubenswrapper[4775]: I1002 03:00:52.990299 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.017376 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgf89\" (UniqueName: \"kubernetes.io/projected/4d617192-6d46-4775-a34c-2fae311c2dff-kube-api-access-qgf89\") pod \"4d617192-6d46-4775-a34c-2fae311c2dff\" (UID: \"4d617192-6d46-4775-a34c-2fae311c2dff\") " Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.032265 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d617192-6d46-4775-a34c-2fae311c2dff-kube-api-access-qgf89" (OuterVolumeSpecName: "kube-api-access-qgf89") pod "4d617192-6d46-4775-a34c-2fae311c2dff" (UID: "4d617192-6d46-4775-a34c-2fae311c2dff"). InnerVolumeSpecName "kube-api-access-qgf89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.122679 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgf89\" (UniqueName: \"kubernetes.io/projected/4d617192-6d46-4775-a34c-2fae311c2dff-kube-api-access-qgf89\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.156738 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.183540 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Oct 02 03:00:53 crc kubenswrapper[4775]: E1002 03:00:53.183994 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d617192-6d46-4775-a34c-2fae311c2dff" containerName="mariadb-client-5-default" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.184010 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d617192-6d46-4775-a34c-2fae311c2dff" containerName="mariadb-client-5-default" Oct 02 03:00:53 crc kubenswrapper[4775]: E1002 03:00:53.184025 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d322d4f2-73f4-40dd-a66e-7564e032f72b" containerName="registry-server" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.184035 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d322d4f2-73f4-40dd-a66e-7564e032f72b" containerName="registry-server" Oct 02 03:00:53 crc kubenswrapper[4775]: E1002 03:00:53.184067 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3be6e12-351e-47db-9d9e-d96a9e143ef6" containerName="extract-utilities" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.184077 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3be6e12-351e-47db-9d9e-d96a9e143ef6" containerName="extract-utilities" Oct 02 03:00:53 crc kubenswrapper[4775]: E1002 03:00:53.184105 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d322d4f2-73f4-40dd-a66e-7564e032f72b" containerName="extract-utilities" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.184113 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d322d4f2-73f4-40dd-a66e-7564e032f72b" containerName="extract-utilities" Oct 02 03:00:53 crc kubenswrapper[4775]: E1002 03:00:53.184137 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3be6e12-351e-47db-9d9e-d96a9e143ef6" containerName="registry-server" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.184145 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3be6e12-351e-47db-9d9e-d96a9e143ef6" containerName="registry-server" Oct 02 03:00:53 crc kubenswrapper[4775]: E1002 03:00:53.184155 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d322d4f2-73f4-40dd-a66e-7564e032f72b" containerName="extract-content" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.184162 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d322d4f2-73f4-40dd-a66e-7564e032f72b" containerName="extract-content" Oct 02 03:00:53 crc kubenswrapper[4775]: E1002 03:00:53.184176 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3be6e12-351e-47db-9d9e-d96a9e143ef6" containerName="extract-content" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.184185 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3be6e12-351e-47db-9d9e-d96a9e143ef6" containerName="extract-content" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.184553 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d322d4f2-73f4-40dd-a66e-7564e032f72b" containerName="registry-server" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.184571 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d617192-6d46-4775-a34c-2fae311c2dff" containerName="mariadb-client-5-default" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.184584 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3be6e12-351e-47db-9d9e-d96a9e143ef6" containerName="registry-server" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.185174 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.196897 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.326493 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3be6e12-351e-47db-9d9e-d96a9e143ef6-utilities\") pod \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\" (UID: \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\") " Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.327004 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xklj\" (UniqueName: \"kubernetes.io/projected/d3be6e12-351e-47db-9d9e-d96a9e143ef6-kube-api-access-4xklj\") pod \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\" (UID: \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\") " Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.327220 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3be6e12-351e-47db-9d9e-d96a9e143ef6-catalog-content\") pod \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\" (UID: \"d3be6e12-351e-47db-9d9e-d96a9e143ef6\") " Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.327877 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3be6e12-351e-47db-9d9e-d96a9e143ef6-utilities" (OuterVolumeSpecName: "utilities") pod "d3be6e12-351e-47db-9d9e-d96a9e143ef6" (UID: "d3be6e12-351e-47db-9d9e-d96a9e143ef6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.328137 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bphj\" (UniqueName: \"kubernetes.io/projected/0081ac16-e0bd-494d-90f4-d5d17e514e74-kube-api-access-8bphj\") pod \"mariadb-client-6-default\" (UID: \"0081ac16-e0bd-494d-90f4-d5d17e514e74\") " pod="openstack/mariadb-client-6-default" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.328510 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3be6e12-351e-47db-9d9e-d96a9e143ef6-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.335018 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3be6e12-351e-47db-9d9e-d96a9e143ef6-kube-api-access-4xklj" (OuterVolumeSpecName: "kube-api-access-4xklj") pod "d3be6e12-351e-47db-9d9e-d96a9e143ef6" (UID: "d3be6e12-351e-47db-9d9e-d96a9e143ef6"). InnerVolumeSpecName "kube-api-access-4xklj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.355788 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92d73e96e9f1a8d7becc9a318d6b5a331728710fd5846af1340be1ee1cf9bad3" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.355875 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.359415 4775 generic.go:334] "Generic (PLEG): container finished" podID="d3be6e12-351e-47db-9d9e-d96a9e143ef6" containerID="4a40e3511f0b9abfa80345136e6fa07d2ba77e34f51aa8fc2f899d1e7930db98" exitCode=0 Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.359519 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2jlt5" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.360101 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2jkl4" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.359515 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2jlt5" event={"ID":"d3be6e12-351e-47db-9d9e-d96a9e143ef6","Type":"ContainerDied","Data":"4a40e3511f0b9abfa80345136e6fa07d2ba77e34f51aa8fc2f899d1e7930db98"} Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.360195 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2jlt5" event={"ID":"d3be6e12-351e-47db-9d9e-d96a9e143ef6","Type":"ContainerDied","Data":"7f68ff38019800d53c87ea46b2a27c322e26048bca2799bda56d4d7de2195433"} Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.360234 4775 scope.go:117] "RemoveContainer" containerID="4a40e3511f0b9abfa80345136e6fa07d2ba77e34f51aa8fc2f899d1e7930db98" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.396199 4775 scope.go:117] "RemoveContainer" containerID="529995ec6d32298785ad9d318d0d7794b310e251aaa410bebbebc4ead9b2b8c1" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.419635 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2jkl4"] Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.427333 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2jkl4"] Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.430125 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bphj\" (UniqueName: \"kubernetes.io/projected/0081ac16-e0bd-494d-90f4-d5d17e514e74-kube-api-access-8bphj\") pod \"mariadb-client-6-default\" (UID: \"0081ac16-e0bd-494d-90f4-d5d17e514e74\") " pod="openstack/mariadb-client-6-default" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.430343 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xklj\" (UniqueName: \"kubernetes.io/projected/d3be6e12-351e-47db-9d9e-d96a9e143ef6-kube-api-access-4xklj\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.446440 4775 scope.go:117] "RemoveContainer" containerID="d5a3a09d39017a660f1c9fb9f1103ed73e2baf0aa929614d7a97d0f316599d53" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.452413 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bphj\" (UniqueName: \"kubernetes.io/projected/0081ac16-e0bd-494d-90f4-d5d17e514e74-kube-api-access-8bphj\") pod \"mariadb-client-6-default\" (UID: \"0081ac16-e0bd-494d-90f4-d5d17e514e74\") " pod="openstack/mariadb-client-6-default" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.466470 4775 scope.go:117] "RemoveContainer" containerID="4a40e3511f0b9abfa80345136e6fa07d2ba77e34f51aa8fc2f899d1e7930db98" Oct 02 03:00:53 crc kubenswrapper[4775]: E1002 03:00:53.467069 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a40e3511f0b9abfa80345136e6fa07d2ba77e34f51aa8fc2f899d1e7930db98\": container with ID starting with 4a40e3511f0b9abfa80345136e6fa07d2ba77e34f51aa8fc2f899d1e7930db98 not found: ID does not exist" containerID="4a40e3511f0b9abfa80345136e6fa07d2ba77e34f51aa8fc2f899d1e7930db98" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.467118 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a40e3511f0b9abfa80345136e6fa07d2ba77e34f51aa8fc2f899d1e7930db98"} err="failed to get container status \"4a40e3511f0b9abfa80345136e6fa07d2ba77e34f51aa8fc2f899d1e7930db98\": rpc error: code = NotFound desc = could not find container \"4a40e3511f0b9abfa80345136e6fa07d2ba77e34f51aa8fc2f899d1e7930db98\": container with ID starting with 4a40e3511f0b9abfa80345136e6fa07d2ba77e34f51aa8fc2f899d1e7930db98 not found: ID does not exist" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.467150 4775 scope.go:117] "RemoveContainer" containerID="529995ec6d32298785ad9d318d0d7794b310e251aaa410bebbebc4ead9b2b8c1" Oct 02 03:00:53 crc kubenswrapper[4775]: E1002 03:00:53.467643 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"529995ec6d32298785ad9d318d0d7794b310e251aaa410bebbebc4ead9b2b8c1\": container with ID starting with 529995ec6d32298785ad9d318d0d7794b310e251aaa410bebbebc4ead9b2b8c1 not found: ID does not exist" containerID="529995ec6d32298785ad9d318d0d7794b310e251aaa410bebbebc4ead9b2b8c1" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.467679 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"529995ec6d32298785ad9d318d0d7794b310e251aaa410bebbebc4ead9b2b8c1"} err="failed to get container status \"529995ec6d32298785ad9d318d0d7794b310e251aaa410bebbebc4ead9b2b8c1\": rpc error: code = NotFound desc = could not find container \"529995ec6d32298785ad9d318d0d7794b310e251aaa410bebbebc4ead9b2b8c1\": container with ID starting with 529995ec6d32298785ad9d318d0d7794b310e251aaa410bebbebc4ead9b2b8c1 not found: ID does not exist" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.467696 4775 scope.go:117] "RemoveContainer" containerID="d5a3a09d39017a660f1c9fb9f1103ed73e2baf0aa929614d7a97d0f316599d53" Oct 02 03:00:53 crc kubenswrapper[4775]: E1002 03:00:53.468128 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5a3a09d39017a660f1c9fb9f1103ed73e2baf0aa929614d7a97d0f316599d53\": container with ID starting with d5a3a09d39017a660f1c9fb9f1103ed73e2baf0aa929614d7a97d0f316599d53 not found: ID does not exist" containerID="d5a3a09d39017a660f1c9fb9f1103ed73e2baf0aa929614d7a97d0f316599d53" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.468159 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5a3a09d39017a660f1c9fb9f1103ed73e2baf0aa929614d7a97d0f316599d53"} err="failed to get container status \"d5a3a09d39017a660f1c9fb9f1103ed73e2baf0aa929614d7a97d0f316599d53\": rpc error: code = NotFound desc = could not find container \"d5a3a09d39017a660f1c9fb9f1103ed73e2baf0aa929614d7a97d0f316599d53\": container with ID starting with d5a3a09d39017a660f1c9fb9f1103ed73e2baf0aa929614d7a97d0f316599d53 not found: ID does not exist" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.523251 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.784852 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d617192-6d46-4775-a34c-2fae311c2dff" path="/var/lib/kubelet/pods/4d617192-6d46-4775-a34c-2fae311c2dff/volumes" Oct 02 03:00:53 crc kubenswrapper[4775]: I1002 03:00:53.786233 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d322d4f2-73f4-40dd-a66e-7564e032f72b" path="/var/lib/kubelet/pods/d322d4f2-73f4-40dd-a66e-7564e032f72b/volumes" Oct 02 03:00:54 crc kubenswrapper[4775]: I1002 03:00:54.150546 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 02 03:00:54 crc kubenswrapper[4775]: I1002 03:00:54.388711 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3be6e12-351e-47db-9d9e-d96a9e143ef6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3be6e12-351e-47db-9d9e-d96a9e143ef6" (UID: "d3be6e12-351e-47db-9d9e-d96a9e143ef6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:00:54 crc kubenswrapper[4775]: I1002 03:00:54.447325 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3be6e12-351e-47db-9d9e-d96a9e143ef6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:54 crc kubenswrapper[4775]: I1002 03:00:54.820936 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2jlt5"] Oct 02 03:00:54 crc kubenswrapper[4775]: I1002 03:00:54.831202 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2jlt5"] Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.105306 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cxrjw"] Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.106109 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cxrjw" podUID="4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" containerName="registry-server" containerID="cri-o://e747dd8cb98aceb9b2d1d0df2a4086059c90a2a0295d8461ebcfc1dcd9896bc4" gracePeriod=2 Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.402340 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"0081ac16-e0bd-494d-90f4-d5d17e514e74","Type":"ContainerStarted","Data":"21d0aa3ba21eea81786f45e1aef1eca754fe91c36cb06f3cdf58ccb4edcf1dd7"} Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.405467 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"0081ac16-e0bd-494d-90f4-d5d17e514e74","Type":"ContainerStarted","Data":"eea011ec2093905f514fc4f793893fae8aec938b84c239760edc6bfa2ed5f0a6"} Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.414386 4775 generic.go:334] "Generic (PLEG): container finished" podID="4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" containerID="e747dd8cb98aceb9b2d1d0df2a4086059c90a2a0295d8461ebcfc1dcd9896bc4" exitCode=0 Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.414431 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxrjw" event={"ID":"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1","Type":"ContainerDied","Data":"e747dd8cb98aceb9b2d1d0df2a4086059c90a2a0295d8461ebcfc1dcd9896bc4"} Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.428083 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-6-default" podStartSLOduration=2.42804415 podStartE2EDuration="2.42804415s" podCreationTimestamp="2025-10-02 03:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:00:55.425916714 +0000 UTC m=+4792.592660764" watchObservedRunningTime="2025-10-02 03:00:55.42804415 +0000 UTC m=+4792.594788200" Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.538884 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_0081ac16-e0bd-494d-90f4-d5d17e514e74/mariadb-client-6-default/0.log" Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.575285 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.666546 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-utilities\") pod \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\" (UID: \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\") " Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.666622 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-catalog-content\") pod \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\" (UID: \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\") " Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.666820 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwmm2\" (UniqueName: \"kubernetes.io/projected/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-kube-api-access-fwmm2\") pod \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\" (UID: \"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1\") " Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.668361 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-utilities" (OuterVolumeSpecName: "utilities") pod "4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" (UID: "4a0cb2e1-b09f-4fe2-a005-33d688e71fc1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.680701 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-kube-api-access-fwmm2" (OuterVolumeSpecName: "kube-api-access-fwmm2") pod "4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" (UID: "4a0cb2e1-b09f-4fe2-a005-33d688e71fc1"). InnerVolumeSpecName "kube-api-access-fwmm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.745112 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" (UID: "4a0cb2e1-b09f-4fe2-a005-33d688e71fc1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.770224 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.770290 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.770316 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwmm2\" (UniqueName: \"kubernetes.io/projected/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1-kube-api-access-fwmm2\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:55 crc kubenswrapper[4775]: I1002 03:00:55.784313 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3be6e12-351e-47db-9d9e-d96a9e143ef6" path="/var/lib/kubelet/pods/d3be6e12-351e-47db-9d9e-d96a9e143ef6/volumes" Oct 02 03:00:56 crc kubenswrapper[4775]: I1002 03:00:56.428615 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cxrjw" Oct 02 03:00:56 crc kubenswrapper[4775]: I1002 03:00:56.428618 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cxrjw" event={"ID":"4a0cb2e1-b09f-4fe2-a005-33d688e71fc1","Type":"ContainerDied","Data":"1ad34586dfbd8dd2377af67d64e519fc0648674b6eda06e2d267f856dd84d6fd"} Oct 02 03:00:56 crc kubenswrapper[4775]: I1002 03:00:56.429137 4775 scope.go:117] "RemoveContainer" containerID="e747dd8cb98aceb9b2d1d0df2a4086059c90a2a0295d8461ebcfc1dcd9896bc4" Oct 02 03:00:56 crc kubenswrapper[4775]: I1002 03:00:56.432731 4775 generic.go:334] "Generic (PLEG): container finished" podID="0081ac16-e0bd-494d-90f4-d5d17e514e74" containerID="21d0aa3ba21eea81786f45e1aef1eca754fe91c36cb06f3cdf58ccb4edcf1dd7" exitCode=0 Oct 02 03:00:56 crc kubenswrapper[4775]: I1002 03:00:56.432809 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"0081ac16-e0bd-494d-90f4-d5d17e514e74","Type":"ContainerDied","Data":"21d0aa3ba21eea81786f45e1aef1eca754fe91c36cb06f3cdf58ccb4edcf1dd7"} Oct 02 03:00:56 crc kubenswrapper[4775]: I1002 03:00:56.478388 4775 scope.go:117] "RemoveContainer" containerID="c5666dce1fb7c056d1a7387853a4e48e4e3527aff89f3478eb34c59f806a717f" Oct 02 03:00:56 crc kubenswrapper[4775]: I1002 03:00:56.491379 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cxrjw"] Oct 02 03:00:56 crc kubenswrapper[4775]: I1002 03:00:56.498720 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cxrjw"] Oct 02 03:00:56 crc kubenswrapper[4775]: I1002 03:00:56.514005 4775 scope.go:117] "RemoveContainer" containerID="e28b87cf4dd9f71a00129377eb8ece7028d69ce260c4fde229a09eb66fc60e50" Oct 02 03:00:57 crc kubenswrapper[4775]: I1002 03:00:57.777256 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" path="/var/lib/kubelet/pods/4a0cb2e1-b09f-4fe2-a005-33d688e71fc1/volumes" Oct 02 03:00:57 crc kubenswrapper[4775]: I1002 03:00:57.879093 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 02 03:00:57 crc kubenswrapper[4775]: I1002 03:00:57.921036 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 02 03:00:57 crc kubenswrapper[4775]: I1002 03:00:57.929010 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.008903 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bphj\" (UniqueName: \"kubernetes.io/projected/0081ac16-e0bd-494d-90f4-d5d17e514e74-kube-api-access-8bphj\") pod \"0081ac16-e0bd-494d-90f4-d5d17e514e74\" (UID: \"0081ac16-e0bd-494d-90f4-d5d17e514e74\") " Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.015288 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0081ac16-e0bd-494d-90f4-d5d17e514e74-kube-api-access-8bphj" (OuterVolumeSpecName: "kube-api-access-8bphj") pod "0081ac16-e0bd-494d-90f4-d5d17e514e74" (UID: "0081ac16-e0bd-494d-90f4-d5d17e514e74"). InnerVolumeSpecName "kube-api-access-8bphj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.082611 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Oct 02 03:00:58 crc kubenswrapper[4775]: E1002 03:00:58.083558 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" containerName="registry-server" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.083780 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" containerName="registry-server" Oct 02 03:00:58 crc kubenswrapper[4775]: E1002 03:00:58.084059 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0081ac16-e0bd-494d-90f4-d5d17e514e74" containerName="mariadb-client-6-default" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.084236 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="0081ac16-e0bd-494d-90f4-d5d17e514e74" containerName="mariadb-client-6-default" Oct 02 03:00:58 crc kubenswrapper[4775]: E1002 03:00:58.084414 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" containerName="extract-utilities" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.084581 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" containerName="extract-utilities" Oct 02 03:00:58 crc kubenswrapper[4775]: E1002 03:00:58.084950 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" containerName="extract-content" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.085338 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" containerName="extract-content" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.086397 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="0081ac16-e0bd-494d-90f4-d5d17e514e74" containerName="mariadb-client-6-default" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.086644 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a0cb2e1-b09f-4fe2-a005-33d688e71fc1" containerName="registry-server" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.109409 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.109575 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.112985 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bphj\" (UniqueName: \"kubernetes.io/projected/0081ac16-e0bd-494d-90f4-d5d17e514e74-kube-api-access-8bphj\") on node \"crc\" DevicePath \"\"" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.214209 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq8h5\" (UniqueName: \"kubernetes.io/projected/f2defccc-51ac-45e0-bccd-0c6738302b7c-kube-api-access-rq8h5\") pod \"mariadb-client-7-default\" (UID: \"f2defccc-51ac-45e0-bccd-0c6738302b7c\") " pod="openstack/mariadb-client-7-default" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.317675 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq8h5\" (UniqueName: \"kubernetes.io/projected/f2defccc-51ac-45e0-bccd-0c6738302b7c-kube-api-access-rq8h5\") pod \"mariadb-client-7-default\" (UID: \"f2defccc-51ac-45e0-bccd-0c6738302b7c\") " pod="openstack/mariadb-client-7-default" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.348646 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq8h5\" (UniqueName: \"kubernetes.io/projected/f2defccc-51ac-45e0-bccd-0c6738302b7c-kube-api-access-rq8h5\") pod \"mariadb-client-7-default\" (UID: \"f2defccc-51ac-45e0-bccd-0c6738302b7c\") " pod="openstack/mariadb-client-7-default" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.437669 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.458792 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eea011ec2093905f514fc4f793893fae8aec938b84c239760edc6bfa2ed5f0a6" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.458917 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Oct 02 03:00:58 crc kubenswrapper[4775]: E1002 03:00:58.872923 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50401de0_705a_4de5_b211_56688c040651.slice\": RecentStats: unable to find data in memory cache]" Oct 02 03:00:58 crc kubenswrapper[4775]: I1002 03:00:58.995707 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 02 03:00:59 crc kubenswrapper[4775]: W1002 03:00:59.004732 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2defccc_51ac_45e0_bccd_0c6738302b7c.slice/crio-4e49d5c37b82db09cb37cd158da33c1e56fdb89a5b89be11f10f540266f9e6f6 WatchSource:0}: Error finding container 4e49d5c37b82db09cb37cd158da33c1e56fdb89a5b89be11f10f540266f9e6f6: Status 404 returned error can't find the container with id 4e49d5c37b82db09cb37cd158da33c1e56fdb89a5b89be11f10f540266f9e6f6 Oct 02 03:00:59 crc kubenswrapper[4775]: I1002 03:00:59.473596 4775 generic.go:334] "Generic (PLEG): container finished" podID="f2defccc-51ac-45e0-bccd-0c6738302b7c" containerID="673ac341e972788fe95dd4b6f7f41fa75ac52ade2a471604ad54e351d6fe2ae0" exitCode=0 Oct 02 03:00:59 crc kubenswrapper[4775]: I1002 03:00:59.473661 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"f2defccc-51ac-45e0-bccd-0c6738302b7c","Type":"ContainerDied","Data":"673ac341e972788fe95dd4b6f7f41fa75ac52ade2a471604ad54e351d6fe2ae0"} Oct 02 03:00:59 crc kubenswrapper[4775]: I1002 03:00:59.473701 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"f2defccc-51ac-45e0-bccd-0c6738302b7c","Type":"ContainerStarted","Data":"4e49d5c37b82db09cb37cd158da33c1e56fdb89a5b89be11f10f540266f9e6f6"} Oct 02 03:00:59 crc kubenswrapper[4775]: I1002 03:00:59.787198 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0081ac16-e0bd-494d-90f4-d5d17e514e74" path="/var/lib/kubelet/pods/0081ac16-e0bd-494d-90f4-d5d17e514e74/volumes" Oct 02 03:01:00 crc kubenswrapper[4775]: I1002 03:01:00.911424 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 02 03:01:00 crc kubenswrapper[4775]: I1002 03:01:00.934085 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_f2defccc-51ac-45e0-bccd-0c6738302b7c/mariadb-client-7-default/0.log" Oct 02 03:01:00 crc kubenswrapper[4775]: I1002 03:01:00.968982 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 02 03:01:00 crc kubenswrapper[4775]: I1002 03:01:00.973330 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.069543 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq8h5\" (UniqueName: \"kubernetes.io/projected/f2defccc-51ac-45e0-bccd-0c6738302b7c-kube-api-access-rq8h5\") pod \"f2defccc-51ac-45e0-bccd-0c6738302b7c\" (UID: \"f2defccc-51ac-45e0-bccd-0c6738302b7c\") " Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.078542 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2defccc-51ac-45e0-bccd-0c6738302b7c-kube-api-access-rq8h5" (OuterVolumeSpecName: "kube-api-access-rq8h5") pod "f2defccc-51ac-45e0-bccd-0c6738302b7c" (UID: "f2defccc-51ac-45e0-bccd-0c6738302b7c"). InnerVolumeSpecName "kube-api-access-rq8h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.175033 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq8h5\" (UniqueName: \"kubernetes.io/projected/f2defccc-51ac-45e0-bccd-0c6738302b7c-kube-api-access-rq8h5\") on node \"crc\" DevicePath \"\"" Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.185220 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Oct 02 03:01:01 crc kubenswrapper[4775]: E1002 03:01:01.185722 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2defccc-51ac-45e0-bccd-0c6738302b7c" containerName="mariadb-client-7-default" Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.185755 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2defccc-51ac-45e0-bccd-0c6738302b7c" containerName="mariadb-client-7-default" Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.186042 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2defccc-51ac-45e0-bccd-0c6738302b7c" containerName="mariadb-client-7-default" Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.186883 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.195249 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.379667 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp8hr\" (UniqueName: \"kubernetes.io/projected/ca80b963-3356-4b8c-9b7c-1641733b88cd-kube-api-access-qp8hr\") pod \"mariadb-client-2\" (UID: \"ca80b963-3356-4b8c-9b7c-1641733b88cd\") " pod="openstack/mariadb-client-2" Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.482077 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp8hr\" (UniqueName: \"kubernetes.io/projected/ca80b963-3356-4b8c-9b7c-1641733b88cd-kube-api-access-qp8hr\") pod \"mariadb-client-2\" (UID: \"ca80b963-3356-4b8c-9b7c-1641733b88cd\") " pod="openstack/mariadb-client-2" Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.498095 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e49d5c37b82db09cb37cd158da33c1e56fdb89a5b89be11f10f540266f9e6f6" Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.498184 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.516440 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp8hr\" (UniqueName: \"kubernetes.io/projected/ca80b963-3356-4b8c-9b7c-1641733b88cd-kube-api-access-qp8hr\") pod \"mariadb-client-2\" (UID: \"ca80b963-3356-4b8c-9b7c-1641733b88cd\") " pod="openstack/mariadb-client-2" Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.524105 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 02 03:01:01 crc kubenswrapper[4775]: I1002 03:01:01.779033 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2defccc-51ac-45e0-bccd-0c6738302b7c" path="/var/lib/kubelet/pods/f2defccc-51ac-45e0-bccd-0c6738302b7c/volumes" Oct 02 03:01:02 crc kubenswrapper[4775]: I1002 03:01:02.096593 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Oct 02 03:01:02 crc kubenswrapper[4775]: W1002 03:01:02.111858 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca80b963_3356_4b8c_9b7c_1641733b88cd.slice/crio-7f1308661c2485ec0af29cecadb053c6069c2e289e294b450913dab8d00b2688 WatchSource:0}: Error finding container 7f1308661c2485ec0af29cecadb053c6069c2e289e294b450913dab8d00b2688: Status 404 returned error can't find the container with id 7f1308661c2485ec0af29cecadb053c6069c2e289e294b450913dab8d00b2688 Oct 02 03:01:02 crc kubenswrapper[4775]: I1002 03:01:02.513547 4775 generic.go:334] "Generic (PLEG): container finished" podID="ca80b963-3356-4b8c-9b7c-1641733b88cd" containerID="f9ee0dd65d23eb09f02ca47efcc3944f82c77c1965c8946ffa8ee44f05c76f2b" exitCode=0 Oct 02 03:01:02 crc kubenswrapper[4775]: I1002 03:01:02.513624 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"ca80b963-3356-4b8c-9b7c-1641733b88cd","Type":"ContainerDied","Data":"f9ee0dd65d23eb09f02ca47efcc3944f82c77c1965c8946ffa8ee44f05c76f2b"} Oct 02 03:01:02 crc kubenswrapper[4775]: I1002 03:01:02.513697 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"ca80b963-3356-4b8c-9b7c-1641733b88cd","Type":"ContainerStarted","Data":"7f1308661c2485ec0af29cecadb053c6069c2e289e294b450913dab8d00b2688"} Oct 02 03:01:03 crc kubenswrapper[4775]: I1002 03:01:03.952704 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 02 03:01:03 crc kubenswrapper[4775]: I1002 03:01:03.973668 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_ca80b963-3356-4b8c-9b7c-1641733b88cd/mariadb-client-2/0.log" Oct 02 03:01:04 crc kubenswrapper[4775]: I1002 03:01:04.002844 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Oct 02 03:01:04 crc kubenswrapper[4775]: I1002 03:01:04.007617 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Oct 02 03:01:04 crc kubenswrapper[4775]: I1002 03:01:04.040168 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qp8hr\" (UniqueName: \"kubernetes.io/projected/ca80b963-3356-4b8c-9b7c-1641733b88cd-kube-api-access-qp8hr\") pod \"ca80b963-3356-4b8c-9b7c-1641733b88cd\" (UID: \"ca80b963-3356-4b8c-9b7c-1641733b88cd\") " Oct 02 03:01:04 crc kubenswrapper[4775]: I1002 03:01:04.045154 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca80b963-3356-4b8c-9b7c-1641733b88cd-kube-api-access-qp8hr" (OuterVolumeSpecName: "kube-api-access-qp8hr") pod "ca80b963-3356-4b8c-9b7c-1641733b88cd" (UID: "ca80b963-3356-4b8c-9b7c-1641733b88cd"). InnerVolumeSpecName "kube-api-access-qp8hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:01:04 crc kubenswrapper[4775]: I1002 03:01:04.142187 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qp8hr\" (UniqueName: \"kubernetes.io/projected/ca80b963-3356-4b8c-9b7c-1641733b88cd-kube-api-access-qp8hr\") on node \"crc\" DevicePath \"\"" Oct 02 03:01:04 crc kubenswrapper[4775]: I1002 03:01:04.539984 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f1308661c2485ec0af29cecadb053c6069c2e289e294b450913dab8d00b2688" Oct 02 03:01:04 crc kubenswrapper[4775]: I1002 03:01:04.540227 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Oct 02 03:01:05 crc kubenswrapper[4775]: I1002 03:01:05.785779 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca80b963-3356-4b8c-9b7c-1641733b88cd" path="/var/lib/kubelet/pods/ca80b963-3356-4b8c-9b7c-1641733b88cd/volumes" Oct 02 03:01:07 crc kubenswrapper[4775]: I1002 03:01:07.241544 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:01:07 crc kubenswrapper[4775]: I1002 03:01:07.242112 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:01:07 crc kubenswrapper[4775]: I1002 03:01:07.242188 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 03:01:07 crc kubenswrapper[4775]: I1002 03:01:07.243452 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"36c9b00ccca04be89dfec3790031865696944363d02afbb5b6006911151d9160"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 03:01:07 crc kubenswrapper[4775]: I1002 03:01:07.243589 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://36c9b00ccca04be89dfec3790031865696944363d02afbb5b6006911151d9160" gracePeriod=600 Oct 02 03:01:07 crc kubenswrapper[4775]: I1002 03:01:07.575632 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="36c9b00ccca04be89dfec3790031865696944363d02afbb5b6006911151d9160" exitCode=0 Oct 02 03:01:07 crc kubenswrapper[4775]: I1002 03:01:07.575713 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"36c9b00ccca04be89dfec3790031865696944363d02afbb5b6006911151d9160"} Oct 02 03:01:07 crc kubenswrapper[4775]: I1002 03:01:07.575791 4775 scope.go:117] "RemoveContainer" containerID="4a394a4fc98c6e0e2f1b5f3c4fc79010046f145a6682d0ea94e6fe96a9dcc51c" Oct 02 03:01:08 crc kubenswrapper[4775]: I1002 03:01:08.593241 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636"} Oct 02 03:01:10 crc kubenswrapper[4775]: I1002 03:01:10.273687 4775 scope.go:117] "RemoveContainer" containerID="693914b1747317796134e52a369fc2ab24a4681cc5644a1182bcb07992968517" Oct 02 03:03:07 crc kubenswrapper[4775]: I1002 03:03:07.233781 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:03:07 crc kubenswrapper[4775]: I1002 03:03:07.234511 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:03:37 crc kubenswrapper[4775]: I1002 03:03:37.233471 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:03:37 crc kubenswrapper[4775]: I1002 03:03:37.234175 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:04:07 crc kubenswrapper[4775]: I1002 03:04:07.234312 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:04:07 crc kubenswrapper[4775]: I1002 03:04:07.235161 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:04:07 crc kubenswrapper[4775]: I1002 03:04:07.235239 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 03:04:07 crc kubenswrapper[4775]: I1002 03:04:07.236226 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 03:04:07 crc kubenswrapper[4775]: I1002 03:04:07.236348 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" gracePeriod=600 Oct 02 03:04:07 crc kubenswrapper[4775]: E1002 03:04:07.369407 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:04:08 crc kubenswrapper[4775]: I1002 03:04:08.372290 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" exitCode=0 Oct 02 03:04:08 crc kubenswrapper[4775]: I1002 03:04:08.372380 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636"} Oct 02 03:04:08 crc kubenswrapper[4775]: I1002 03:04:08.372823 4775 scope.go:117] "RemoveContainer" containerID="36c9b00ccca04be89dfec3790031865696944363d02afbb5b6006911151d9160" Oct 02 03:04:08 crc kubenswrapper[4775]: I1002 03:04:08.373555 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:04:08 crc kubenswrapper[4775]: E1002 03:04:08.373907 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:04:23 crc kubenswrapper[4775]: I1002 03:04:23.770848 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:04:23 crc kubenswrapper[4775]: E1002 03:04:23.772384 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:04:35 crc kubenswrapper[4775]: I1002 03:04:35.767094 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:04:35 crc kubenswrapper[4775]: E1002 03:04:35.768610 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:04:48 crc kubenswrapper[4775]: I1002 03:04:48.765540 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:04:48 crc kubenswrapper[4775]: E1002 03:04:48.766704 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:05:00 crc kubenswrapper[4775]: I1002 03:05:00.766213 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:05:00 crc kubenswrapper[4775]: E1002 03:05:00.766993 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:05:11 crc kubenswrapper[4775]: I1002 03:05:11.766211 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:05:11 crc kubenswrapper[4775]: E1002 03:05:11.767393 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:05:24 crc kubenswrapper[4775]: I1002 03:05:24.765659 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:05:24 crc kubenswrapper[4775]: E1002 03:05:24.766732 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:05:37 crc kubenswrapper[4775]: I1002 03:05:37.765310 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:05:37 crc kubenswrapper[4775]: E1002 03:05:37.766854 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:05:38 crc kubenswrapper[4775]: I1002 03:05:38.085913 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Oct 02 03:05:38 crc kubenswrapper[4775]: E1002 03:05:38.086652 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca80b963-3356-4b8c-9b7c-1641733b88cd" containerName="mariadb-client-2" Oct 02 03:05:38 crc kubenswrapper[4775]: I1002 03:05:38.086696 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca80b963-3356-4b8c-9b7c-1641733b88cd" containerName="mariadb-client-2" Oct 02 03:05:38 crc kubenswrapper[4775]: I1002 03:05:38.087148 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca80b963-3356-4b8c-9b7c-1641733b88cd" containerName="mariadb-client-2" Oct 02 03:05:38 crc kubenswrapper[4775]: I1002 03:05:38.088061 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 02 03:05:38 crc kubenswrapper[4775]: I1002 03:05:38.091315 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-tnzbn" Oct 02 03:05:38 crc kubenswrapper[4775]: I1002 03:05:38.100935 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 02 03:05:38 crc kubenswrapper[4775]: I1002 03:05:38.254898 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t48cr\" (UniqueName: \"kubernetes.io/projected/3f39e43d-dd89-41c5-9725-ffc4011d49e7-kube-api-access-t48cr\") pod \"mariadb-copy-data\" (UID: \"3f39e43d-dd89-41c5-9725-ffc4011d49e7\") " pod="openstack/mariadb-copy-data" Oct 02 03:05:38 crc kubenswrapper[4775]: I1002 03:05:38.255033 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0437535e-8196-42c1-a3da-3e40ab45cc18\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0437535e-8196-42c1-a3da-3e40ab45cc18\") pod \"mariadb-copy-data\" (UID: \"3f39e43d-dd89-41c5-9725-ffc4011d49e7\") " pod="openstack/mariadb-copy-data" Oct 02 03:05:38 crc kubenswrapper[4775]: I1002 03:05:38.356835 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t48cr\" (UniqueName: \"kubernetes.io/projected/3f39e43d-dd89-41c5-9725-ffc4011d49e7-kube-api-access-t48cr\") pod \"mariadb-copy-data\" (UID: \"3f39e43d-dd89-41c5-9725-ffc4011d49e7\") " pod="openstack/mariadb-copy-data" Oct 02 03:05:38 crc kubenswrapper[4775]: I1002 03:05:38.356946 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0437535e-8196-42c1-a3da-3e40ab45cc18\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0437535e-8196-42c1-a3da-3e40ab45cc18\") pod \"mariadb-copy-data\" (UID: \"3f39e43d-dd89-41c5-9725-ffc4011d49e7\") " pod="openstack/mariadb-copy-data" Oct 02 03:05:38 crc kubenswrapper[4775]: I1002 03:05:38.364193 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 03:05:38 crc kubenswrapper[4775]: I1002 03:05:38.364271 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0437535e-8196-42c1-a3da-3e40ab45cc18\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0437535e-8196-42c1-a3da-3e40ab45cc18\") pod \"mariadb-copy-data\" (UID: \"3f39e43d-dd89-41c5-9725-ffc4011d49e7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/5e0e84e8a8446a27e16f0068d3cce5f50e24e72e4c32ad4ad83893d3b6d0a0d2/globalmount\"" pod="openstack/mariadb-copy-data" Oct 02 03:05:38 crc kubenswrapper[4775]: I1002 03:05:38.491609 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t48cr\" (UniqueName: \"kubernetes.io/projected/3f39e43d-dd89-41c5-9725-ffc4011d49e7-kube-api-access-t48cr\") pod \"mariadb-copy-data\" (UID: \"3f39e43d-dd89-41c5-9725-ffc4011d49e7\") " pod="openstack/mariadb-copy-data" Oct 02 03:05:38 crc kubenswrapper[4775]: I1002 03:05:38.740103 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0437535e-8196-42c1-a3da-3e40ab45cc18\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0437535e-8196-42c1-a3da-3e40ab45cc18\") pod \"mariadb-copy-data\" (UID: \"3f39e43d-dd89-41c5-9725-ffc4011d49e7\") " pod="openstack/mariadb-copy-data" Oct 02 03:05:39 crc kubenswrapper[4775]: I1002 03:05:39.028373 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Oct 02 03:05:39 crc kubenswrapper[4775]: I1002 03:05:39.474184 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Oct 02 03:05:40 crc kubenswrapper[4775]: I1002 03:05:40.325517 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"3f39e43d-dd89-41c5-9725-ffc4011d49e7","Type":"ContainerStarted","Data":"9658eb311b07646537782bad4488e9c9ad86d998b64013397f8b6f2719d26c4b"} Oct 02 03:05:40 crc kubenswrapper[4775]: I1002 03:05:40.325586 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"3f39e43d-dd89-41c5-9725-ffc4011d49e7","Type":"ContainerStarted","Data":"2a1a68db48c1f28b7f3709fd05940e2d38b980cccd497af366ea0aba33da6ed5"} Oct 02 03:05:40 crc kubenswrapper[4775]: I1002 03:05:40.354512 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.354484704 podStartE2EDuration="3.354484704s" podCreationTimestamp="2025-10-02 03:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:05:40.346161499 +0000 UTC m=+5077.512905629" watchObservedRunningTime="2025-10-02 03:05:40.354484704 +0000 UTC m=+5077.521228774" Oct 02 03:05:42 crc kubenswrapper[4775]: I1002 03:05:42.400073 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 02 03:05:42 crc kubenswrapper[4775]: I1002 03:05:42.402089 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 02 03:05:42 crc kubenswrapper[4775]: I1002 03:05:42.409849 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 02 03:05:42 crc kubenswrapper[4775]: I1002 03:05:42.542308 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lrls\" (UniqueName: \"kubernetes.io/projected/4935c71d-4123-478c-98c6-bcb86eceb39e-kube-api-access-7lrls\") pod \"mariadb-client\" (UID: \"4935c71d-4123-478c-98c6-bcb86eceb39e\") " pod="openstack/mariadb-client" Oct 02 03:05:42 crc kubenswrapper[4775]: I1002 03:05:42.644983 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lrls\" (UniqueName: \"kubernetes.io/projected/4935c71d-4123-478c-98c6-bcb86eceb39e-kube-api-access-7lrls\") pod \"mariadb-client\" (UID: \"4935c71d-4123-478c-98c6-bcb86eceb39e\") " pod="openstack/mariadb-client" Oct 02 03:05:42 crc kubenswrapper[4775]: I1002 03:05:42.677724 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lrls\" (UniqueName: \"kubernetes.io/projected/4935c71d-4123-478c-98c6-bcb86eceb39e-kube-api-access-7lrls\") pod \"mariadb-client\" (UID: \"4935c71d-4123-478c-98c6-bcb86eceb39e\") " pod="openstack/mariadb-client" Oct 02 03:05:42 crc kubenswrapper[4775]: I1002 03:05:42.733654 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 02 03:05:43 crc kubenswrapper[4775]: I1002 03:05:43.227652 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 02 03:05:43 crc kubenswrapper[4775]: I1002 03:05:43.362948 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"4935c71d-4123-478c-98c6-bcb86eceb39e","Type":"ContainerStarted","Data":"5d1ac78ce28c2027533a387bd2366676d6ccf471c4b10612648fa7b37ba312d9"} Oct 02 03:05:44 crc kubenswrapper[4775]: I1002 03:05:44.378122 4775 generic.go:334] "Generic (PLEG): container finished" podID="4935c71d-4123-478c-98c6-bcb86eceb39e" containerID="a130041296f0f400ecb06409082211457f73f30e50e1c5ea06c45ba9df9350ea" exitCode=0 Oct 02 03:05:44 crc kubenswrapper[4775]: I1002 03:05:44.378190 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"4935c71d-4123-478c-98c6-bcb86eceb39e","Type":"ContainerDied","Data":"a130041296f0f400ecb06409082211457f73f30e50e1c5ea06c45ba9df9350ea"} Oct 02 03:05:45 crc kubenswrapper[4775]: I1002 03:05:45.785298 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 02 03:05:45 crc kubenswrapper[4775]: I1002 03:05:45.815392 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_4935c71d-4123-478c-98c6-bcb86eceb39e/mariadb-client/0.log" Oct 02 03:05:45 crc kubenswrapper[4775]: I1002 03:05:45.869797 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 02 03:05:45 crc kubenswrapper[4775]: I1002 03:05:45.876788 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 02 03:05:45 crc kubenswrapper[4775]: I1002 03:05:45.901754 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lrls\" (UniqueName: \"kubernetes.io/projected/4935c71d-4123-478c-98c6-bcb86eceb39e-kube-api-access-7lrls\") pod \"4935c71d-4123-478c-98c6-bcb86eceb39e\" (UID: \"4935c71d-4123-478c-98c6-bcb86eceb39e\") " Oct 02 03:05:45 crc kubenswrapper[4775]: I1002 03:05:45.908443 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4935c71d-4123-478c-98c6-bcb86eceb39e-kube-api-access-7lrls" (OuterVolumeSpecName: "kube-api-access-7lrls") pod "4935c71d-4123-478c-98c6-bcb86eceb39e" (UID: "4935c71d-4123-478c-98c6-bcb86eceb39e"). InnerVolumeSpecName "kube-api-access-7lrls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:05:46 crc kubenswrapper[4775]: I1002 03:05:46.003960 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lrls\" (UniqueName: \"kubernetes.io/projected/4935c71d-4123-478c-98c6-bcb86eceb39e-kube-api-access-7lrls\") on node \"crc\" DevicePath \"\"" Oct 02 03:05:46 crc kubenswrapper[4775]: I1002 03:05:46.015093 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Oct 02 03:05:46 crc kubenswrapper[4775]: E1002 03:05:46.015552 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4935c71d-4123-478c-98c6-bcb86eceb39e" containerName="mariadb-client" Oct 02 03:05:46 crc kubenswrapper[4775]: I1002 03:05:46.015575 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4935c71d-4123-478c-98c6-bcb86eceb39e" containerName="mariadb-client" Oct 02 03:05:46 crc kubenswrapper[4775]: I1002 03:05:46.015804 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4935c71d-4123-478c-98c6-bcb86eceb39e" containerName="mariadb-client" Oct 02 03:05:46 crc kubenswrapper[4775]: I1002 03:05:46.016490 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 02 03:05:46 crc kubenswrapper[4775]: I1002 03:05:46.022575 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 02 03:05:46 crc kubenswrapper[4775]: I1002 03:05:46.105984 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcqj2\" (UniqueName: \"kubernetes.io/projected/a597bf86-3838-4697-b65e-b472e32ad7c6-kube-api-access-kcqj2\") pod \"mariadb-client\" (UID: \"a597bf86-3838-4697-b65e-b472e32ad7c6\") " pod="openstack/mariadb-client" Oct 02 03:05:46 crc kubenswrapper[4775]: I1002 03:05:46.207931 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcqj2\" (UniqueName: \"kubernetes.io/projected/a597bf86-3838-4697-b65e-b472e32ad7c6-kube-api-access-kcqj2\") pod \"mariadb-client\" (UID: \"a597bf86-3838-4697-b65e-b472e32ad7c6\") " pod="openstack/mariadb-client" Oct 02 03:05:46 crc kubenswrapper[4775]: I1002 03:05:46.228651 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcqj2\" (UniqueName: \"kubernetes.io/projected/a597bf86-3838-4697-b65e-b472e32ad7c6-kube-api-access-kcqj2\") pod \"mariadb-client\" (UID: \"a597bf86-3838-4697-b65e-b472e32ad7c6\") " pod="openstack/mariadb-client" Oct 02 03:05:46 crc kubenswrapper[4775]: I1002 03:05:46.338942 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 02 03:05:46 crc kubenswrapper[4775]: I1002 03:05:46.399609 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d1ac78ce28c2027533a387bd2366676d6ccf471c4b10612648fa7b37ba312d9" Oct 02 03:05:46 crc kubenswrapper[4775]: I1002 03:05:46.399683 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 02 03:05:46 crc kubenswrapper[4775]: I1002 03:05:46.426685 4775 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/mariadb-client" oldPodUID="4935c71d-4123-478c-98c6-bcb86eceb39e" podUID="a597bf86-3838-4697-b65e-b472e32ad7c6" Oct 02 03:05:46 crc kubenswrapper[4775]: I1002 03:05:46.617264 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Oct 02 03:05:46 crc kubenswrapper[4775]: W1002 03:05:46.625541 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda597bf86_3838_4697_b65e_b472e32ad7c6.slice/crio-d3e6af9252ab5593138708830a50b09dcd2c635f16abe28dbf0744b5db87d5ec WatchSource:0}: Error finding container d3e6af9252ab5593138708830a50b09dcd2c635f16abe28dbf0744b5db87d5ec: Status 404 returned error can't find the container with id d3e6af9252ab5593138708830a50b09dcd2c635f16abe28dbf0744b5db87d5ec Oct 02 03:05:47 crc kubenswrapper[4775]: I1002 03:05:47.411595 4775 generic.go:334] "Generic (PLEG): container finished" podID="a597bf86-3838-4697-b65e-b472e32ad7c6" containerID="d2fa9f54cc55f33effa2e0feb0aebe2f4a5da2cfa68b3365931a5c57a4eb0ca8" exitCode=0 Oct 02 03:05:47 crc kubenswrapper[4775]: I1002 03:05:47.411676 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"a597bf86-3838-4697-b65e-b472e32ad7c6","Type":"ContainerDied","Data":"d2fa9f54cc55f33effa2e0feb0aebe2f4a5da2cfa68b3365931a5c57a4eb0ca8"} Oct 02 03:05:47 crc kubenswrapper[4775]: I1002 03:05:47.411732 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"a597bf86-3838-4697-b65e-b472e32ad7c6","Type":"ContainerStarted","Data":"d3e6af9252ab5593138708830a50b09dcd2c635f16abe28dbf0744b5db87d5ec"} Oct 02 03:05:47 crc kubenswrapper[4775]: I1002 03:05:47.782776 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4935c71d-4123-478c-98c6-bcb86eceb39e" path="/var/lib/kubelet/pods/4935c71d-4123-478c-98c6-bcb86eceb39e/volumes" Oct 02 03:05:48 crc kubenswrapper[4775]: I1002 03:05:48.899834 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 02 03:05:48 crc kubenswrapper[4775]: I1002 03:05:48.922887 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_a597bf86-3838-4697-b65e-b472e32ad7c6/mariadb-client/0.log" Oct 02 03:05:48 crc kubenswrapper[4775]: I1002 03:05:48.955468 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcqj2\" (UniqueName: \"kubernetes.io/projected/a597bf86-3838-4697-b65e-b472e32ad7c6-kube-api-access-kcqj2\") pod \"a597bf86-3838-4697-b65e-b472e32ad7c6\" (UID: \"a597bf86-3838-4697-b65e-b472e32ad7c6\") " Oct 02 03:05:48 crc kubenswrapper[4775]: I1002 03:05:48.962104 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Oct 02 03:05:48 crc kubenswrapper[4775]: I1002 03:05:48.964264 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a597bf86-3838-4697-b65e-b472e32ad7c6-kube-api-access-kcqj2" (OuterVolumeSpecName: "kube-api-access-kcqj2") pod "a597bf86-3838-4697-b65e-b472e32ad7c6" (UID: "a597bf86-3838-4697-b65e-b472e32ad7c6"). InnerVolumeSpecName "kube-api-access-kcqj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:05:48 crc kubenswrapper[4775]: I1002 03:05:48.972700 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Oct 02 03:05:49 crc kubenswrapper[4775]: I1002 03:05:49.058154 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcqj2\" (UniqueName: \"kubernetes.io/projected/a597bf86-3838-4697-b65e-b472e32ad7c6-kube-api-access-kcqj2\") on node \"crc\" DevicePath \"\"" Oct 02 03:05:49 crc kubenswrapper[4775]: I1002 03:05:49.445431 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Oct 02 03:05:49 crc kubenswrapper[4775]: I1002 03:05:49.445835 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3e6af9252ab5593138708830a50b09dcd2c635f16abe28dbf0744b5db87d5ec" Oct 02 03:05:49 crc kubenswrapper[4775]: I1002 03:05:49.786069 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a597bf86-3838-4697-b65e-b472e32ad7c6" path="/var/lib/kubelet/pods/a597bf86-3838-4697-b65e-b472e32ad7c6/volumes" Oct 02 03:05:51 crc kubenswrapper[4775]: I1002 03:05:51.765467 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:05:51 crc kubenswrapper[4775]: E1002 03:05:51.766205 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:06:03 crc kubenswrapper[4775]: I1002 03:06:03.787189 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:06:03 crc kubenswrapper[4775]: E1002 03:06:03.788773 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:06:14 crc kubenswrapper[4775]: I1002 03:06:14.766486 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:06:14 crc kubenswrapper[4775]: E1002 03:06:14.767472 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.066734 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 03:06:24 crc kubenswrapper[4775]: E1002 03:06:24.067795 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a597bf86-3838-4697-b65e-b472e32ad7c6" containerName="mariadb-client" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.067818 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a597bf86-3838-4697-b65e-b472e32ad7c6" containerName="mariadb-client" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.068163 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a597bf86-3838-4697-b65e-b472e32ad7c6" containerName="mariadb-client" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.069651 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.074877 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.075533 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.075688 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-ffp2n" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.085074 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.087319 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.091742 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.097170 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.100012 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.109349 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.115652 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.252413 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/790fc282-a15f-4530-a6e0-20b5e44401ad-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.252491 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/730a72f5-678f-4ae4-8008-397b6ba49170-config\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.252540 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-config\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.252639 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/730a72f5-678f-4ae4-8008-397b6ba49170-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.252688 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hgxm\" (UniqueName: \"kubernetes.io/projected/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-kube-api-access-2hgxm\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.252803 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xp99r\" (UniqueName: \"kubernetes.io/projected/790fc282-a15f-4530-a6e0-20b5e44401ad-kube-api-access-xp99r\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.252892 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d45c8ecd-4a0a-4210-926a-7e1005a5a8b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d45c8ecd-4a0a-4210-926a-7e1005a5a8b5\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.252977 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.253026 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-330d2df4-bafa-499f-9ac7-29f699e6bea3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-330d2df4-bafa-499f-9ac7-29f699e6bea3\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.253057 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/790fc282-a15f-4530-a6e0-20b5e44401ad-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.253236 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/730a72f5-678f-4ae4-8008-397b6ba49170-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.253308 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.253345 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/730a72f5-678f-4ae4-8008-397b6ba49170-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.253391 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.253420 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/790fc282-a15f-4530-a6e0-20b5e44401ad-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.253448 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/790fc282-a15f-4530-a6e0-20b5e44401ad-config\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.253595 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq2ft\" (UniqueName: \"kubernetes.io/projected/730a72f5-678f-4ae4-8008-397b6ba49170-kube-api-access-dq2ft\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.253644 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b9463e4f-7464-493a-9fa7-7ef769ad2ebd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9463e4f-7464-493a-9fa7-7ef769ad2ebd\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.256410 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.258760 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.264030 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-vzw2b" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.264337 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.264674 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.287873 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.290501 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.300376 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.308880 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.316816 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.318400 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.332929 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355390 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-config\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355441 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/730a72f5-678f-4ae4-8008-397b6ba49170-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355484 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hgxm\" (UniqueName: \"kubernetes.io/projected/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-kube-api-access-2hgxm\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355511 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xp99r\" (UniqueName: \"kubernetes.io/projected/790fc282-a15f-4530-a6e0-20b5e44401ad-kube-api-access-xp99r\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355538 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d45c8ecd-4a0a-4210-926a-7e1005a5a8b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d45c8ecd-4a0a-4210-926a-7e1005a5a8b5\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355566 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355589 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-330d2df4-bafa-499f-9ac7-29f699e6bea3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-330d2df4-bafa-499f-9ac7-29f699e6bea3\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355608 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/790fc282-a15f-4530-a6e0-20b5e44401ad-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355647 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-config\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355687 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/730a72f5-678f-4ae4-8008-397b6ba49170-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355715 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355739 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355764 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/730a72f5-678f-4ae4-8008-397b6ba49170-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355787 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355810 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/790fc282-a15f-4530-a6e0-20b5e44401ad-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355834 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/790fc282-a15f-4530-a6e0-20b5e44401ad-config\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355874 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3279b5b4-4067-46b9-ad67-ef948750d75d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3279b5b4-4067-46b9-ad67-ef948750d75d\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355902 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq2ft\" (UniqueName: \"kubernetes.io/projected/730a72f5-678f-4ae4-8008-397b6ba49170-kube-api-access-dq2ft\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355928 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b9463e4f-7464-493a-9fa7-7ef769ad2ebd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9463e4f-7464-493a-9fa7-7ef769ad2ebd\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.355980 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.356006 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.356040 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/790fc282-a15f-4530-a6e0-20b5e44401ad-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.356063 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q6c6\" (UniqueName: \"kubernetes.io/projected/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-kube-api-access-9q6c6\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.356088 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/730a72f5-678f-4ae4-8008-397b6ba49170-config\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.356684 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.357486 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-config\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.358622 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/730a72f5-678f-4ae4-8008-397b6ba49170-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.359108 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.359837 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/730a72f5-678f-4ae4-8008-397b6ba49170-config\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.360240 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/790fc282-a15f-4530-a6e0-20b5e44401ad-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.360251 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/790fc282-a15f-4530-a6e0-20b5e44401ad-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.360491 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/730a72f5-678f-4ae4-8008-397b6ba49170-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.361454 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/790fc282-a15f-4530-a6e0-20b5e44401ad-config\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.363395 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.364090 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.364123 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-330d2df4-bafa-499f-9ac7-29f699e6bea3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-330d2df4-bafa-499f-9ac7-29f699e6bea3\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ff0998e8cbe5ac9b931ce56bf67523940a6a081f068f99db82212a90cb4b3532/globalmount\"" pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.366120 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.366156 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d45c8ecd-4a0a-4210-926a-7e1005a5a8b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d45c8ecd-4a0a-4210-926a-7e1005a5a8b5\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/37a4729e4b4d27b1b75057de76b7718dcc113d8e767e79a22de8219685a895fe/globalmount\"" pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.366421 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.366490 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b9463e4f-7464-493a-9fa7-7ef769ad2ebd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9463e4f-7464-493a-9fa7-7ef769ad2ebd\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a25c1ae835e2f66a4ece660ae833fa6a52ac111392e784da0eb469c76c798aa9/globalmount\"" pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.377648 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/730a72f5-678f-4ae4-8008-397b6ba49170-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.380379 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xp99r\" (UniqueName: \"kubernetes.io/projected/790fc282-a15f-4530-a6e0-20b5e44401ad-kube-api-access-xp99r\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.381873 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq2ft\" (UniqueName: \"kubernetes.io/projected/730a72f5-678f-4ae4-8008-397b6ba49170-kube-api-access-dq2ft\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.384559 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hgxm\" (UniqueName: \"kubernetes.io/projected/ddccc962-e478-4e58-b4d3-cfcf109fd0f7-kube-api-access-2hgxm\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.387198 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/790fc282-a15f-4530-a6e0-20b5e44401ad-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.403569 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d45c8ecd-4a0a-4210-926a-7e1005a5a8b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d45c8ecd-4a0a-4210-926a-7e1005a5a8b5\") pod \"ovsdbserver-nb-1\" (UID: \"ddccc962-e478-4e58-b4d3-cfcf109fd0f7\") " pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.406859 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-330d2df4-bafa-499f-9ac7-29f699e6bea3\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-330d2df4-bafa-499f-9ac7-29f699e6bea3\") pod \"ovsdbserver-nb-0\" (UID: \"790fc282-a15f-4530-a6e0-20b5e44401ad\") " pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.416916 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b9463e4f-7464-493a-9fa7-7ef769ad2ebd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b9463e4f-7464-493a-9fa7-7ef769ad2ebd\") pod \"ovsdbserver-nb-2\" (UID: \"730a72f5-678f-4ae4-8008-397b6ba49170\") " pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.425543 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.446028 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.458864 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.458937 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hrcx\" (UniqueName: \"kubernetes.io/projected/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-kube-api-access-9hrcx\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.458970 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b18a1684-4fbd-4a52-95de-8d63ce13b38d-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.458997 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzq96\" (UniqueName: \"kubernetes.io/projected/b18a1684-4fbd-4a52-95de-8d63ce13b38d-kube-api-access-pzq96\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.459035 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3279b5b4-4067-46b9-ad67-ef948750d75d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3279b5b4-4067-46b9-ad67-ef948750d75d\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.459069 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.459095 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.459124 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ae59565d-a0f2-4d0a-95c7-e810ffe8a322\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae59565d-a0f2-4d0a-95c7-e810ffe8a322\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.459148 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q6c6\" (UniqueName: \"kubernetes.io/projected/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-kube-api-access-9q6c6\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.459167 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.459186 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.459206 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-config\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.459225 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b18a1684-4fbd-4a52-95de-8d63ce13b38d-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.459266 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b18a1684-4fbd-4a52-95de-8d63ce13b38d-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.459287 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-d1eb9c70-ae8e-48f8-8d54-caf25e015327\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d1eb9c70-ae8e-48f8-8d54-caf25e015327\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.459304 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b18a1684-4fbd-4a52-95de-8d63ce13b38d-config\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.459325 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-config\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.459352 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.460630 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.461415 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-config\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.463705 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.466182 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.466245 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3279b5b4-4067-46b9-ad67-ef948750d75d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3279b5b4-4067-46b9-ad67-ef948750d75d\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/abcd31951ee5442a2af4d8c389f0e1443c4b0dc8f14c962bcb09e47bfca02bc0/globalmount\"" pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.471288 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.482847 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q6c6\" (UniqueName: \"kubernetes.io/projected/1964d668-d2ef-4f06-9ffc-c49f14eb81c3-kube-api-access-9q6c6\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.494783 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3279b5b4-4067-46b9-ad67-ef948750d75d\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3279b5b4-4067-46b9-ad67-ef948750d75d\") pod \"ovsdbserver-sb-0\" (UID: \"1964d668-d2ef-4f06-9ffc-c49f14eb81c3\") " pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.561087 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ae59565d-a0f2-4d0a-95c7-e810ffe8a322\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae59565d-a0f2-4d0a-95c7-e810ffe8a322\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.561571 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.561599 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.561624 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-config\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.561645 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b18a1684-4fbd-4a52-95de-8d63ce13b38d-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.561697 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b18a1684-4fbd-4a52-95de-8d63ce13b38d-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.561717 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-d1eb9c70-ae8e-48f8-8d54-caf25e015327\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d1eb9c70-ae8e-48f8-8d54-caf25e015327\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.561734 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b18a1684-4fbd-4a52-95de-8d63ce13b38d-config\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.561771 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.561800 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hrcx\" (UniqueName: \"kubernetes.io/projected/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-kube-api-access-9hrcx\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.561817 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b18a1684-4fbd-4a52-95de-8d63ce13b38d-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.561839 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzq96\" (UniqueName: \"kubernetes.io/projected/b18a1684-4fbd-4a52-95de-8d63ce13b38d-kube-api-access-pzq96\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.562204 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b18a1684-4fbd-4a52-95de-8d63ce13b38d-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.562573 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-config\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.562615 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.562750 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.564913 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b18a1684-4fbd-4a52-95de-8d63ce13b38d-config\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.565488 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b18a1684-4fbd-4a52-95de-8d63ce13b38d-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.565580 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.565602 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ae59565d-a0f2-4d0a-95c7-e810ffe8a322\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae59565d-a0f2-4d0a-95c7-e810ffe8a322\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c7743099aa46ecf3cf3da023a6409e231298f834f8574a5ac38de32b1205b533/globalmount\"" pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.569520 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b18a1684-4fbd-4a52-95de-8d63ce13b38d-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.570339 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.577199 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.577222 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-d1eb9c70-ae8e-48f8-8d54-caf25e015327\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d1eb9c70-ae8e-48f8-8d54-caf25e015327\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/603733223dcfe889484eb9d72d140c12ad0212c083e205f1c81df9069912f9b8/globalmount\"" pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.578100 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzq96\" (UniqueName: \"kubernetes.io/projected/b18a1684-4fbd-4a52-95de-8d63ce13b38d-kube-api-access-pzq96\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.582196 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hrcx\" (UniqueName: \"kubernetes.io/projected/94de3ec6-4d84-4ef5-a127-284ff4e64ea2-kube-api-access-9hrcx\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.582634 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.610659 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ae59565d-a0f2-4d0a-95c7-e810ffe8a322\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ae59565d-a0f2-4d0a-95c7-e810ffe8a322\") pod \"ovsdbserver-sb-2\" (UID: \"b18a1684-4fbd-4a52-95de-8d63ce13b38d\") " pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.614541 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-d1eb9c70-ae8e-48f8-8d54-caf25e015327\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-d1eb9c70-ae8e-48f8-8d54-caf25e015327\") pod \"ovsdbserver-sb-1\" (UID: \"94de3ec6-4d84-4ef5-a127-284ff4e64ea2\") " pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.620359 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.638755 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.707981 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.869841 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Oct 02 03:06:24 crc kubenswrapper[4775]: I1002 03:06:24.961670 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Oct 02 03:06:24 crc kubenswrapper[4775]: W1002 03:06:24.991780 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddccc962_e478_4e58_b4d3_cfcf109fd0f7.slice/crio-6b2cf36c0f61330a22f76d32483be22e6dac5f1b2bbab7be41e3948cbfd40bb2 WatchSource:0}: Error finding container 6b2cf36c0f61330a22f76d32483be22e6dac5f1b2bbab7be41e3948cbfd40bb2: Status 404 returned error can't find the container with id 6b2cf36c0f61330a22f76d32483be22e6dac5f1b2bbab7be41e3948cbfd40bb2 Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.107604 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 02 03:06:25 crc kubenswrapper[4775]: W1002 03:06:25.121941 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod790fc282_a15f_4530_a6e0_20b5e44401ad.slice/crio-a9be090ab7b31e5faa90903830a77975314dfd6c57b3e9392294ed71948267f4 WatchSource:0}: Error finding container a9be090ab7b31e5faa90903830a77975314dfd6c57b3e9392294ed71948267f4: Status 404 returned error can't find the container with id a9be090ab7b31e5faa90903830a77975314dfd6c57b3e9392294ed71948267f4 Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.204014 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.821384 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b18a1684-4fbd-4a52-95de-8d63ce13b38d","Type":"ContainerStarted","Data":"1d48a947ab27c607e4b1e8b8ef0beb194e10534e8cd5e482bb86a968a56daacc"} Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.821425 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b18a1684-4fbd-4a52-95de-8d63ce13b38d","Type":"ContainerStarted","Data":"a3607492b7274a26fd1084a95ca74e1610f2624cbda62ea4473b25ab09806358"} Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.821436 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"b18a1684-4fbd-4a52-95de-8d63ce13b38d","Type":"ContainerStarted","Data":"b08eee2cdc3a51b2030a2e041ced17f34640236578e81e1345db5d6a60eb0b43"} Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.828479 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"ddccc962-e478-4e58-b4d3-cfcf109fd0f7","Type":"ContainerStarted","Data":"490f6d2cae9cdedd153a6b1fb5a958147ab227e2140a6fd7ba335978e5b64169"} Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.828519 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"ddccc962-e478-4e58-b4d3-cfcf109fd0f7","Type":"ContainerStarted","Data":"8eb2387684de70b99e01c03dfad3f10ffd4a2060102263ccb2b4c7552eb5d86a"} Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.828529 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"ddccc962-e478-4e58-b4d3-cfcf109fd0f7","Type":"ContainerStarted","Data":"6b2cf36c0f61330a22f76d32483be22e6dac5f1b2bbab7be41e3948cbfd40bb2"} Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.831894 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"790fc282-a15f-4530-a6e0-20b5e44401ad","Type":"ContainerStarted","Data":"a26692d0118ede62d3b5532d3a4ba0bea9b92290a5d116d5c7f99b54f4a80243"} Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.831927 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"790fc282-a15f-4530-a6e0-20b5e44401ad","Type":"ContainerStarted","Data":"02ca11b1f90056bb63749a219fa66012570723a28640b0442ab267a13ec07379"} Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.831939 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"790fc282-a15f-4530-a6e0-20b5e44401ad","Type":"ContainerStarted","Data":"a9be090ab7b31e5faa90903830a77975314dfd6c57b3e9392294ed71948267f4"} Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.834717 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"730a72f5-678f-4ae4-8008-397b6ba49170","Type":"ContainerStarted","Data":"649d3ec5ebc62261496c960b8d3d5629ca78631e6139e5339583adff9e01b89f"} Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.834742 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"730a72f5-678f-4ae4-8008-397b6ba49170","Type":"ContainerStarted","Data":"0beb8f217704b8255c0c6dfbb737719495bbc1b101a5351759d8e445268521bb"} Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.834751 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"730a72f5-678f-4ae4-8008-397b6ba49170","Type":"ContainerStarted","Data":"6eca3d136674c952a60dfa3ea7225902645da76cd71a305d2f6f7a7803a3ec8f"} Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.847077 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=2.847049827 podStartE2EDuration="2.847049827s" podCreationTimestamp="2025-10-02 03:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:06:25.846936414 +0000 UTC m=+5123.013680454" watchObservedRunningTime="2025-10-02 03:06:25.847049827 +0000 UTC m=+5123.013793897" Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.903074 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=2.90304906 podStartE2EDuration="2.90304906s" podCreationTimestamp="2025-10-02 03:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:06:25.873041007 +0000 UTC m=+5123.039785047" watchObservedRunningTime="2025-10-02 03:06:25.90304906 +0000 UTC m=+5123.069793120" Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.927905 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=2.92788279 podStartE2EDuration="2.92788279s" podCreationTimestamp="2025-10-02 03:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:06:25.898354169 +0000 UTC m=+5123.065098239" watchObservedRunningTime="2025-10-02 03:06:25.92788279 +0000 UTC m=+5123.094626840" Oct 02 03:06:25 crc kubenswrapper[4775]: I1002 03:06:25.932295 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=2.932262173 podStartE2EDuration="2.932262173s" podCreationTimestamp="2025-10-02 03:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:06:25.925390836 +0000 UTC m=+5123.092134906" watchObservedRunningTime="2025-10-02 03:06:25.932262173 +0000 UTC m=+5123.099006243" Oct 02 03:06:26 crc kubenswrapper[4775]: I1002 03:06:26.196847 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 02 03:06:26 crc kubenswrapper[4775]: W1002 03:06:26.206241 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1964d668_d2ef_4f06_9ffc_c49f14eb81c3.slice/crio-74b17112f3faac2aae59e074d3444efb250e400a71e5228ac54463da99678f66 WatchSource:0}: Error finding container 74b17112f3faac2aae59e074d3444efb250e400a71e5228ac54463da99678f66: Status 404 returned error can't find the container with id 74b17112f3faac2aae59e074d3444efb250e400a71e5228ac54463da99678f66 Oct 02 03:06:26 crc kubenswrapper[4775]: I1002 03:06:26.323638 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Oct 02 03:06:26 crc kubenswrapper[4775]: W1002 03:06:26.333578 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94de3ec6_4d84_4ef5_a127_284ff4e64ea2.slice/crio-013aa2d26a239bcdcb58456a03ea05e083e45d9bcf7a3091dc48c319ea08df45 WatchSource:0}: Error finding container 013aa2d26a239bcdcb58456a03ea05e083e45d9bcf7a3091dc48c319ea08df45: Status 404 returned error can't find the container with id 013aa2d26a239bcdcb58456a03ea05e083e45d9bcf7a3091dc48c319ea08df45 Oct 02 03:06:26 crc kubenswrapper[4775]: I1002 03:06:26.846381 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1964d668-d2ef-4f06-9ffc-c49f14eb81c3","Type":"ContainerStarted","Data":"94ab27e50d9c4052ffb179b2158b8dbf5887095326c9e5c096e9d3f5e9c844f3"} Oct 02 03:06:26 crc kubenswrapper[4775]: I1002 03:06:26.846748 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1964d668-d2ef-4f06-9ffc-c49f14eb81c3","Type":"ContainerStarted","Data":"ff658b548e7d09d6f3d2007f50671d78b66c1a6dd30c22b8c46454f3f94ef67f"} Oct 02 03:06:26 crc kubenswrapper[4775]: I1002 03:06:26.846773 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"1964d668-d2ef-4f06-9ffc-c49f14eb81c3","Type":"ContainerStarted","Data":"74b17112f3faac2aae59e074d3444efb250e400a71e5228ac54463da99678f66"} Oct 02 03:06:26 crc kubenswrapper[4775]: I1002 03:06:26.850614 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"94de3ec6-4d84-4ef5-a127-284ff4e64ea2","Type":"ContainerStarted","Data":"37e28e292a3a01bdab96e4f1a31663587e0852521ffca53c902bffc00e3e003a"} Oct 02 03:06:26 crc kubenswrapper[4775]: I1002 03:06:26.850679 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"94de3ec6-4d84-4ef5-a127-284ff4e64ea2","Type":"ContainerStarted","Data":"8c18c358d05b4d2cf82321490d7cb05b89629ef3d016db36c5fb8ff4477c6520"} Oct 02 03:06:26 crc kubenswrapper[4775]: I1002 03:06:26.850709 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"94de3ec6-4d84-4ef5-a127-284ff4e64ea2","Type":"ContainerStarted","Data":"013aa2d26a239bcdcb58456a03ea05e083e45d9bcf7a3091dc48c319ea08df45"} Oct 02 03:06:26 crc kubenswrapper[4775]: I1002 03:06:26.871650 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=3.871579837 podStartE2EDuration="3.871579837s" podCreationTimestamp="2025-10-02 03:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:06:26.863405086 +0000 UTC m=+5124.030149136" watchObservedRunningTime="2025-10-02 03:06:26.871579837 +0000 UTC m=+5124.038323897" Oct 02 03:06:26 crc kubenswrapper[4775]: I1002 03:06:26.897616 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=3.897592577 podStartE2EDuration="3.897592577s" podCreationTimestamp="2025-10-02 03:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:06:26.883844863 +0000 UTC m=+5124.050588943" watchObservedRunningTime="2025-10-02 03:06:26.897592577 +0000 UTC m=+5124.064336627" Oct 02 03:06:27 crc kubenswrapper[4775]: I1002 03:06:27.426758 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:27 crc kubenswrapper[4775]: I1002 03:06:27.446552 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:27 crc kubenswrapper[4775]: I1002 03:06:27.583541 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:27 crc kubenswrapper[4775]: I1002 03:06:27.622438 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:27 crc kubenswrapper[4775]: I1002 03:06:27.641085 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:27 crc kubenswrapper[4775]: I1002 03:06:27.709069 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:29 crc kubenswrapper[4775]: I1002 03:06:29.425826 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:29 crc kubenswrapper[4775]: I1002 03:06:29.447015 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:29 crc kubenswrapper[4775]: I1002 03:06:29.583813 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:29 crc kubenswrapper[4775]: I1002 03:06:29.622735 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:29 crc kubenswrapper[4775]: I1002 03:06:29.640769 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:29 crc kubenswrapper[4775]: I1002 03:06:29.709231 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:29 crc kubenswrapper[4775]: I1002 03:06:29.767234 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:06:29 crc kubenswrapper[4775]: E1002 03:06:29.767703 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.504155 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.511762 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.587283 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.592765 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.644679 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.701058 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.731405 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.756711 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.759532 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.828154 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.842006 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7789f5c469-8qpf5"] Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.843415 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.849223 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.872951 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7789f5c469-8qpf5"] Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.995577 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bdl7\" (UniqueName: \"kubernetes.io/projected/7f514b1c-ebee-4a89-be14-59841552e497-kube-api-access-5bdl7\") pod \"dnsmasq-dns-7789f5c469-8qpf5\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.995664 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-ovsdbserver-nb\") pod \"dnsmasq-dns-7789f5c469-8qpf5\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.995796 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-config\") pod \"dnsmasq-dns-7789f5c469-8qpf5\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:30 crc kubenswrapper[4775]: I1002 03:06:30.995843 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-dns-svc\") pod \"dnsmasq-dns-7789f5c469-8qpf5\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.096404 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bdl7\" (UniqueName: \"kubernetes.io/projected/7f514b1c-ebee-4a89-be14-59841552e497-kube-api-access-5bdl7\") pod \"dnsmasq-dns-7789f5c469-8qpf5\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.096464 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-ovsdbserver-nb\") pod \"dnsmasq-dns-7789f5c469-8qpf5\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.096516 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-config\") pod \"dnsmasq-dns-7789f5c469-8qpf5\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.096543 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-dns-svc\") pod \"dnsmasq-dns-7789f5c469-8qpf5\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.097449 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-config\") pod \"dnsmasq-dns-7789f5c469-8qpf5\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.097462 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-ovsdbserver-nb\") pod \"dnsmasq-dns-7789f5c469-8qpf5\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.097512 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-dns-svc\") pod \"dnsmasq-dns-7789f5c469-8qpf5\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.129092 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bdl7\" (UniqueName: \"kubernetes.io/projected/7f514b1c-ebee-4a89-be14-59841552e497-kube-api-access-5bdl7\") pod \"dnsmasq-dns-7789f5c469-8qpf5\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.138668 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7789f5c469-8qpf5"] Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.139746 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.170067 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dff454985-787r8"] Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.171743 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.175052 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.180936 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dff454985-787r8"] Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.198238 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-ovsdbserver-nb\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.198292 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-config\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.198322 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-ovsdbserver-sb\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.198345 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw6wg\" (UniqueName: \"kubernetes.io/projected/7480c773-8658-4a0d-9150-eb0b3ecc7211-kube-api-access-mw6wg\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.198419 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-dns-svc\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.302884 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-config\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.303180 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-ovsdbserver-sb\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.303203 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw6wg\" (UniqueName: \"kubernetes.io/projected/7480c773-8658-4a0d-9150-eb0b3ecc7211-kube-api-access-mw6wg\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.303267 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-dns-svc\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.303332 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-ovsdbserver-nb\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.304142 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-ovsdbserver-nb\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.304406 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-ovsdbserver-sb\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.304491 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-dns-svc\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.304792 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-config\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.323011 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw6wg\" (UniqueName: \"kubernetes.io/projected/7480c773-8658-4a0d-9150-eb0b3ecc7211-kube-api-access-mw6wg\") pod \"dnsmasq-dns-dff454985-787r8\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.547236 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.655416 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7789f5c469-8qpf5"] Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.828543 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dff454985-787r8"] Oct 02 03:06:31 crc kubenswrapper[4775]: W1002 03:06:31.830341 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7480c773_8658_4a0d_9150_eb0b3ecc7211.slice/crio-7355e90447b9d841e249c32b96ec0b78ada8bca55ca4bb8856fe41d32321a98d WatchSource:0}: Error finding container 7355e90447b9d841e249c32b96ec0b78ada8bca55ca4bb8856fe41d32321a98d: Status 404 returned error can't find the container with id 7355e90447b9d841e249c32b96ec0b78ada8bca55ca4bb8856fe41d32321a98d Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.925123 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dff454985-787r8" event={"ID":"7480c773-8658-4a0d-9150-eb0b3ecc7211","Type":"ContainerStarted","Data":"7355e90447b9d841e249c32b96ec0b78ada8bca55ca4bb8856fe41d32321a98d"} Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.929414 4775 generic.go:334] "Generic (PLEG): container finished" podID="7f514b1c-ebee-4a89-be14-59841552e497" containerID="ebc2f5de553b834f1b0821db4c39c2ac9c04b55c87d208fa3dc1849a28045757" exitCode=0 Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.929461 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" event={"ID":"7f514b1c-ebee-4a89-be14-59841552e497","Type":"ContainerDied","Data":"ebc2f5de553b834f1b0821db4c39c2ac9c04b55c87d208fa3dc1849a28045757"} Oct 02 03:06:31 crc kubenswrapper[4775]: I1002 03:06:31.929504 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" event={"ID":"7f514b1c-ebee-4a89-be14-59841552e497","Type":"ContainerStarted","Data":"7004bdc99194e6c067df1201e6c9fd855699c8ba733a899bbf682fd3647ece47"} Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.334340 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.433809 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-config\") pod \"7f514b1c-ebee-4a89-be14-59841552e497\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.433996 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-ovsdbserver-nb\") pod \"7f514b1c-ebee-4a89-be14-59841552e497\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.434065 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bdl7\" (UniqueName: \"kubernetes.io/projected/7f514b1c-ebee-4a89-be14-59841552e497-kube-api-access-5bdl7\") pod \"7f514b1c-ebee-4a89-be14-59841552e497\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.434158 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-dns-svc\") pod \"7f514b1c-ebee-4a89-be14-59841552e497\" (UID: \"7f514b1c-ebee-4a89-be14-59841552e497\") " Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.440454 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f514b1c-ebee-4a89-be14-59841552e497-kube-api-access-5bdl7" (OuterVolumeSpecName: "kube-api-access-5bdl7") pod "7f514b1c-ebee-4a89-be14-59841552e497" (UID: "7f514b1c-ebee-4a89-be14-59841552e497"). InnerVolumeSpecName "kube-api-access-5bdl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.456163 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7f514b1c-ebee-4a89-be14-59841552e497" (UID: "7f514b1c-ebee-4a89-be14-59841552e497"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.459056 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7f514b1c-ebee-4a89-be14-59841552e497" (UID: "7f514b1c-ebee-4a89-be14-59841552e497"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.463591 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-config" (OuterVolumeSpecName: "config") pod "7f514b1c-ebee-4a89-be14-59841552e497" (UID: "7f514b1c-ebee-4a89-be14-59841552e497"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.536619 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.536670 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.536692 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f514b1c-ebee-4a89-be14-59841552e497-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.536716 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bdl7\" (UniqueName: \"kubernetes.io/projected/7f514b1c-ebee-4a89-be14-59841552e497-kube-api-access-5bdl7\") on node \"crc\" DevicePath \"\"" Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.942093 4775 generic.go:334] "Generic (PLEG): container finished" podID="7480c773-8658-4a0d-9150-eb0b3ecc7211" containerID="aa012c6ebc8c7bd2102d53e5acab2b3196cd4feb3263df17135cf748bcbcf98e" exitCode=0 Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.942146 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dff454985-787r8" event={"ID":"7480c773-8658-4a0d-9150-eb0b3ecc7211","Type":"ContainerDied","Data":"aa012c6ebc8c7bd2102d53e5acab2b3196cd4feb3263df17135cf748bcbcf98e"} Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.944015 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" event={"ID":"7f514b1c-ebee-4a89-be14-59841552e497","Type":"ContainerDied","Data":"7004bdc99194e6c067df1201e6c9fd855699c8ba733a899bbf682fd3647ece47"} Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.944066 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7789f5c469-8qpf5" Oct 02 03:06:32 crc kubenswrapper[4775]: I1002 03:06:32.944090 4775 scope.go:117] "RemoveContainer" containerID="ebc2f5de553b834f1b0821db4c39c2ac9c04b55c87d208fa3dc1849a28045757" Oct 02 03:06:33 crc kubenswrapper[4775]: I1002 03:06:33.036738 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7789f5c469-8qpf5"] Oct 02 03:06:33 crc kubenswrapper[4775]: I1002 03:06:33.046902 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7789f5c469-8qpf5"] Oct 02 03:06:33 crc kubenswrapper[4775]: I1002 03:06:33.785999 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f514b1c-ebee-4a89-be14-59841552e497" path="/var/lib/kubelet/pods/7f514b1c-ebee-4a89-be14-59841552e497/volumes" Oct 02 03:06:33 crc kubenswrapper[4775]: I1002 03:06:33.959183 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dff454985-787r8" event={"ID":"7480c773-8658-4a0d-9150-eb0b3ecc7211","Type":"ContainerStarted","Data":"41dae23bbe5d004c673724794d9350d7a365967d3597529bf8e479e1d294d8d9"} Oct 02 03:06:33 crc kubenswrapper[4775]: I1002 03:06:33.959459 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:33 crc kubenswrapper[4775]: I1002 03:06:33.992774 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dff454985-787r8" podStartSLOduration=2.992757111 podStartE2EDuration="2.992757111s" podCreationTimestamp="2025-10-02 03:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:06:33.984768855 +0000 UTC m=+5131.151512915" watchObservedRunningTime="2025-10-02 03:06:33.992757111 +0000 UTC m=+5131.159501161" Oct 02 03:06:34 crc kubenswrapper[4775]: I1002 03:06:34.655943 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 02 03:06:34 crc kubenswrapper[4775]: I1002 03:06:34.739394 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.386770 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Oct 02 03:06:37 crc kubenswrapper[4775]: E1002 03:06:37.387826 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f514b1c-ebee-4a89-be14-59841552e497" containerName="init" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.387850 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f514b1c-ebee-4a89-be14-59841552e497" containerName="init" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.388144 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f514b1c-ebee-4a89-be14-59841552e497" containerName="init" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.389049 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.392344 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.397211 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.535132 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2c2l\" (UniqueName: \"kubernetes.io/projected/4dc66a88-f227-4e53-b079-20480022af30-kube-api-access-j2c2l\") pod \"ovn-copy-data\" (UID: \"4dc66a88-f227-4e53-b079-20480022af30\") " pod="openstack/ovn-copy-data" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.535355 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b1cbce3e-283d-44b5-89d2-4ff9f2456a4b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b1cbce3e-283d-44b5-89d2-4ff9f2456a4b\") pod \"ovn-copy-data\" (UID: \"4dc66a88-f227-4e53-b079-20480022af30\") " pod="openstack/ovn-copy-data" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.535575 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/4dc66a88-f227-4e53-b079-20480022af30-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"4dc66a88-f227-4e53-b079-20480022af30\") " pod="openstack/ovn-copy-data" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.637161 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/4dc66a88-f227-4e53-b079-20480022af30-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"4dc66a88-f227-4e53-b079-20480022af30\") " pod="openstack/ovn-copy-data" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.637719 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2c2l\" (UniqueName: \"kubernetes.io/projected/4dc66a88-f227-4e53-b079-20480022af30-kube-api-access-j2c2l\") pod \"ovn-copy-data\" (UID: \"4dc66a88-f227-4e53-b079-20480022af30\") " pod="openstack/ovn-copy-data" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.637842 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b1cbce3e-283d-44b5-89d2-4ff9f2456a4b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b1cbce3e-283d-44b5-89d2-4ff9f2456a4b\") pod \"ovn-copy-data\" (UID: \"4dc66a88-f227-4e53-b079-20480022af30\") " pod="openstack/ovn-copy-data" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.644065 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.644211 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b1cbce3e-283d-44b5-89d2-4ff9f2456a4b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b1cbce3e-283d-44b5-89d2-4ff9f2456a4b\") pod \"ovn-copy-data\" (UID: \"4dc66a88-f227-4e53-b079-20480022af30\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f3e18ac5c4c14f75f00258d2edcbe7622f16af5a651cfd1c116fe64b3af7c3bb/globalmount\"" pod="openstack/ovn-copy-data" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.646391 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/4dc66a88-f227-4e53-b079-20480022af30-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"4dc66a88-f227-4e53-b079-20480022af30\") " pod="openstack/ovn-copy-data" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.668564 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2c2l\" (UniqueName: \"kubernetes.io/projected/4dc66a88-f227-4e53-b079-20480022af30-kube-api-access-j2c2l\") pod \"ovn-copy-data\" (UID: \"4dc66a88-f227-4e53-b079-20480022af30\") " pod="openstack/ovn-copy-data" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.676143 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b1cbce3e-283d-44b5-89d2-4ff9f2456a4b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b1cbce3e-283d-44b5-89d2-4ff9f2456a4b\") pod \"ovn-copy-data\" (UID: \"4dc66a88-f227-4e53-b079-20480022af30\") " pod="openstack/ovn-copy-data" Oct 02 03:06:37 crc kubenswrapper[4775]: I1002 03:06:37.724710 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Oct 02 03:06:38 crc kubenswrapper[4775]: I1002 03:06:38.331399 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Oct 02 03:06:38 crc kubenswrapper[4775]: W1002 03:06:38.342045 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4dc66a88_f227_4e53_b079_20480022af30.slice/crio-f4f8cf6b75128e4046cdc586b008a9e066a6c0606fafea06a65497fe66d7b143 WatchSource:0}: Error finding container f4f8cf6b75128e4046cdc586b008a9e066a6c0606fafea06a65497fe66d7b143: Status 404 returned error can't find the container with id f4f8cf6b75128e4046cdc586b008a9e066a6c0606fafea06a65497fe66d7b143 Oct 02 03:06:39 crc kubenswrapper[4775]: I1002 03:06:39.008113 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"4dc66a88-f227-4e53-b079-20480022af30","Type":"ContainerStarted","Data":"4fa20029e1e23a43b3f83646d241b54b43a98d2547aa56587370084c5a8c10f0"} Oct 02 03:06:39 crc kubenswrapper[4775]: I1002 03:06:39.008452 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"4dc66a88-f227-4e53-b079-20480022af30","Type":"ContainerStarted","Data":"f4f8cf6b75128e4046cdc586b008a9e066a6c0606fafea06a65497fe66d7b143"} Oct 02 03:06:39 crc kubenswrapper[4775]: I1002 03:06:39.035234 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.03521071 podStartE2EDuration="3.03521071s" podCreationTimestamp="2025-10-02 03:06:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:06:39.031367901 +0000 UTC m=+5136.198111971" watchObservedRunningTime="2025-10-02 03:06:39.03521071 +0000 UTC m=+5136.201954780" Oct 02 03:06:41 crc kubenswrapper[4775]: I1002 03:06:41.550308 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:06:41 crc kubenswrapper[4775]: I1002 03:06:41.638116 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-bgzg5"] Oct 02 03:06:41 crc kubenswrapper[4775]: I1002 03:06:41.638420 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" podUID="186f8050-b541-4d3f-b79a-8113264792d8" containerName="dnsmasq-dns" containerID="cri-o://7a70bee8fc23fff70a91faa6041da6b3f16970a6325dc89ce6826d7b9f587102" gracePeriod=10 Oct 02 03:06:42 crc kubenswrapper[4775]: I1002 03:06:42.039335 4775 generic.go:334] "Generic (PLEG): container finished" podID="186f8050-b541-4d3f-b79a-8113264792d8" containerID="7a70bee8fc23fff70a91faa6041da6b3f16970a6325dc89ce6826d7b9f587102" exitCode=0 Oct 02 03:06:42 crc kubenswrapper[4775]: I1002 03:06:42.039607 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" event={"ID":"186f8050-b541-4d3f-b79a-8113264792d8","Type":"ContainerDied","Data":"7a70bee8fc23fff70a91faa6041da6b3f16970a6325dc89ce6826d7b9f587102"} Oct 02 03:06:42 crc kubenswrapper[4775]: I1002 03:06:42.101482 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 03:06:42 crc kubenswrapper[4775]: I1002 03:06:42.233396 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/186f8050-b541-4d3f-b79a-8113264792d8-dns-svc\") pod \"186f8050-b541-4d3f-b79a-8113264792d8\" (UID: \"186f8050-b541-4d3f-b79a-8113264792d8\") " Oct 02 03:06:42 crc kubenswrapper[4775]: I1002 03:06:42.233498 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xcxd\" (UniqueName: \"kubernetes.io/projected/186f8050-b541-4d3f-b79a-8113264792d8-kube-api-access-7xcxd\") pod \"186f8050-b541-4d3f-b79a-8113264792d8\" (UID: \"186f8050-b541-4d3f-b79a-8113264792d8\") " Oct 02 03:06:42 crc kubenswrapper[4775]: I1002 03:06:42.233586 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/186f8050-b541-4d3f-b79a-8113264792d8-config\") pod \"186f8050-b541-4d3f-b79a-8113264792d8\" (UID: \"186f8050-b541-4d3f-b79a-8113264792d8\") " Oct 02 03:06:42 crc kubenswrapper[4775]: I1002 03:06:42.244237 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/186f8050-b541-4d3f-b79a-8113264792d8-kube-api-access-7xcxd" (OuterVolumeSpecName: "kube-api-access-7xcxd") pod "186f8050-b541-4d3f-b79a-8113264792d8" (UID: "186f8050-b541-4d3f-b79a-8113264792d8"). InnerVolumeSpecName "kube-api-access-7xcxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:06:42 crc kubenswrapper[4775]: I1002 03:06:42.292815 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/186f8050-b541-4d3f-b79a-8113264792d8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "186f8050-b541-4d3f-b79a-8113264792d8" (UID: "186f8050-b541-4d3f-b79a-8113264792d8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:06:42 crc kubenswrapper[4775]: I1002 03:06:42.307301 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/186f8050-b541-4d3f-b79a-8113264792d8-config" (OuterVolumeSpecName: "config") pod "186f8050-b541-4d3f-b79a-8113264792d8" (UID: "186f8050-b541-4d3f-b79a-8113264792d8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:06:42 crc kubenswrapper[4775]: I1002 03:06:42.336108 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/186f8050-b541-4d3f-b79a-8113264792d8-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 03:06:42 crc kubenswrapper[4775]: I1002 03:06:42.336149 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xcxd\" (UniqueName: \"kubernetes.io/projected/186f8050-b541-4d3f-b79a-8113264792d8-kube-api-access-7xcxd\") on node \"crc\" DevicePath \"\"" Oct 02 03:06:42 crc kubenswrapper[4775]: I1002 03:06:42.336171 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/186f8050-b541-4d3f-b79a-8113264792d8-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:06:42 crc kubenswrapper[4775]: I1002 03:06:42.766200 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:06:42 crc kubenswrapper[4775]: E1002 03:06:42.766854 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:06:43 crc kubenswrapper[4775]: I1002 03:06:43.051941 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" event={"ID":"186f8050-b541-4d3f-b79a-8113264792d8","Type":"ContainerDied","Data":"00150e8dc885934dc8db44b73e8b1931b217bd4ef1c97b7a636762bc9920d340"} Oct 02 03:06:43 crc kubenswrapper[4775]: I1002 03:06:43.052005 4775 scope.go:117] "RemoveContainer" containerID="7a70bee8fc23fff70a91faa6041da6b3f16970a6325dc89ce6826d7b9f587102" Oct 02 03:06:43 crc kubenswrapper[4775]: I1002 03:06:43.052267 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-bgzg5" Oct 02 03:06:43 crc kubenswrapper[4775]: I1002 03:06:43.081478 4775 scope.go:117] "RemoveContainer" containerID="9a670f7be5f11680f800b7815672b4f6d8fac5de283a0f31aed8b01effed334d" Oct 02 03:06:43 crc kubenswrapper[4775]: I1002 03:06:43.109695 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-bgzg5"] Oct 02 03:06:43 crc kubenswrapper[4775]: I1002 03:06:43.123056 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-bgzg5"] Oct 02 03:06:43 crc kubenswrapper[4775]: E1002 03:06:43.504475 4775 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.47:37270->38.102.83.47:45167: read tcp 38.102.83.47:37270->38.102.83.47:45167: read: connection reset by peer Oct 02 03:06:43 crc kubenswrapper[4775]: I1002 03:06:43.783940 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="186f8050-b541-4d3f-b79a-8113264792d8" path="/var/lib/kubelet/pods/186f8050-b541-4d3f-b79a-8113264792d8/volumes" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.514022 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 02 03:06:44 crc kubenswrapper[4775]: E1002 03:06:44.514476 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="186f8050-b541-4d3f-b79a-8113264792d8" containerName="dnsmasq-dns" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.514495 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="186f8050-b541-4d3f-b79a-8113264792d8" containerName="dnsmasq-dns" Oct 02 03:06:44 crc kubenswrapper[4775]: E1002 03:06:44.514538 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="186f8050-b541-4d3f-b79a-8113264792d8" containerName="init" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.514546 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="186f8050-b541-4d3f-b79a-8113264792d8" containerName="init" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.514773 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="186f8050-b541-4d3f-b79a-8113264792d8" containerName="dnsmasq-dns" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.516142 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.519835 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-l4pnk" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.520061 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.526112 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.526212 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.679590 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-config\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.679751 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw672\" (UniqueName: \"kubernetes.io/projected/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-kube-api-access-zw672\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.679775 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-scripts\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.679798 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.679856 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.781356 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw672\" (UniqueName: \"kubernetes.io/projected/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-kube-api-access-zw672\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.781449 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-scripts\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.781500 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.781597 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.781783 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-config\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.782190 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.782715 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-scripts\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.783494 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-config\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.788210 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.802882 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw672\" (UniqueName: \"kubernetes.io/projected/04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671-kube-api-access-zw672\") pod \"ovn-northd-0\" (UID: \"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671\") " pod="openstack/ovn-northd-0" Oct 02 03:06:44 crc kubenswrapper[4775]: I1002 03:06:44.845190 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 02 03:06:45 crc kubenswrapper[4775]: I1002 03:06:45.344347 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 02 03:06:45 crc kubenswrapper[4775]: W1002 03:06:45.351843 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04bab7f7_0bbf_4cd8_9f93_c0f8f07b4671.slice/crio-a67c96bcbb78a2fbe094dbc08be78ea9a09cd3d2efd23d1d01364f09ec3f366c WatchSource:0}: Error finding container a67c96bcbb78a2fbe094dbc08be78ea9a09cd3d2efd23d1d01364f09ec3f366c: Status 404 returned error can't find the container with id a67c96bcbb78a2fbe094dbc08be78ea9a09cd3d2efd23d1d01364f09ec3f366c Oct 02 03:06:46 crc kubenswrapper[4775]: I1002 03:06:46.082754 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671","Type":"ContainerStarted","Data":"ebe1cc157ba86b066925b78c56b88b2e310a06f81b35e09f06351492af0832f3"} Oct 02 03:06:46 crc kubenswrapper[4775]: I1002 03:06:46.083287 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671","Type":"ContainerStarted","Data":"ebe21132e8a0aeb1fd14c9387e78272b3f541ee241f0b578a8d41a00ad5aec90"} Oct 02 03:06:46 crc kubenswrapper[4775]: I1002 03:06:46.083320 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 02 03:06:46 crc kubenswrapper[4775]: I1002 03:06:46.083339 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671","Type":"ContainerStarted","Data":"a67c96bcbb78a2fbe094dbc08be78ea9a09cd3d2efd23d1d01364f09ec3f366c"} Oct 02 03:06:46 crc kubenswrapper[4775]: I1002 03:06:46.134473 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.134437349 podStartE2EDuration="2.134437349s" podCreationTimestamp="2025-10-02 03:06:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:06:46.118882678 +0000 UTC m=+5143.285626748" watchObservedRunningTime="2025-10-02 03:06:46.134437349 +0000 UTC m=+5143.301181429" Oct 02 03:06:50 crc kubenswrapper[4775]: I1002 03:06:50.207425 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-kfp5w"] Oct 02 03:06:50 crc kubenswrapper[4775]: I1002 03:06:50.209473 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kfp5w" Oct 02 03:06:50 crc kubenswrapper[4775]: I1002 03:06:50.216181 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-kfp5w"] Oct 02 03:06:50 crc kubenswrapper[4775]: I1002 03:06:50.287774 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qw6wv\" (UniqueName: \"kubernetes.io/projected/a23444b7-1f4c-4927-a42e-fc8703728295-kube-api-access-qw6wv\") pod \"keystone-db-create-kfp5w\" (UID: \"a23444b7-1f4c-4927-a42e-fc8703728295\") " pod="openstack/keystone-db-create-kfp5w" Oct 02 03:06:50 crc kubenswrapper[4775]: I1002 03:06:50.388870 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qw6wv\" (UniqueName: \"kubernetes.io/projected/a23444b7-1f4c-4927-a42e-fc8703728295-kube-api-access-qw6wv\") pod \"keystone-db-create-kfp5w\" (UID: \"a23444b7-1f4c-4927-a42e-fc8703728295\") " pod="openstack/keystone-db-create-kfp5w" Oct 02 03:06:50 crc kubenswrapper[4775]: I1002 03:06:50.423067 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qw6wv\" (UniqueName: \"kubernetes.io/projected/a23444b7-1f4c-4927-a42e-fc8703728295-kube-api-access-qw6wv\") pod \"keystone-db-create-kfp5w\" (UID: \"a23444b7-1f4c-4927-a42e-fc8703728295\") " pod="openstack/keystone-db-create-kfp5w" Oct 02 03:06:50 crc kubenswrapper[4775]: I1002 03:06:50.537135 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kfp5w" Oct 02 03:06:51 crc kubenswrapper[4775]: I1002 03:06:51.047539 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-kfp5w"] Oct 02 03:06:51 crc kubenswrapper[4775]: W1002 03:06:51.047611 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda23444b7_1f4c_4927_a42e_fc8703728295.slice/crio-4b114d4e0d423e3d7a9d9d5fab33788ffea4ef5db11a8c6f7030d2af51ccef27 WatchSource:0}: Error finding container 4b114d4e0d423e3d7a9d9d5fab33788ffea4ef5db11a8c6f7030d2af51ccef27: Status 404 returned error can't find the container with id 4b114d4e0d423e3d7a9d9d5fab33788ffea4ef5db11a8c6f7030d2af51ccef27 Oct 02 03:06:51 crc kubenswrapper[4775]: I1002 03:06:51.135201 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-kfp5w" event={"ID":"a23444b7-1f4c-4927-a42e-fc8703728295","Type":"ContainerStarted","Data":"4b114d4e0d423e3d7a9d9d5fab33788ffea4ef5db11a8c6f7030d2af51ccef27"} Oct 02 03:06:52 crc kubenswrapper[4775]: I1002 03:06:52.147668 4775 generic.go:334] "Generic (PLEG): container finished" podID="a23444b7-1f4c-4927-a42e-fc8703728295" containerID="882b4cc388ee5d3e24b19519ebf9c9b39356accb865da6007f8a8f056098640c" exitCode=0 Oct 02 03:06:52 crc kubenswrapper[4775]: I1002 03:06:52.147743 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-kfp5w" event={"ID":"a23444b7-1f4c-4927-a42e-fc8703728295","Type":"ContainerDied","Data":"882b4cc388ee5d3e24b19519ebf9c9b39356accb865da6007f8a8f056098640c"} Oct 02 03:06:53 crc kubenswrapper[4775]: I1002 03:06:53.576639 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kfp5w" Oct 02 03:06:53 crc kubenswrapper[4775]: I1002 03:06:53.650005 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qw6wv\" (UniqueName: \"kubernetes.io/projected/a23444b7-1f4c-4927-a42e-fc8703728295-kube-api-access-qw6wv\") pod \"a23444b7-1f4c-4927-a42e-fc8703728295\" (UID: \"a23444b7-1f4c-4927-a42e-fc8703728295\") " Oct 02 03:06:53 crc kubenswrapper[4775]: I1002 03:06:53.660250 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a23444b7-1f4c-4927-a42e-fc8703728295-kube-api-access-qw6wv" (OuterVolumeSpecName: "kube-api-access-qw6wv") pod "a23444b7-1f4c-4927-a42e-fc8703728295" (UID: "a23444b7-1f4c-4927-a42e-fc8703728295"). InnerVolumeSpecName "kube-api-access-qw6wv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:06:53 crc kubenswrapper[4775]: I1002 03:06:53.752773 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qw6wv\" (UniqueName: \"kubernetes.io/projected/a23444b7-1f4c-4927-a42e-fc8703728295-kube-api-access-qw6wv\") on node \"crc\" DevicePath \"\"" Oct 02 03:06:53 crc kubenswrapper[4775]: I1002 03:06:53.780249 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:06:53 crc kubenswrapper[4775]: E1002 03:06:53.780637 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:06:54 crc kubenswrapper[4775]: I1002 03:06:54.169781 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-kfp5w" event={"ID":"a23444b7-1f4c-4927-a42e-fc8703728295","Type":"ContainerDied","Data":"4b114d4e0d423e3d7a9d9d5fab33788ffea4ef5db11a8c6f7030d2af51ccef27"} Oct 02 03:06:54 crc kubenswrapper[4775]: I1002 03:06:54.169837 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b114d4e0d423e3d7a9d9d5fab33788ffea4ef5db11a8c6f7030d2af51ccef27" Oct 02 03:06:54 crc kubenswrapper[4775]: I1002 03:06:54.170342 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-kfp5w" Oct 02 03:06:59 crc kubenswrapper[4775]: I1002 03:06:59.944447 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 02 03:07:00 crc kubenswrapper[4775]: I1002 03:07:00.222027 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-3671-account-create-szs6m"] Oct 02 03:07:00 crc kubenswrapper[4775]: E1002 03:07:00.222583 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a23444b7-1f4c-4927-a42e-fc8703728295" containerName="mariadb-database-create" Oct 02 03:07:00 crc kubenswrapper[4775]: I1002 03:07:00.222605 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a23444b7-1f4c-4927-a42e-fc8703728295" containerName="mariadb-database-create" Oct 02 03:07:00 crc kubenswrapper[4775]: I1002 03:07:00.222815 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a23444b7-1f4c-4927-a42e-fc8703728295" containerName="mariadb-database-create" Oct 02 03:07:00 crc kubenswrapper[4775]: I1002 03:07:00.223478 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3671-account-create-szs6m" Oct 02 03:07:00 crc kubenswrapper[4775]: I1002 03:07:00.226711 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 02 03:07:00 crc kubenswrapper[4775]: I1002 03:07:00.243265 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3671-account-create-szs6m"] Oct 02 03:07:00 crc kubenswrapper[4775]: I1002 03:07:00.298396 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdz6k\" (UniqueName: \"kubernetes.io/projected/1e2dc5c3-2bcb-476b-9220-8643c983a17b-kube-api-access-tdz6k\") pod \"keystone-3671-account-create-szs6m\" (UID: \"1e2dc5c3-2bcb-476b-9220-8643c983a17b\") " pod="openstack/keystone-3671-account-create-szs6m" Oct 02 03:07:00 crc kubenswrapper[4775]: I1002 03:07:00.399859 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdz6k\" (UniqueName: \"kubernetes.io/projected/1e2dc5c3-2bcb-476b-9220-8643c983a17b-kube-api-access-tdz6k\") pod \"keystone-3671-account-create-szs6m\" (UID: \"1e2dc5c3-2bcb-476b-9220-8643c983a17b\") " pod="openstack/keystone-3671-account-create-szs6m" Oct 02 03:07:00 crc kubenswrapper[4775]: I1002 03:07:00.434881 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdz6k\" (UniqueName: \"kubernetes.io/projected/1e2dc5c3-2bcb-476b-9220-8643c983a17b-kube-api-access-tdz6k\") pod \"keystone-3671-account-create-szs6m\" (UID: \"1e2dc5c3-2bcb-476b-9220-8643c983a17b\") " pod="openstack/keystone-3671-account-create-szs6m" Oct 02 03:07:00 crc kubenswrapper[4775]: I1002 03:07:00.553376 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3671-account-create-szs6m" Oct 02 03:07:01 crc kubenswrapper[4775]: I1002 03:07:01.037345 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3671-account-create-szs6m"] Oct 02 03:07:01 crc kubenswrapper[4775]: I1002 03:07:01.243316 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3671-account-create-szs6m" event={"ID":"1e2dc5c3-2bcb-476b-9220-8643c983a17b","Type":"ContainerStarted","Data":"c945add0fbcf34ed6f9d3387d1905b090be1149b84ae6fd953b445b98005ecbc"} Oct 02 03:07:02 crc kubenswrapper[4775]: I1002 03:07:02.259377 4775 generic.go:334] "Generic (PLEG): container finished" podID="1e2dc5c3-2bcb-476b-9220-8643c983a17b" containerID="1ad1ca145a2617ab096b2670115314f7c65e16bd32d25181a414eecf4ead577b" exitCode=0 Oct 02 03:07:02 crc kubenswrapper[4775]: I1002 03:07:02.259460 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3671-account-create-szs6m" event={"ID":"1e2dc5c3-2bcb-476b-9220-8643c983a17b","Type":"ContainerDied","Data":"1ad1ca145a2617ab096b2670115314f7c65e16bd32d25181a414eecf4ead577b"} Oct 02 03:07:03 crc kubenswrapper[4775]: I1002 03:07:03.684932 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3671-account-create-szs6m" Oct 02 03:07:03 crc kubenswrapper[4775]: I1002 03:07:03.768390 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdz6k\" (UniqueName: \"kubernetes.io/projected/1e2dc5c3-2bcb-476b-9220-8643c983a17b-kube-api-access-tdz6k\") pod \"1e2dc5c3-2bcb-476b-9220-8643c983a17b\" (UID: \"1e2dc5c3-2bcb-476b-9220-8643c983a17b\") " Oct 02 03:07:03 crc kubenswrapper[4775]: I1002 03:07:03.780510 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e2dc5c3-2bcb-476b-9220-8643c983a17b-kube-api-access-tdz6k" (OuterVolumeSpecName: "kube-api-access-tdz6k") pod "1e2dc5c3-2bcb-476b-9220-8643c983a17b" (UID: "1e2dc5c3-2bcb-476b-9220-8643c983a17b"). InnerVolumeSpecName "kube-api-access-tdz6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:07:03 crc kubenswrapper[4775]: I1002 03:07:03.872474 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdz6k\" (UniqueName: \"kubernetes.io/projected/1e2dc5c3-2bcb-476b-9220-8643c983a17b-kube-api-access-tdz6k\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:04 crc kubenswrapper[4775]: I1002 03:07:04.302823 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3671-account-create-szs6m" event={"ID":"1e2dc5c3-2bcb-476b-9220-8643c983a17b","Type":"ContainerDied","Data":"c945add0fbcf34ed6f9d3387d1905b090be1149b84ae6fd953b445b98005ecbc"} Oct 02 03:07:04 crc kubenswrapper[4775]: I1002 03:07:04.302878 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c945add0fbcf34ed6f9d3387d1905b090be1149b84ae6fd953b445b98005ecbc" Oct 02 03:07:04 crc kubenswrapper[4775]: I1002 03:07:04.303024 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3671-account-create-szs6m" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.639434 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-nc6qh"] Oct 02 03:07:05 crc kubenswrapper[4775]: E1002 03:07:05.639915 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e2dc5c3-2bcb-476b-9220-8643c983a17b" containerName="mariadb-account-create" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.641209 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e2dc5c3-2bcb-476b-9220-8643c983a17b" containerName="mariadb-account-create" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.641707 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e2dc5c3-2bcb-476b-9220-8643c983a17b" containerName="mariadb-account-create" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.642520 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nc6qh" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.645818 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.646262 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5tx2s" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.652553 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.655776 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-nc6qh"] Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.656911 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.717025 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf757e5-76fb-4aee-80c3-d362a58a977b-combined-ca-bundle\") pod \"keystone-db-sync-nc6qh\" (UID: \"daf757e5-76fb-4aee-80c3-d362a58a977b\") " pod="openstack/keystone-db-sync-nc6qh" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.717079 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf757e5-76fb-4aee-80c3-d362a58a977b-config-data\") pod \"keystone-db-sync-nc6qh\" (UID: \"daf757e5-76fb-4aee-80c3-d362a58a977b\") " pod="openstack/keystone-db-sync-nc6qh" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.717109 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6lht\" (UniqueName: \"kubernetes.io/projected/daf757e5-76fb-4aee-80c3-d362a58a977b-kube-api-access-z6lht\") pod \"keystone-db-sync-nc6qh\" (UID: \"daf757e5-76fb-4aee-80c3-d362a58a977b\") " pod="openstack/keystone-db-sync-nc6qh" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.766255 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:07:05 crc kubenswrapper[4775]: E1002 03:07:05.766469 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.819259 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf757e5-76fb-4aee-80c3-d362a58a977b-combined-ca-bundle\") pod \"keystone-db-sync-nc6qh\" (UID: \"daf757e5-76fb-4aee-80c3-d362a58a977b\") " pod="openstack/keystone-db-sync-nc6qh" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.819354 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf757e5-76fb-4aee-80c3-d362a58a977b-config-data\") pod \"keystone-db-sync-nc6qh\" (UID: \"daf757e5-76fb-4aee-80c3-d362a58a977b\") " pod="openstack/keystone-db-sync-nc6qh" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.819388 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6lht\" (UniqueName: \"kubernetes.io/projected/daf757e5-76fb-4aee-80c3-d362a58a977b-kube-api-access-z6lht\") pod \"keystone-db-sync-nc6qh\" (UID: \"daf757e5-76fb-4aee-80c3-d362a58a977b\") " pod="openstack/keystone-db-sync-nc6qh" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.825324 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf757e5-76fb-4aee-80c3-d362a58a977b-combined-ca-bundle\") pod \"keystone-db-sync-nc6qh\" (UID: \"daf757e5-76fb-4aee-80c3-d362a58a977b\") " pod="openstack/keystone-db-sync-nc6qh" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.834669 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf757e5-76fb-4aee-80c3-d362a58a977b-config-data\") pod \"keystone-db-sync-nc6qh\" (UID: \"daf757e5-76fb-4aee-80c3-d362a58a977b\") " pod="openstack/keystone-db-sync-nc6qh" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.848002 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6lht\" (UniqueName: \"kubernetes.io/projected/daf757e5-76fb-4aee-80c3-d362a58a977b-kube-api-access-z6lht\") pod \"keystone-db-sync-nc6qh\" (UID: \"daf757e5-76fb-4aee-80c3-d362a58a977b\") " pod="openstack/keystone-db-sync-nc6qh" Oct 02 03:07:05 crc kubenswrapper[4775]: I1002 03:07:05.986497 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nc6qh" Oct 02 03:07:06 crc kubenswrapper[4775]: I1002 03:07:06.446346 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-nc6qh"] Oct 02 03:07:06 crc kubenswrapper[4775]: W1002 03:07:06.460619 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddaf757e5_76fb_4aee_80c3_d362a58a977b.slice/crio-19caed43843d1922751cff275809551dfbffac2ee125078f761c3f7439cabed2 WatchSource:0}: Error finding container 19caed43843d1922751cff275809551dfbffac2ee125078f761c3f7439cabed2: Status 404 returned error can't find the container with id 19caed43843d1922751cff275809551dfbffac2ee125078f761c3f7439cabed2 Oct 02 03:07:07 crc kubenswrapper[4775]: I1002 03:07:07.329919 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nc6qh" event={"ID":"daf757e5-76fb-4aee-80c3-d362a58a977b","Type":"ContainerStarted","Data":"5cd71b527c8d218a66d5ee9d6594d6ae90876f30023e50e8ba78b75de4541854"} Oct 02 03:07:07 crc kubenswrapper[4775]: I1002 03:07:07.330357 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nc6qh" event={"ID":"daf757e5-76fb-4aee-80c3-d362a58a977b","Type":"ContainerStarted","Data":"19caed43843d1922751cff275809551dfbffac2ee125078f761c3f7439cabed2"} Oct 02 03:07:07 crc kubenswrapper[4775]: I1002 03:07:07.354887 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-nc6qh" podStartSLOduration=2.354871145 podStartE2EDuration="2.354871145s" podCreationTimestamp="2025-10-02 03:07:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:07:07.35156541 +0000 UTC m=+5164.518309450" watchObservedRunningTime="2025-10-02 03:07:07.354871145 +0000 UTC m=+5164.521615185" Oct 02 03:07:08 crc kubenswrapper[4775]: I1002 03:07:08.348531 4775 generic.go:334] "Generic (PLEG): container finished" podID="daf757e5-76fb-4aee-80c3-d362a58a977b" containerID="5cd71b527c8d218a66d5ee9d6594d6ae90876f30023e50e8ba78b75de4541854" exitCode=0 Oct 02 03:07:08 crc kubenswrapper[4775]: I1002 03:07:08.348618 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nc6qh" event={"ID":"daf757e5-76fb-4aee-80c3-d362a58a977b","Type":"ContainerDied","Data":"5cd71b527c8d218a66d5ee9d6594d6ae90876f30023e50e8ba78b75de4541854"} Oct 02 03:07:09 crc kubenswrapper[4775]: I1002 03:07:09.751279 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nc6qh" Oct 02 03:07:09 crc kubenswrapper[4775]: I1002 03:07:09.796446 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf757e5-76fb-4aee-80c3-d362a58a977b-config-data\") pod \"daf757e5-76fb-4aee-80c3-d362a58a977b\" (UID: \"daf757e5-76fb-4aee-80c3-d362a58a977b\") " Oct 02 03:07:09 crc kubenswrapper[4775]: I1002 03:07:09.796711 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6lht\" (UniqueName: \"kubernetes.io/projected/daf757e5-76fb-4aee-80c3-d362a58a977b-kube-api-access-z6lht\") pod \"daf757e5-76fb-4aee-80c3-d362a58a977b\" (UID: \"daf757e5-76fb-4aee-80c3-d362a58a977b\") " Oct 02 03:07:09 crc kubenswrapper[4775]: I1002 03:07:09.796757 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf757e5-76fb-4aee-80c3-d362a58a977b-combined-ca-bundle\") pod \"daf757e5-76fb-4aee-80c3-d362a58a977b\" (UID: \"daf757e5-76fb-4aee-80c3-d362a58a977b\") " Oct 02 03:07:09 crc kubenswrapper[4775]: I1002 03:07:09.803665 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daf757e5-76fb-4aee-80c3-d362a58a977b-kube-api-access-z6lht" (OuterVolumeSpecName: "kube-api-access-z6lht") pod "daf757e5-76fb-4aee-80c3-d362a58a977b" (UID: "daf757e5-76fb-4aee-80c3-d362a58a977b"). InnerVolumeSpecName "kube-api-access-z6lht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:07:09 crc kubenswrapper[4775]: I1002 03:07:09.837445 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf757e5-76fb-4aee-80c3-d362a58a977b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "daf757e5-76fb-4aee-80c3-d362a58a977b" (UID: "daf757e5-76fb-4aee-80c3-d362a58a977b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:07:09 crc kubenswrapper[4775]: I1002 03:07:09.876130 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf757e5-76fb-4aee-80c3-d362a58a977b-config-data" (OuterVolumeSpecName: "config-data") pod "daf757e5-76fb-4aee-80c3-d362a58a977b" (UID: "daf757e5-76fb-4aee-80c3-d362a58a977b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:07:09 crc kubenswrapper[4775]: I1002 03:07:09.899306 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6lht\" (UniqueName: \"kubernetes.io/projected/daf757e5-76fb-4aee-80c3-d362a58a977b-kube-api-access-z6lht\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:09 crc kubenswrapper[4775]: I1002 03:07:09.899361 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf757e5-76fb-4aee-80c3-d362a58a977b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:09 crc kubenswrapper[4775]: I1002 03:07:09.899379 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf757e5-76fb-4aee-80c3-d362a58a977b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.379420 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-nc6qh" event={"ID":"daf757e5-76fb-4aee-80c3-d362a58a977b","Type":"ContainerDied","Data":"19caed43843d1922751cff275809551dfbffac2ee125078f761c3f7439cabed2"} Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.379771 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19caed43843d1922751cff275809551dfbffac2ee125078f761c3f7439cabed2" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.380270 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-nc6qh" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.522934 4775 scope.go:117] "RemoveContainer" containerID="98f3f05d32c954cc623282251bab7200b1197c5d21755e5900ba47c85b6b16c0" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.554091 4775 scope.go:117] "RemoveContainer" containerID="21d0aa3ba21eea81786f45e1aef1eca754fe91c36cb06f3cdf58ccb4edcf1dd7" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.585720 4775 scope.go:117] "RemoveContainer" containerID="af0ce6373af58eec3da363e858ca3f553944585860a600e6b0297212ab9b3264" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.613104 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f7dfc9995-xlkfd"] Oct 02 03:07:10 crc kubenswrapper[4775]: E1002 03:07:10.613534 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf757e5-76fb-4aee-80c3-d362a58a977b" containerName="keystone-db-sync" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.613550 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf757e5-76fb-4aee-80c3-d362a58a977b" containerName="keystone-db-sync" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.613745 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="daf757e5-76fb-4aee-80c3-d362a58a977b" containerName="keystone-db-sync" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.615012 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.629467 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f7dfc9995-xlkfd"] Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.633087 4775 scope.go:117] "RemoveContainer" containerID="2a225463593bf5f379339e9642c5e4bd93bfc749740d97e7c0ff663fc62eed76" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.670151 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-r5w8c"] Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.671538 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.674190 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.674467 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.674618 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5tx2s" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.675749 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.676405 4775 scope.go:117] "RemoveContainer" containerID="369cc334c79e0621d07ce7bded2932fb04244a920c2671e9b5db1d96bb9a16e3" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.694101 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-r5w8c"] Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.715428 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-dns-svc\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.715468 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcqjl\" (UniqueName: \"kubernetes.io/projected/05e48010-0ce7-433d-b099-95f5a371ffa1-kube-api-access-qcqjl\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.715655 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-ovsdbserver-sb\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.715697 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-config\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.715732 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-fernet-keys\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.715788 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-ovsdbserver-nb\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.715810 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-credential-keys\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.715890 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-config-data\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.715905 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-combined-ca-bundle\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.715939 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gprcf\" (UniqueName: \"kubernetes.io/projected/fc13c15d-c9db-4039-8526-1151e2c080cc-kube-api-access-gprcf\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.716005 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-scripts\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.721984 4775 scope.go:117] "RemoveContainer" containerID="673ac341e972788fe95dd4b6f7f41fa75ac52ade2a471604ad54e351d6fe2ae0" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.754562 4775 scope.go:117] "RemoveContainer" containerID="f9ee0dd65d23eb09f02ca47efcc3944f82c77c1965c8946ffa8ee44f05c76f2b" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.796482 4775 scope.go:117] "RemoveContainer" containerID="fad2bc6324d014ce2d0b23509b04be74261b40e138fd6a2e46369347cb669b7f" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.818435 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-fernet-keys\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.818489 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-ovsdbserver-nb\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.818509 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-credential-keys\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.818553 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-config-data\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.818568 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-combined-ca-bundle\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.818589 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gprcf\" (UniqueName: \"kubernetes.io/projected/fc13c15d-c9db-4039-8526-1151e2c080cc-kube-api-access-gprcf\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.818627 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-scripts\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.818723 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-dns-svc\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.818742 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcqjl\" (UniqueName: \"kubernetes.io/projected/05e48010-0ce7-433d-b099-95f5a371ffa1-kube-api-access-qcqjl\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.818800 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-ovsdbserver-sb\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.818819 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-config\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.819637 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-config\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.819843 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-ovsdbserver-nb\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.820693 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-ovsdbserver-sb\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.826622 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-combined-ca-bundle\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.827047 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-fernet-keys\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.827218 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-dns-svc\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.828535 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-scripts\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.832509 4775 scope.go:117] "RemoveContainer" containerID="1b5206f9745d26c88b7308321a3826fc8f589314f6f280977a90d726c96ea4f3" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.834810 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-credential-keys\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.837811 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-config-data\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.838725 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcqjl\" (UniqueName: \"kubernetes.io/projected/05e48010-0ce7-433d-b099-95f5a371ffa1-kube-api-access-qcqjl\") pod \"keystone-bootstrap-r5w8c\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.851354 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gprcf\" (UniqueName: \"kubernetes.io/projected/fc13c15d-c9db-4039-8526-1151e2c080cc-kube-api-access-gprcf\") pod \"dnsmasq-dns-f7dfc9995-xlkfd\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:10 crc kubenswrapper[4775]: I1002 03:07:10.936107 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:11 crc kubenswrapper[4775]: I1002 03:07:11.077183 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:11 crc kubenswrapper[4775]: I1002 03:07:11.442979 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f7dfc9995-xlkfd"] Oct 02 03:07:11 crc kubenswrapper[4775]: W1002 03:07:11.443652 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc13c15d_c9db_4039_8526_1151e2c080cc.slice/crio-18c28f488d8f74ab79ee1ea9299a7f4f009888b1031eece2b77ab0256a1a8ef3 WatchSource:0}: Error finding container 18c28f488d8f74ab79ee1ea9299a7f4f009888b1031eece2b77ab0256a1a8ef3: Status 404 returned error can't find the container with id 18c28f488d8f74ab79ee1ea9299a7f4f009888b1031eece2b77ab0256a1a8ef3 Oct 02 03:07:11 crc kubenswrapper[4775]: I1002 03:07:11.528878 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-r5w8c"] Oct 02 03:07:11 crc kubenswrapper[4775]: W1002 03:07:11.540484 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05e48010_0ce7_433d_b099_95f5a371ffa1.slice/crio-e4223a8dca7aada7124abf017c6a7ee5a569fee563cb9d7bd5f4d8c2737a2a45 WatchSource:0}: Error finding container e4223a8dca7aada7124abf017c6a7ee5a569fee563cb9d7bd5f4d8c2737a2a45: Status 404 returned error can't find the container with id e4223a8dca7aada7124abf017c6a7ee5a569fee563cb9d7bd5f4d8c2737a2a45 Oct 02 03:07:12 crc kubenswrapper[4775]: I1002 03:07:12.399084 4775 generic.go:334] "Generic (PLEG): container finished" podID="fc13c15d-c9db-4039-8526-1151e2c080cc" containerID="ff51368a5a8e78612a5228b0debf8f912ec18d1e713c7508b98c85b0670975de" exitCode=0 Oct 02 03:07:12 crc kubenswrapper[4775]: I1002 03:07:12.399210 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" event={"ID":"fc13c15d-c9db-4039-8526-1151e2c080cc","Type":"ContainerDied","Data":"ff51368a5a8e78612a5228b0debf8f912ec18d1e713c7508b98c85b0670975de"} Oct 02 03:07:12 crc kubenswrapper[4775]: I1002 03:07:12.401471 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" event={"ID":"fc13c15d-c9db-4039-8526-1151e2c080cc","Type":"ContainerStarted","Data":"18c28f488d8f74ab79ee1ea9299a7f4f009888b1031eece2b77ab0256a1a8ef3"} Oct 02 03:07:12 crc kubenswrapper[4775]: I1002 03:07:12.404520 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r5w8c" event={"ID":"05e48010-0ce7-433d-b099-95f5a371ffa1","Type":"ContainerStarted","Data":"b152a10dfa6bd2a9f1aa43eaf12766e437aeb3db19855b25781a30df5c8d960e"} Oct 02 03:07:12 crc kubenswrapper[4775]: I1002 03:07:12.404717 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r5w8c" event={"ID":"05e48010-0ce7-433d-b099-95f5a371ffa1","Type":"ContainerStarted","Data":"e4223a8dca7aada7124abf017c6a7ee5a569fee563cb9d7bd5f4d8c2737a2a45"} Oct 02 03:07:12 crc kubenswrapper[4775]: I1002 03:07:12.475679 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-r5w8c" podStartSLOduration=2.475648535 podStartE2EDuration="2.475648535s" podCreationTimestamp="2025-10-02 03:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:07:12.465271697 +0000 UTC m=+5169.632015797" watchObservedRunningTime="2025-10-02 03:07:12.475648535 +0000 UTC m=+5169.642392615" Oct 02 03:07:13 crc kubenswrapper[4775]: I1002 03:07:13.418903 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" event={"ID":"fc13c15d-c9db-4039-8526-1151e2c080cc","Type":"ContainerStarted","Data":"11ba9dcd44e21f64181ec61b63a445c44883856a6d4441608fe04c5c93765def"} Oct 02 03:07:13 crc kubenswrapper[4775]: I1002 03:07:13.457338 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" podStartSLOduration=3.457308449 podStartE2EDuration="3.457308449s" podCreationTimestamp="2025-10-02 03:07:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:07:13.443728589 +0000 UTC m=+5170.610472669" watchObservedRunningTime="2025-10-02 03:07:13.457308449 +0000 UTC m=+5170.624052529" Oct 02 03:07:14 crc kubenswrapper[4775]: I1002 03:07:14.431914 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:15 crc kubenswrapper[4775]: I1002 03:07:15.446675 4775 generic.go:334] "Generic (PLEG): container finished" podID="05e48010-0ce7-433d-b099-95f5a371ffa1" containerID="b152a10dfa6bd2a9f1aa43eaf12766e437aeb3db19855b25781a30df5c8d960e" exitCode=0 Oct 02 03:07:15 crc kubenswrapper[4775]: I1002 03:07:15.446808 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r5w8c" event={"ID":"05e48010-0ce7-433d-b099-95f5a371ffa1","Type":"ContainerDied","Data":"b152a10dfa6bd2a9f1aa43eaf12766e437aeb3db19855b25781a30df5c8d960e"} Oct 02 03:07:16 crc kubenswrapper[4775]: I1002 03:07:16.875485 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:16 crc kubenswrapper[4775]: I1002 03:07:16.967548 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-fernet-keys\") pod \"05e48010-0ce7-433d-b099-95f5a371ffa1\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " Oct 02 03:07:16 crc kubenswrapper[4775]: I1002 03:07:16.967647 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcqjl\" (UniqueName: \"kubernetes.io/projected/05e48010-0ce7-433d-b099-95f5a371ffa1-kube-api-access-qcqjl\") pod \"05e48010-0ce7-433d-b099-95f5a371ffa1\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " Oct 02 03:07:16 crc kubenswrapper[4775]: I1002 03:07:16.967714 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-scripts\") pod \"05e48010-0ce7-433d-b099-95f5a371ffa1\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " Oct 02 03:07:16 crc kubenswrapper[4775]: I1002 03:07:16.967848 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-credential-keys\") pod \"05e48010-0ce7-433d-b099-95f5a371ffa1\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " Oct 02 03:07:16 crc kubenswrapper[4775]: I1002 03:07:16.967883 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-config-data\") pod \"05e48010-0ce7-433d-b099-95f5a371ffa1\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " Oct 02 03:07:16 crc kubenswrapper[4775]: I1002 03:07:16.967986 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-combined-ca-bundle\") pod \"05e48010-0ce7-433d-b099-95f5a371ffa1\" (UID: \"05e48010-0ce7-433d-b099-95f5a371ffa1\") " Oct 02 03:07:16 crc kubenswrapper[4775]: I1002 03:07:16.977579 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-scripts" (OuterVolumeSpecName: "scripts") pod "05e48010-0ce7-433d-b099-95f5a371ffa1" (UID: "05e48010-0ce7-433d-b099-95f5a371ffa1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:07:16 crc kubenswrapper[4775]: I1002 03:07:16.977655 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05e48010-0ce7-433d-b099-95f5a371ffa1-kube-api-access-qcqjl" (OuterVolumeSpecName: "kube-api-access-qcqjl") pod "05e48010-0ce7-433d-b099-95f5a371ffa1" (UID: "05e48010-0ce7-433d-b099-95f5a371ffa1"). InnerVolumeSpecName "kube-api-access-qcqjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:07:16 crc kubenswrapper[4775]: I1002 03:07:16.979381 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "05e48010-0ce7-433d-b099-95f5a371ffa1" (UID: "05e48010-0ce7-433d-b099-95f5a371ffa1"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:07:16 crc kubenswrapper[4775]: I1002 03:07:16.980778 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "05e48010-0ce7-433d-b099-95f5a371ffa1" (UID: "05e48010-0ce7-433d-b099-95f5a371ffa1"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.010577 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-config-data" (OuterVolumeSpecName: "config-data") pod "05e48010-0ce7-433d-b099-95f5a371ffa1" (UID: "05e48010-0ce7-433d-b099-95f5a371ffa1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.012047 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05e48010-0ce7-433d-b099-95f5a371ffa1" (UID: "05e48010-0ce7-433d-b099-95f5a371ffa1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.070225 4775 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.070430 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.070563 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.070678 4775 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.070799 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcqjl\" (UniqueName: \"kubernetes.io/projected/05e48010-0ce7-433d-b099-95f5a371ffa1-kube-api-access-qcqjl\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.070913 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05e48010-0ce7-433d-b099-95f5a371ffa1-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.490512 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-r5w8c" event={"ID":"05e48010-0ce7-433d-b099-95f5a371ffa1","Type":"ContainerDied","Data":"e4223a8dca7aada7124abf017c6a7ee5a569fee563cb9d7bd5f4d8c2737a2a45"} Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.490562 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4223a8dca7aada7124abf017c6a7ee5a569fee563cb9d7bd5f4d8c2737a2a45" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.490565 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-r5w8c" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.593538 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-r5w8c"] Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.601853 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-r5w8c"] Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.662943 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rw6jm"] Oct 02 03:07:17 crc kubenswrapper[4775]: E1002 03:07:17.663822 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e48010-0ce7-433d-b099-95f5a371ffa1" containerName="keystone-bootstrap" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.663892 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e48010-0ce7-433d-b099-95f5a371ffa1" containerName="keystone-bootstrap" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.664495 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="05e48010-0ce7-433d-b099-95f5a371ffa1" containerName="keystone-bootstrap" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.666307 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.668379 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.668425 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.668813 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.670375 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5tx2s" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.674937 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rw6jm"] Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.777450 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05e48010-0ce7-433d-b099-95f5a371ffa1" path="/var/lib/kubelet/pods/05e48010-0ce7-433d-b099-95f5a371ffa1/volumes" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.787469 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-credential-keys\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.787527 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqbjt\" (UniqueName: \"kubernetes.io/projected/c1d8f136-3ac3-485b-af3a-8b0410f2b902-kube-api-access-dqbjt\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.787547 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-config-data\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.787607 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-scripts\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.787655 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-combined-ca-bundle\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.787700 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-fernet-keys\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.889455 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-credential-keys\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.890249 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqbjt\" (UniqueName: \"kubernetes.io/projected/c1d8f136-3ac3-485b-af3a-8b0410f2b902-kube-api-access-dqbjt\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.890328 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-config-data\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.890583 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-scripts\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.890720 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-combined-ca-bundle\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.890839 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-fernet-keys\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.895859 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-combined-ca-bundle\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.896933 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-scripts\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.898331 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-fernet-keys\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.901225 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-config-data\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.902345 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-credential-keys\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:17 crc kubenswrapper[4775]: I1002 03:07:17.910589 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqbjt\" (UniqueName: \"kubernetes.io/projected/c1d8f136-3ac3-485b-af3a-8b0410f2b902-kube-api-access-dqbjt\") pod \"keystone-bootstrap-rw6jm\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:18 crc kubenswrapper[4775]: I1002 03:07:18.010526 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:18 crc kubenswrapper[4775]: I1002 03:07:18.500251 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rw6jm"] Oct 02 03:07:19 crc kubenswrapper[4775]: I1002 03:07:19.522990 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rw6jm" event={"ID":"c1d8f136-3ac3-485b-af3a-8b0410f2b902","Type":"ContainerStarted","Data":"3090386b30abf1216016133f8506b8d20c86c4cce95801155441acf59ebeb663"} Oct 02 03:07:19 crc kubenswrapper[4775]: I1002 03:07:19.523301 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rw6jm" event={"ID":"c1d8f136-3ac3-485b-af3a-8b0410f2b902","Type":"ContainerStarted","Data":"180e1eddc92a1f44cfa710a51bead272337cbfbd9efb47cb43a619261d1a7c3a"} Oct 02 03:07:19 crc kubenswrapper[4775]: I1002 03:07:19.556855 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rw6jm" podStartSLOduration=2.556824627 podStartE2EDuration="2.556824627s" podCreationTimestamp="2025-10-02 03:07:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:07:19.553141242 +0000 UTC m=+5176.719885352" watchObservedRunningTime="2025-10-02 03:07:19.556824627 +0000 UTC m=+5176.723568697" Oct 02 03:07:19 crc kubenswrapper[4775]: I1002 03:07:19.766204 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:07:19 crc kubenswrapper[4775]: E1002 03:07:19.766644 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:07:20 crc kubenswrapper[4775]: I1002 03:07:20.938022 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.026330 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dff454985-787r8"] Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.026682 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dff454985-787r8" podUID="7480c773-8658-4a0d-9150-eb0b3ecc7211" containerName="dnsmasq-dns" containerID="cri-o://41dae23bbe5d004c673724794d9350d7a365967d3597529bf8e479e1d294d8d9" gracePeriod=10 Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.548053 4775 generic.go:334] "Generic (PLEG): container finished" podID="c1d8f136-3ac3-485b-af3a-8b0410f2b902" containerID="3090386b30abf1216016133f8506b8d20c86c4cce95801155441acf59ebeb663" exitCode=0 Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.548178 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rw6jm" event={"ID":"c1d8f136-3ac3-485b-af3a-8b0410f2b902","Type":"ContainerDied","Data":"3090386b30abf1216016133f8506b8d20c86c4cce95801155441acf59ebeb663"} Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.552644 4775 generic.go:334] "Generic (PLEG): container finished" podID="7480c773-8658-4a0d-9150-eb0b3ecc7211" containerID="41dae23bbe5d004c673724794d9350d7a365967d3597529bf8e479e1d294d8d9" exitCode=0 Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.552706 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dff454985-787r8" event={"ID":"7480c773-8658-4a0d-9150-eb0b3ecc7211","Type":"ContainerDied","Data":"41dae23bbe5d004c673724794d9350d7a365967d3597529bf8e479e1d294d8d9"} Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.719356 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.775544 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-config\") pod \"7480c773-8658-4a0d-9150-eb0b3ecc7211\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.775637 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-ovsdbserver-nb\") pod \"7480c773-8658-4a0d-9150-eb0b3ecc7211\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.775690 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-ovsdbserver-sb\") pod \"7480c773-8658-4a0d-9150-eb0b3ecc7211\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.775910 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-dns-svc\") pod \"7480c773-8658-4a0d-9150-eb0b3ecc7211\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.775969 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mw6wg\" (UniqueName: \"kubernetes.io/projected/7480c773-8658-4a0d-9150-eb0b3ecc7211-kube-api-access-mw6wg\") pod \"7480c773-8658-4a0d-9150-eb0b3ecc7211\" (UID: \"7480c773-8658-4a0d-9150-eb0b3ecc7211\") " Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.782673 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7480c773-8658-4a0d-9150-eb0b3ecc7211-kube-api-access-mw6wg" (OuterVolumeSpecName: "kube-api-access-mw6wg") pod "7480c773-8658-4a0d-9150-eb0b3ecc7211" (UID: "7480c773-8658-4a0d-9150-eb0b3ecc7211"). InnerVolumeSpecName "kube-api-access-mw6wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.810435 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7480c773-8658-4a0d-9150-eb0b3ecc7211" (UID: "7480c773-8658-4a0d-9150-eb0b3ecc7211"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.824570 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7480c773-8658-4a0d-9150-eb0b3ecc7211" (UID: "7480c773-8658-4a0d-9150-eb0b3ecc7211"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.827835 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-config" (OuterVolumeSpecName: "config") pod "7480c773-8658-4a0d-9150-eb0b3ecc7211" (UID: "7480c773-8658-4a0d-9150-eb0b3ecc7211"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.831426 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7480c773-8658-4a0d-9150-eb0b3ecc7211" (UID: "7480c773-8658-4a0d-9150-eb0b3ecc7211"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.877510 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.877533 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.877543 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.877552 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7480c773-8658-4a0d-9150-eb0b3ecc7211-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:21 crc kubenswrapper[4775]: I1002 03:07:21.877563 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mw6wg\" (UniqueName: \"kubernetes.io/projected/7480c773-8658-4a0d-9150-eb0b3ecc7211-kube-api-access-mw6wg\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:22 crc kubenswrapper[4775]: I1002 03:07:22.567119 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dff454985-787r8" Oct 02 03:07:22 crc kubenswrapper[4775]: I1002 03:07:22.567230 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dff454985-787r8" event={"ID":"7480c773-8658-4a0d-9150-eb0b3ecc7211","Type":"ContainerDied","Data":"7355e90447b9d841e249c32b96ec0b78ada8bca55ca4bb8856fe41d32321a98d"} Oct 02 03:07:22 crc kubenswrapper[4775]: I1002 03:07:22.567299 4775 scope.go:117] "RemoveContainer" containerID="41dae23bbe5d004c673724794d9350d7a365967d3597529bf8e479e1d294d8d9" Oct 02 03:07:22 crc kubenswrapper[4775]: I1002 03:07:22.630287 4775 scope.go:117] "RemoveContainer" containerID="aa012c6ebc8c7bd2102d53e5acab2b3196cd4feb3263df17135cf748bcbcf98e" Oct 02 03:07:22 crc kubenswrapper[4775]: I1002 03:07:22.637902 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dff454985-787r8"] Oct 02 03:07:22 crc kubenswrapper[4775]: I1002 03:07:22.648738 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dff454985-787r8"] Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.060697 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.098252 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-scripts\") pod \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.098453 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-combined-ca-bundle\") pod \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.098504 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqbjt\" (UniqueName: \"kubernetes.io/projected/c1d8f136-3ac3-485b-af3a-8b0410f2b902-kube-api-access-dqbjt\") pod \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.098551 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-config-data\") pod \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.098663 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-credential-keys\") pod \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.098827 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-fernet-keys\") pod \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\" (UID: \"c1d8f136-3ac3-485b-af3a-8b0410f2b902\") " Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.105903 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-scripts" (OuterVolumeSpecName: "scripts") pod "c1d8f136-3ac3-485b-af3a-8b0410f2b902" (UID: "c1d8f136-3ac3-485b-af3a-8b0410f2b902"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.106650 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c1d8f136-3ac3-485b-af3a-8b0410f2b902" (UID: "c1d8f136-3ac3-485b-af3a-8b0410f2b902"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.107408 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1d8f136-3ac3-485b-af3a-8b0410f2b902-kube-api-access-dqbjt" (OuterVolumeSpecName: "kube-api-access-dqbjt") pod "c1d8f136-3ac3-485b-af3a-8b0410f2b902" (UID: "c1d8f136-3ac3-485b-af3a-8b0410f2b902"). InnerVolumeSpecName "kube-api-access-dqbjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.108899 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c1d8f136-3ac3-485b-af3a-8b0410f2b902" (UID: "c1d8f136-3ac3-485b-af3a-8b0410f2b902"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.139714 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-config-data" (OuterVolumeSpecName: "config-data") pod "c1d8f136-3ac3-485b-af3a-8b0410f2b902" (UID: "c1d8f136-3ac3-485b-af3a-8b0410f2b902"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.150641 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1d8f136-3ac3-485b-af3a-8b0410f2b902" (UID: "c1d8f136-3ac3-485b-af3a-8b0410f2b902"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.201150 4775 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.201193 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.201212 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.201230 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqbjt\" (UniqueName: \"kubernetes.io/projected/c1d8f136-3ac3-485b-af3a-8b0410f2b902-kube-api-access-dqbjt\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.201249 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.201265 4775 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c1d8f136-3ac3-485b-af3a-8b0410f2b902-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.582316 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rw6jm" event={"ID":"c1d8f136-3ac3-485b-af3a-8b0410f2b902","Type":"ContainerDied","Data":"180e1eddc92a1f44cfa710a51bead272337cbfbd9efb47cb43a619261d1a7c3a"} Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.582614 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="180e1eddc92a1f44cfa710a51bead272337cbfbd9efb47cb43a619261d1a7c3a" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.582796 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rw6jm" Oct 02 03:07:23 crc kubenswrapper[4775]: I1002 03:07:23.785190 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7480c773-8658-4a0d-9150-eb0b3ecc7211" path="/var/lib/kubelet/pods/7480c773-8658-4a0d-9150-eb0b3ecc7211/volumes" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.293916 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6959f9548-mcs66"] Oct 02 03:07:24 crc kubenswrapper[4775]: E1002 03:07:24.294341 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1d8f136-3ac3-485b-af3a-8b0410f2b902" containerName="keystone-bootstrap" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.294365 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1d8f136-3ac3-485b-af3a-8b0410f2b902" containerName="keystone-bootstrap" Oct 02 03:07:24 crc kubenswrapper[4775]: E1002 03:07:24.294388 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7480c773-8658-4a0d-9150-eb0b3ecc7211" containerName="init" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.294397 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7480c773-8658-4a0d-9150-eb0b3ecc7211" containerName="init" Oct 02 03:07:24 crc kubenswrapper[4775]: E1002 03:07:24.294413 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7480c773-8658-4a0d-9150-eb0b3ecc7211" containerName="dnsmasq-dns" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.294421 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7480c773-8658-4a0d-9150-eb0b3ecc7211" containerName="dnsmasq-dns" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.294595 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7480c773-8658-4a0d-9150-eb0b3ecc7211" containerName="dnsmasq-dns" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.294619 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1d8f136-3ac3-485b-af3a-8b0410f2b902" containerName="keystone-bootstrap" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.295243 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.297441 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.298251 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.298534 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.298767 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-5tx2s" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.318053 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6959f9548-mcs66"] Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.424527 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-fernet-keys\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.424689 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-credential-keys\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.424792 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-combined-ca-bundle\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.424823 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wdvq\" (UniqueName: \"kubernetes.io/projected/4161a715-fa2c-452b-9259-ce6a0e165533-kube-api-access-7wdvq\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.424887 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-config-data\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.424919 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-scripts\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.526250 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-credential-keys\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.526329 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-combined-ca-bundle\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.526346 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wdvq\" (UniqueName: \"kubernetes.io/projected/4161a715-fa2c-452b-9259-ce6a0e165533-kube-api-access-7wdvq\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.526378 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-config-data\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.526398 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-scripts\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.526422 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-fernet-keys\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.531065 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-fernet-keys\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.532648 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-config-data\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.533115 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-scripts\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.533459 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-combined-ca-bundle\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.533633 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4161a715-fa2c-452b-9259-ce6a0e165533-credential-keys\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.547436 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wdvq\" (UniqueName: \"kubernetes.io/projected/4161a715-fa2c-452b-9259-ce6a0e165533-kube-api-access-7wdvq\") pod \"keystone-6959f9548-mcs66\" (UID: \"4161a715-fa2c-452b-9259-ce6a0e165533\") " pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:24 crc kubenswrapper[4775]: I1002 03:07:24.622706 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:25 crc kubenswrapper[4775]: I1002 03:07:25.119251 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6959f9548-mcs66"] Oct 02 03:07:25 crc kubenswrapper[4775]: W1002 03:07:25.124680 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4161a715_fa2c_452b_9259_ce6a0e165533.slice/crio-35d688e024ab28711309d84e8304c837721de2d36d5cfad8619d0ecd12eb6836 WatchSource:0}: Error finding container 35d688e024ab28711309d84e8304c837721de2d36d5cfad8619d0ecd12eb6836: Status 404 returned error can't find the container with id 35d688e024ab28711309d84e8304c837721de2d36d5cfad8619d0ecd12eb6836 Oct 02 03:07:25 crc kubenswrapper[4775]: I1002 03:07:25.605362 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6959f9548-mcs66" event={"ID":"4161a715-fa2c-452b-9259-ce6a0e165533","Type":"ContainerStarted","Data":"145365e62ae5f740ad1c8285f8cbc40183afd7d801af5815352d174ede06c215"} Oct 02 03:07:25 crc kubenswrapper[4775]: I1002 03:07:25.605426 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6959f9548-mcs66" event={"ID":"4161a715-fa2c-452b-9259-ce6a0e165533","Type":"ContainerStarted","Data":"35d688e024ab28711309d84e8304c837721de2d36d5cfad8619d0ecd12eb6836"} Oct 02 03:07:25 crc kubenswrapper[4775]: I1002 03:07:25.605642 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:25 crc kubenswrapper[4775]: I1002 03:07:25.627932 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6959f9548-mcs66" podStartSLOduration=1.6278533309999998 podStartE2EDuration="1.627853331s" podCreationTimestamp="2025-10-02 03:07:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:07:25.622971625 +0000 UTC m=+5182.789715685" watchObservedRunningTime="2025-10-02 03:07:25.627853331 +0000 UTC m=+5182.794597441" Oct 02 03:07:26 crc kubenswrapper[4775]: I1002 03:07:26.547986 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-dff454985-787r8" podUID="7480c773-8658-4a0d-9150-eb0b3ecc7211" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.1.11:5353: i/o timeout" Oct 02 03:07:31 crc kubenswrapper[4775]: I1002 03:07:31.766078 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:07:31 crc kubenswrapper[4775]: E1002 03:07:31.767347 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:07:44 crc kubenswrapper[4775]: I1002 03:07:44.765876 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:07:44 crc kubenswrapper[4775]: E1002 03:07:44.766779 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:07:56 crc kubenswrapper[4775]: I1002 03:07:56.043730 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6959f9548-mcs66" Oct 02 03:07:57 crc kubenswrapper[4775]: I1002 03:07:57.766470 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:07:57 crc kubenswrapper[4775]: E1002 03:07:57.767452 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.290099 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.292181 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.294806 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.295465 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-kglx5" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.309397 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.329400 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 02 03:08:00 crc kubenswrapper[4775]: E1002 03:08:00.331479 4775 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"openstack-config-secret\": Failed to watch *v1.Secret: unknown (get secrets)" logger="UnhandledError" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.347143 4775 status_manager.go:875] "Failed to update status for pod" pod="openstack/openstackclient" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"afdabe08-aeb1-4f02-b6a8-d669651e96b4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T03:08:00Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T03:08:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T03:08:00Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-02T03:08:00Z\\\",\\\"message\\\":\\\"containers with unready status: [openstackclient]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"openstackclient\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/clouds.yaml\\\",\\\"name\\\":\\\"openstack-config\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/.config/openstack/secure.yaml\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/home/cloud-admin/cloudrc\\\",\\\"name\\\":\\\"openstack-config-secret\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-79h5r\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-02T03:08:00Z\\\"}}\" for pod \"openstack\"/\"openstackclient\": pods \"openstackclient\" is forbidden: node \"crc\" can only update pod status for pods with spec.nodeName set to itself" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.350944 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 02 03:08:00 crc kubenswrapper[4775]: E1002 03:08:00.352054 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-79h5r openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="afdabe08-aeb1-4f02-b6a8-d669651e96b4" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.388146 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.396458 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.397627 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.400597 4775 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="afdabe08-aeb1-4f02-b6a8-d669651e96b4" podUID="bcc7bd03-693b-4733-9f7f-dbd7fd15e490" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.405417 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.413820 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/afdabe08-aeb1-4f02-b6a8-d669651e96b4-openstack-config\") pod \"openstackclient\" (UID: \"afdabe08-aeb1-4f02-b6a8-d669651e96b4\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.413901 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79h5r\" (UniqueName: \"kubernetes.io/projected/afdabe08-aeb1-4f02-b6a8-d669651e96b4-kube-api-access-79h5r\") pod \"openstackclient\" (UID: \"afdabe08-aeb1-4f02-b6a8-d669651e96b4\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.413975 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/afdabe08-aeb1-4f02-b6a8-d669651e96b4-openstack-config-secret\") pod \"openstackclient\" (UID: \"afdabe08-aeb1-4f02-b6a8-d669651e96b4\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.515919 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79h5r\" (UniqueName: \"kubernetes.io/projected/afdabe08-aeb1-4f02-b6a8-d669651e96b4-kube-api-access-79h5r\") pod \"openstackclient\" (UID: \"afdabe08-aeb1-4f02-b6a8-d669651e96b4\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.516291 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp6qw\" (UniqueName: \"kubernetes.io/projected/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-kube-api-access-zp6qw\") pod \"openstackclient\" (UID: \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.516428 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-openstack-config-secret\") pod \"openstackclient\" (UID: \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.518126 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/afdabe08-aeb1-4f02-b6a8-d669651e96b4-openstack-config-secret\") pod \"openstackclient\" (UID: \"afdabe08-aeb1-4f02-b6a8-d669651e96b4\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.518231 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-openstack-config\") pod \"openstackclient\" (UID: \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: E1002 03:08:00.518255 4775 projected.go:194] Error preparing data for projected volume kube-api-access-79h5r for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (afdabe08-aeb1-4f02-b6a8-d669651e96b4) does not match the UID in record. The object might have been deleted and then recreated Oct 02 03:08:00 crc kubenswrapper[4775]: E1002 03:08:00.518332 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/afdabe08-aeb1-4f02-b6a8-d669651e96b4-kube-api-access-79h5r podName:afdabe08-aeb1-4f02-b6a8-d669651e96b4 nodeName:}" failed. No retries permitted until 2025-10-02 03:08:01.0183071 +0000 UTC m=+5218.185051160 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-79h5r" (UniqueName: "kubernetes.io/projected/afdabe08-aeb1-4f02-b6a8-d669651e96b4-kube-api-access-79h5r") pod "openstackclient" (UID: "afdabe08-aeb1-4f02-b6a8-d669651e96b4") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (afdabe08-aeb1-4f02-b6a8-d669651e96b4) does not match the UID in record. The object might have been deleted and then recreated Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.518445 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/afdabe08-aeb1-4f02-b6a8-d669651e96b4-openstack-config\") pod \"openstackclient\" (UID: \"afdabe08-aeb1-4f02-b6a8-d669651e96b4\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.519308 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/afdabe08-aeb1-4f02-b6a8-d669651e96b4-openstack-config\") pod \"openstackclient\" (UID: \"afdabe08-aeb1-4f02-b6a8-d669651e96b4\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.527924 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/afdabe08-aeb1-4f02-b6a8-d669651e96b4-openstack-config-secret\") pod \"openstackclient\" (UID: \"afdabe08-aeb1-4f02-b6a8-d669651e96b4\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.619879 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp6qw\" (UniqueName: \"kubernetes.io/projected/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-kube-api-access-zp6qw\") pod \"openstackclient\" (UID: \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.619931 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-openstack-config-secret\") pod \"openstackclient\" (UID: \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.619975 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-openstack-config\") pod \"openstackclient\" (UID: \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.620729 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-openstack-config\") pod \"openstackclient\" (UID: \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.627419 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-openstack-config-secret\") pod \"openstackclient\" (UID: \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.640597 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp6qw\" (UniqueName: \"kubernetes.io/projected/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-kube-api-access-zp6qw\") pod \"openstackclient\" (UID: \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\") " pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.714225 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.966985 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.970086 4775 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="afdabe08-aeb1-4f02-b6a8-d669651e96b4" podUID="bcc7bd03-693b-4733-9f7f-dbd7fd15e490" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.978820 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 03:08:00 crc kubenswrapper[4775]: I1002 03:08:00.981546 4775 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="afdabe08-aeb1-4f02-b6a8-d669651e96b4" podUID="bcc7bd03-693b-4733-9f7f-dbd7fd15e490" Oct 02 03:08:01 crc kubenswrapper[4775]: I1002 03:08:01.027039 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79h5r\" (UniqueName: \"kubernetes.io/projected/afdabe08-aeb1-4f02-b6a8-d669651e96b4-kube-api-access-79h5r\") pod \"openstackclient\" (UID: \"afdabe08-aeb1-4f02-b6a8-d669651e96b4\") " pod="openstack/openstackclient" Oct 02 03:08:01 crc kubenswrapper[4775]: E1002 03:08:01.028900 4775 projected.go:194] Error preparing data for projected volume kube-api-access-79h5r for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (afdabe08-aeb1-4f02-b6a8-d669651e96b4) does not match the UID in record. The object might have been deleted and then recreated Oct 02 03:08:01 crc kubenswrapper[4775]: E1002 03:08:01.029080 4775 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/afdabe08-aeb1-4f02-b6a8-d669651e96b4-kube-api-access-79h5r podName:afdabe08-aeb1-4f02-b6a8-d669651e96b4 nodeName:}" failed. No retries permitted until 2025-10-02 03:08:02.029058466 +0000 UTC m=+5219.195802516 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-79h5r" (UniqueName: "kubernetes.io/projected/afdabe08-aeb1-4f02-b6a8-d669651e96b4-kube-api-access-79h5r") pod "openstackclient" (UID: "afdabe08-aeb1-4f02-b6a8-d669651e96b4") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (afdabe08-aeb1-4f02-b6a8-d669651e96b4) does not match the UID in record. The object might have been deleted and then recreated Oct 02 03:08:01 crc kubenswrapper[4775]: I1002 03:08:01.128540 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/afdabe08-aeb1-4f02-b6a8-d669651e96b4-openstack-config\") pod \"afdabe08-aeb1-4f02-b6a8-d669651e96b4\" (UID: \"afdabe08-aeb1-4f02-b6a8-d669651e96b4\") " Oct 02 03:08:01 crc kubenswrapper[4775]: I1002 03:08:01.128664 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/afdabe08-aeb1-4f02-b6a8-d669651e96b4-openstack-config-secret\") pod \"afdabe08-aeb1-4f02-b6a8-d669651e96b4\" (UID: \"afdabe08-aeb1-4f02-b6a8-d669651e96b4\") " Oct 02 03:08:01 crc kubenswrapper[4775]: I1002 03:08:01.129142 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79h5r\" (UniqueName: \"kubernetes.io/projected/afdabe08-aeb1-4f02-b6a8-d669651e96b4-kube-api-access-79h5r\") on node \"crc\" DevicePath \"\"" Oct 02 03:08:01 crc kubenswrapper[4775]: I1002 03:08:01.129368 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/afdabe08-aeb1-4f02-b6a8-d669651e96b4-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "afdabe08-aeb1-4f02-b6a8-d669651e96b4" (UID: "afdabe08-aeb1-4f02-b6a8-d669651e96b4"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:08:01 crc kubenswrapper[4775]: I1002 03:08:01.134419 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdabe08-aeb1-4f02-b6a8-d669651e96b4-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "afdabe08-aeb1-4f02-b6a8-d669651e96b4" (UID: "afdabe08-aeb1-4f02-b6a8-d669651e96b4"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:08:01 crc kubenswrapper[4775]: I1002 03:08:01.202714 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 03:08:01 crc kubenswrapper[4775]: W1002 03:08:01.205366 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcc7bd03_693b_4733_9f7f_dbd7fd15e490.slice/crio-34c513c7dfb16ef511e3749a97abc90b9f9149d39bc75e3e56a952bb94c209fc WatchSource:0}: Error finding container 34c513c7dfb16ef511e3749a97abc90b9f9149d39bc75e3e56a952bb94c209fc: Status 404 returned error can't find the container with id 34c513c7dfb16ef511e3749a97abc90b9f9149d39bc75e3e56a952bb94c209fc Oct 02 03:08:01 crc kubenswrapper[4775]: I1002 03:08:01.230746 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/afdabe08-aeb1-4f02-b6a8-d669651e96b4-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 03:08:01 crc kubenswrapper[4775]: I1002 03:08:01.230785 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/afdabe08-aeb1-4f02-b6a8-d669651e96b4-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:08:01 crc kubenswrapper[4775]: I1002 03:08:01.582446 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 02 03:08:01 crc kubenswrapper[4775]: I1002 03:08:01.777291 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afdabe08-aeb1-4f02-b6a8-d669651e96b4" path="/var/lib/kubelet/pods/afdabe08-aeb1-4f02-b6a8-d669651e96b4/volumes" Oct 02 03:08:01 crc kubenswrapper[4775]: I1002 03:08:01.981927 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 03:08:01 crc kubenswrapper[4775]: I1002 03:08:01.981912 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bcc7bd03-693b-4733-9f7f-dbd7fd15e490","Type":"ContainerStarted","Data":"d89c6c8b2bc6e48a2f1b06ce8dc58577b83fcdb8ccb8e824108f02adb5655b1d"} Oct 02 03:08:01 crc kubenswrapper[4775]: I1002 03:08:01.982526 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bcc7bd03-693b-4733-9f7f-dbd7fd15e490","Type":"ContainerStarted","Data":"34c513c7dfb16ef511e3749a97abc90b9f9149d39bc75e3e56a952bb94c209fc"} Oct 02 03:08:02 crc kubenswrapper[4775]: I1002 03:08:02.005790 4775 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="afdabe08-aeb1-4f02-b6a8-d669651e96b4" podUID="bcc7bd03-693b-4733-9f7f-dbd7fd15e490" Oct 02 03:08:02 crc kubenswrapper[4775]: I1002 03:08:02.008178 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.008145475 podStartE2EDuration="2.008145475s" podCreationTimestamp="2025-10-02 03:08:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:08:02.002271643 +0000 UTC m=+5219.169015723" watchObservedRunningTime="2025-10-02 03:08:02.008145475 +0000 UTC m=+5219.174889555" Oct 02 03:08:08 crc kubenswrapper[4775]: I1002 03:08:08.765879 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:08:08 crc kubenswrapper[4775]: E1002 03:08:08.766907 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:08:21 crc kubenswrapper[4775]: I1002 03:08:21.765895 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:08:21 crc kubenswrapper[4775]: E1002 03:08:21.766922 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:08:33 crc kubenswrapper[4775]: I1002 03:08:33.774931 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:08:33 crc kubenswrapper[4775]: E1002 03:08:33.777628 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:08:45 crc kubenswrapper[4775]: I1002 03:08:45.766278 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:08:45 crc kubenswrapper[4775]: E1002 03:08:45.767687 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:08:57 crc kubenswrapper[4775]: I1002 03:08:57.765315 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:08:57 crc kubenswrapper[4775]: E1002 03:08:57.766249 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:09:07 crc kubenswrapper[4775]: I1002 03:09:07.568411 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wbbzt"] Oct 02 03:09:07 crc kubenswrapper[4775]: I1002 03:09:07.571866 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:07 crc kubenswrapper[4775]: I1002 03:09:07.581634 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wbbzt"] Oct 02 03:09:07 crc kubenswrapper[4775]: I1002 03:09:07.684379 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af80badb-06f8-4025-a49b-e036affcdca4-catalog-content\") pod \"redhat-marketplace-wbbzt\" (UID: \"af80badb-06f8-4025-a49b-e036affcdca4\") " pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:07 crc kubenswrapper[4775]: I1002 03:09:07.684511 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5w4s\" (UniqueName: \"kubernetes.io/projected/af80badb-06f8-4025-a49b-e036affcdca4-kube-api-access-c5w4s\") pod \"redhat-marketplace-wbbzt\" (UID: \"af80badb-06f8-4025-a49b-e036affcdca4\") " pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:07 crc kubenswrapper[4775]: I1002 03:09:07.684572 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af80badb-06f8-4025-a49b-e036affcdca4-utilities\") pod \"redhat-marketplace-wbbzt\" (UID: \"af80badb-06f8-4025-a49b-e036affcdca4\") " pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:07 crc kubenswrapper[4775]: I1002 03:09:07.785822 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af80badb-06f8-4025-a49b-e036affcdca4-catalog-content\") pod \"redhat-marketplace-wbbzt\" (UID: \"af80badb-06f8-4025-a49b-e036affcdca4\") " pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:07 crc kubenswrapper[4775]: I1002 03:09:07.785932 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5w4s\" (UniqueName: \"kubernetes.io/projected/af80badb-06f8-4025-a49b-e036affcdca4-kube-api-access-c5w4s\") pod \"redhat-marketplace-wbbzt\" (UID: \"af80badb-06f8-4025-a49b-e036affcdca4\") " pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:07 crc kubenswrapper[4775]: I1002 03:09:07.786015 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af80badb-06f8-4025-a49b-e036affcdca4-utilities\") pod \"redhat-marketplace-wbbzt\" (UID: \"af80badb-06f8-4025-a49b-e036affcdca4\") " pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:07 crc kubenswrapper[4775]: I1002 03:09:07.786556 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af80badb-06f8-4025-a49b-e036affcdca4-utilities\") pod \"redhat-marketplace-wbbzt\" (UID: \"af80badb-06f8-4025-a49b-e036affcdca4\") " pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:07 crc kubenswrapper[4775]: I1002 03:09:07.786993 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af80badb-06f8-4025-a49b-e036affcdca4-catalog-content\") pod \"redhat-marketplace-wbbzt\" (UID: \"af80badb-06f8-4025-a49b-e036affcdca4\") " pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:07 crc kubenswrapper[4775]: I1002 03:09:07.817388 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5w4s\" (UniqueName: \"kubernetes.io/projected/af80badb-06f8-4025-a49b-e036affcdca4-kube-api-access-c5w4s\") pod \"redhat-marketplace-wbbzt\" (UID: \"af80badb-06f8-4025-a49b-e036affcdca4\") " pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:07 crc kubenswrapper[4775]: I1002 03:09:07.902187 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:08 crc kubenswrapper[4775]: I1002 03:09:08.395049 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wbbzt"] Oct 02 03:09:08 crc kubenswrapper[4775]: I1002 03:09:08.730349 4775 generic.go:334] "Generic (PLEG): container finished" podID="af80badb-06f8-4025-a49b-e036affcdca4" containerID="b9b96364f7a4e88bbedabd797d7c6916c8bfd7d54258465bcc8a72601178b439" exitCode=0 Oct 02 03:09:08 crc kubenswrapper[4775]: I1002 03:09:08.730411 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wbbzt" event={"ID":"af80badb-06f8-4025-a49b-e036affcdca4","Type":"ContainerDied","Data":"b9b96364f7a4e88bbedabd797d7c6916c8bfd7d54258465bcc8a72601178b439"} Oct 02 03:09:08 crc kubenswrapper[4775]: I1002 03:09:08.730447 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wbbzt" event={"ID":"af80badb-06f8-4025-a49b-e036affcdca4","Type":"ContainerStarted","Data":"0445391d841461c09cebb90bc996006fe3037d703fd1a409486d796fe196d6bf"} Oct 02 03:09:08 crc kubenswrapper[4775]: I1002 03:09:08.735002 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 03:09:08 crc kubenswrapper[4775]: I1002 03:09:08.766325 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:09:09 crc kubenswrapper[4775]: I1002 03:09:09.742374 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"5c0c35a295e1473ae177f82875e76ff0ae056f4db6af4a094e5ee23221bb6584"} Oct 02 03:09:10 crc kubenswrapper[4775]: I1002 03:09:10.752564 4775 generic.go:334] "Generic (PLEG): container finished" podID="af80badb-06f8-4025-a49b-e036affcdca4" containerID="e4e3b58d355ddb9f8e98f8d7537f8bd398b2c277b77d6b018617a1d5238ad8e3" exitCode=0 Oct 02 03:09:10 crc kubenswrapper[4775]: I1002 03:09:10.752776 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wbbzt" event={"ID":"af80badb-06f8-4025-a49b-e036affcdca4","Type":"ContainerDied","Data":"e4e3b58d355ddb9f8e98f8d7537f8bd398b2c277b77d6b018617a1d5238ad8e3"} Oct 02 03:09:11 crc kubenswrapper[4775]: I1002 03:09:11.767266 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wbbzt" event={"ID":"af80badb-06f8-4025-a49b-e036affcdca4","Type":"ContainerStarted","Data":"7981d3f0e7a4689b5772e1dfcd9891daf59689229aaa83b5c0a03cd70803b9f4"} Oct 02 03:09:11 crc kubenswrapper[4775]: I1002 03:09:11.792233 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wbbzt" podStartSLOduration=2.327836429 podStartE2EDuration="4.792208624s" podCreationTimestamp="2025-10-02 03:09:07 +0000 UTC" firstStartedPulling="2025-10-02 03:09:08.734458703 +0000 UTC m=+5285.901202783" lastFinishedPulling="2025-10-02 03:09:11.198830908 +0000 UTC m=+5288.365574978" observedRunningTime="2025-10-02 03:09:11.788465528 +0000 UTC m=+5288.955209608" watchObservedRunningTime="2025-10-02 03:09:11.792208624 +0000 UTC m=+5288.958952704" Oct 02 03:09:17 crc kubenswrapper[4775]: I1002 03:09:17.903187 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:17 crc kubenswrapper[4775]: I1002 03:09:17.903606 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:17 crc kubenswrapper[4775]: I1002 03:09:17.969695 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:18 crc kubenswrapper[4775]: I1002 03:09:18.912983 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:18 crc kubenswrapper[4775]: I1002 03:09:18.964456 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wbbzt"] Oct 02 03:09:20 crc kubenswrapper[4775]: I1002 03:09:20.871632 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wbbzt" podUID="af80badb-06f8-4025-a49b-e036affcdca4" containerName="registry-server" containerID="cri-o://7981d3f0e7a4689b5772e1dfcd9891daf59689229aaa83b5c0a03cd70803b9f4" gracePeriod=2 Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.375177 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.457641 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5w4s\" (UniqueName: \"kubernetes.io/projected/af80badb-06f8-4025-a49b-e036affcdca4-kube-api-access-c5w4s\") pod \"af80badb-06f8-4025-a49b-e036affcdca4\" (UID: \"af80badb-06f8-4025-a49b-e036affcdca4\") " Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.457790 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af80badb-06f8-4025-a49b-e036affcdca4-catalog-content\") pod \"af80badb-06f8-4025-a49b-e036affcdca4\" (UID: \"af80badb-06f8-4025-a49b-e036affcdca4\") " Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.457863 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af80badb-06f8-4025-a49b-e036affcdca4-utilities\") pod \"af80badb-06f8-4025-a49b-e036affcdca4\" (UID: \"af80badb-06f8-4025-a49b-e036affcdca4\") " Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.459212 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af80badb-06f8-4025-a49b-e036affcdca4-utilities" (OuterVolumeSpecName: "utilities") pod "af80badb-06f8-4025-a49b-e036affcdca4" (UID: "af80badb-06f8-4025-a49b-e036affcdca4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.467513 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af80badb-06f8-4025-a49b-e036affcdca4-kube-api-access-c5w4s" (OuterVolumeSpecName: "kube-api-access-c5w4s") pod "af80badb-06f8-4025-a49b-e036affcdca4" (UID: "af80badb-06f8-4025-a49b-e036affcdca4"). InnerVolumeSpecName "kube-api-access-c5w4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.476684 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af80badb-06f8-4025-a49b-e036affcdca4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "af80badb-06f8-4025-a49b-e036affcdca4" (UID: "af80badb-06f8-4025-a49b-e036affcdca4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.563480 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/af80badb-06f8-4025-a49b-e036affcdca4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.563823 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/af80badb-06f8-4025-a49b-e036affcdca4-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.563926 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5w4s\" (UniqueName: \"kubernetes.io/projected/af80badb-06f8-4025-a49b-e036affcdca4-kube-api-access-c5w4s\") on node \"crc\" DevicePath \"\"" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.884496 4775 generic.go:334] "Generic (PLEG): container finished" podID="af80badb-06f8-4025-a49b-e036affcdca4" containerID="7981d3f0e7a4689b5772e1dfcd9891daf59689229aaa83b5c0a03cd70803b9f4" exitCode=0 Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.884563 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wbbzt" event={"ID":"af80badb-06f8-4025-a49b-e036affcdca4","Type":"ContainerDied","Data":"7981d3f0e7a4689b5772e1dfcd9891daf59689229aaa83b5c0a03cd70803b9f4"} Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.884639 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wbbzt" event={"ID":"af80badb-06f8-4025-a49b-e036affcdca4","Type":"ContainerDied","Data":"0445391d841461c09cebb90bc996006fe3037d703fd1a409486d796fe196d6bf"} Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.884630 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wbbzt" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.884671 4775 scope.go:117] "RemoveContainer" containerID="7981d3f0e7a4689b5772e1dfcd9891daf59689229aaa83b5c0a03cd70803b9f4" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.915417 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wbbzt"] Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.924475 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wbbzt"] Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.926972 4775 scope.go:117] "RemoveContainer" containerID="e4e3b58d355ddb9f8e98f8d7537f8bd398b2c277b77d6b018617a1d5238ad8e3" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.961988 4775 scope.go:117] "RemoveContainer" containerID="b9b96364f7a4e88bbedabd797d7c6916c8bfd7d54258465bcc8a72601178b439" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.995334 4775 scope.go:117] "RemoveContainer" containerID="7981d3f0e7a4689b5772e1dfcd9891daf59689229aaa83b5c0a03cd70803b9f4" Oct 02 03:09:21 crc kubenswrapper[4775]: E1002 03:09:21.995864 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7981d3f0e7a4689b5772e1dfcd9891daf59689229aaa83b5c0a03cd70803b9f4\": container with ID starting with 7981d3f0e7a4689b5772e1dfcd9891daf59689229aaa83b5c0a03cd70803b9f4 not found: ID does not exist" containerID="7981d3f0e7a4689b5772e1dfcd9891daf59689229aaa83b5c0a03cd70803b9f4" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.995898 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7981d3f0e7a4689b5772e1dfcd9891daf59689229aaa83b5c0a03cd70803b9f4"} err="failed to get container status \"7981d3f0e7a4689b5772e1dfcd9891daf59689229aaa83b5c0a03cd70803b9f4\": rpc error: code = NotFound desc = could not find container \"7981d3f0e7a4689b5772e1dfcd9891daf59689229aaa83b5c0a03cd70803b9f4\": container with ID starting with 7981d3f0e7a4689b5772e1dfcd9891daf59689229aaa83b5c0a03cd70803b9f4 not found: ID does not exist" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.995922 4775 scope.go:117] "RemoveContainer" containerID="e4e3b58d355ddb9f8e98f8d7537f8bd398b2c277b77d6b018617a1d5238ad8e3" Oct 02 03:09:21 crc kubenswrapper[4775]: E1002 03:09:21.996411 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4e3b58d355ddb9f8e98f8d7537f8bd398b2c277b77d6b018617a1d5238ad8e3\": container with ID starting with e4e3b58d355ddb9f8e98f8d7537f8bd398b2c277b77d6b018617a1d5238ad8e3 not found: ID does not exist" containerID="e4e3b58d355ddb9f8e98f8d7537f8bd398b2c277b77d6b018617a1d5238ad8e3" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.996449 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4e3b58d355ddb9f8e98f8d7537f8bd398b2c277b77d6b018617a1d5238ad8e3"} err="failed to get container status \"e4e3b58d355ddb9f8e98f8d7537f8bd398b2c277b77d6b018617a1d5238ad8e3\": rpc error: code = NotFound desc = could not find container \"e4e3b58d355ddb9f8e98f8d7537f8bd398b2c277b77d6b018617a1d5238ad8e3\": container with ID starting with e4e3b58d355ddb9f8e98f8d7537f8bd398b2c277b77d6b018617a1d5238ad8e3 not found: ID does not exist" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.996477 4775 scope.go:117] "RemoveContainer" containerID="b9b96364f7a4e88bbedabd797d7c6916c8bfd7d54258465bcc8a72601178b439" Oct 02 03:09:21 crc kubenswrapper[4775]: E1002 03:09:21.996937 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9b96364f7a4e88bbedabd797d7c6916c8bfd7d54258465bcc8a72601178b439\": container with ID starting with b9b96364f7a4e88bbedabd797d7c6916c8bfd7d54258465bcc8a72601178b439 not found: ID does not exist" containerID="b9b96364f7a4e88bbedabd797d7c6916c8bfd7d54258465bcc8a72601178b439" Oct 02 03:09:21 crc kubenswrapper[4775]: I1002 03:09:21.997058 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9b96364f7a4e88bbedabd797d7c6916c8bfd7d54258465bcc8a72601178b439"} err="failed to get container status \"b9b96364f7a4e88bbedabd797d7c6916c8bfd7d54258465bcc8a72601178b439\": rpc error: code = NotFound desc = could not find container \"b9b96364f7a4e88bbedabd797d7c6916c8bfd7d54258465bcc8a72601178b439\": container with ID starting with b9b96364f7a4e88bbedabd797d7c6916c8bfd7d54258465bcc8a72601178b439 not found: ID does not exist" Oct 02 03:09:23 crc kubenswrapper[4775]: I1002 03:09:23.786850 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af80badb-06f8-4025-a49b-e036affcdca4" path="/var/lib/kubelet/pods/af80badb-06f8-4025-a49b-e036affcdca4/volumes" Oct 02 03:09:46 crc kubenswrapper[4775]: I1002 03:09:46.122551 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-zwvbk"] Oct 02 03:09:46 crc kubenswrapper[4775]: E1002 03:09:46.123804 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af80badb-06f8-4025-a49b-e036affcdca4" containerName="extract-utilities" Oct 02 03:09:46 crc kubenswrapper[4775]: I1002 03:09:46.123820 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="af80badb-06f8-4025-a49b-e036affcdca4" containerName="extract-utilities" Oct 02 03:09:46 crc kubenswrapper[4775]: E1002 03:09:46.123836 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af80badb-06f8-4025-a49b-e036affcdca4" containerName="registry-server" Oct 02 03:09:46 crc kubenswrapper[4775]: I1002 03:09:46.123842 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="af80badb-06f8-4025-a49b-e036affcdca4" containerName="registry-server" Oct 02 03:09:46 crc kubenswrapper[4775]: E1002 03:09:46.123862 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af80badb-06f8-4025-a49b-e036affcdca4" containerName="extract-content" Oct 02 03:09:46 crc kubenswrapper[4775]: I1002 03:09:46.123867 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="af80badb-06f8-4025-a49b-e036affcdca4" containerName="extract-content" Oct 02 03:09:46 crc kubenswrapper[4775]: I1002 03:09:46.124024 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="af80badb-06f8-4025-a49b-e036affcdca4" containerName="registry-server" Oct 02 03:09:46 crc kubenswrapper[4775]: I1002 03:09:46.124535 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zwvbk" Oct 02 03:09:46 crc kubenswrapper[4775]: I1002 03:09:46.133718 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-zwvbk"] Oct 02 03:09:46 crc kubenswrapper[4775]: I1002 03:09:46.159263 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwvgd\" (UniqueName: \"kubernetes.io/projected/02d4d565-63c4-4522-a2c5-daa9dc836057-kube-api-access-qwvgd\") pod \"barbican-db-create-zwvbk\" (UID: \"02d4d565-63c4-4522-a2c5-daa9dc836057\") " pod="openstack/barbican-db-create-zwvbk" Oct 02 03:09:46 crc kubenswrapper[4775]: I1002 03:09:46.260567 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwvgd\" (UniqueName: \"kubernetes.io/projected/02d4d565-63c4-4522-a2c5-daa9dc836057-kube-api-access-qwvgd\") pod \"barbican-db-create-zwvbk\" (UID: \"02d4d565-63c4-4522-a2c5-daa9dc836057\") " pod="openstack/barbican-db-create-zwvbk" Oct 02 03:09:46 crc kubenswrapper[4775]: I1002 03:09:46.284848 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwvgd\" (UniqueName: \"kubernetes.io/projected/02d4d565-63c4-4522-a2c5-daa9dc836057-kube-api-access-qwvgd\") pod \"barbican-db-create-zwvbk\" (UID: \"02d4d565-63c4-4522-a2c5-daa9dc836057\") " pod="openstack/barbican-db-create-zwvbk" Oct 02 03:09:46 crc kubenswrapper[4775]: I1002 03:09:46.448044 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zwvbk" Oct 02 03:09:46 crc kubenswrapper[4775]: I1002 03:09:46.744239 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-zwvbk"] Oct 02 03:09:47 crc kubenswrapper[4775]: I1002 03:09:47.166812 4775 generic.go:334] "Generic (PLEG): container finished" podID="02d4d565-63c4-4522-a2c5-daa9dc836057" containerID="45ebcf3bd424f58ba6a18ad63bead97953a421eb5bbee73f38099235edb4495b" exitCode=0 Oct 02 03:09:47 crc kubenswrapper[4775]: I1002 03:09:47.166878 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-zwvbk" event={"ID":"02d4d565-63c4-4522-a2c5-daa9dc836057","Type":"ContainerDied","Data":"45ebcf3bd424f58ba6a18ad63bead97953a421eb5bbee73f38099235edb4495b"} Oct 02 03:09:47 crc kubenswrapper[4775]: I1002 03:09:47.166920 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-zwvbk" event={"ID":"02d4d565-63c4-4522-a2c5-daa9dc836057","Type":"ContainerStarted","Data":"5339b57c9603a77a070f9addce00a0ea0c6129b7117ad17cffd0bc5f9737526e"} Oct 02 03:09:48 crc kubenswrapper[4775]: I1002 03:09:48.497488 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zwvbk" Oct 02 03:09:48 crc kubenswrapper[4775]: I1002 03:09:48.605045 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwvgd\" (UniqueName: \"kubernetes.io/projected/02d4d565-63c4-4522-a2c5-daa9dc836057-kube-api-access-qwvgd\") pod \"02d4d565-63c4-4522-a2c5-daa9dc836057\" (UID: \"02d4d565-63c4-4522-a2c5-daa9dc836057\") " Oct 02 03:09:48 crc kubenswrapper[4775]: I1002 03:09:48.611598 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02d4d565-63c4-4522-a2c5-daa9dc836057-kube-api-access-qwvgd" (OuterVolumeSpecName: "kube-api-access-qwvgd") pod "02d4d565-63c4-4522-a2c5-daa9dc836057" (UID: "02d4d565-63c4-4522-a2c5-daa9dc836057"). InnerVolumeSpecName "kube-api-access-qwvgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:09:48 crc kubenswrapper[4775]: I1002 03:09:48.706716 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwvgd\" (UniqueName: \"kubernetes.io/projected/02d4d565-63c4-4522-a2c5-daa9dc836057-kube-api-access-qwvgd\") on node \"crc\" DevicePath \"\"" Oct 02 03:09:49 crc kubenswrapper[4775]: I1002 03:09:49.189174 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-zwvbk" event={"ID":"02d4d565-63c4-4522-a2c5-daa9dc836057","Type":"ContainerDied","Data":"5339b57c9603a77a070f9addce00a0ea0c6129b7117ad17cffd0bc5f9737526e"} Oct 02 03:09:49 crc kubenswrapper[4775]: I1002 03:09:49.189227 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5339b57c9603a77a070f9addce00a0ea0c6129b7117ad17cffd0bc5f9737526e" Oct 02 03:09:49 crc kubenswrapper[4775]: I1002 03:09:49.189634 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-zwvbk" Oct 02 03:09:56 crc kubenswrapper[4775]: I1002 03:09:56.244585 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-c027-account-create-rn5zs"] Oct 02 03:09:56 crc kubenswrapper[4775]: E1002 03:09:56.245502 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02d4d565-63c4-4522-a2c5-daa9dc836057" containerName="mariadb-database-create" Oct 02 03:09:56 crc kubenswrapper[4775]: I1002 03:09:56.245519 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="02d4d565-63c4-4522-a2c5-daa9dc836057" containerName="mariadb-database-create" Oct 02 03:09:56 crc kubenswrapper[4775]: I1002 03:09:56.245858 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="02d4d565-63c4-4522-a2c5-daa9dc836057" containerName="mariadb-database-create" Oct 02 03:09:56 crc kubenswrapper[4775]: I1002 03:09:56.246627 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c027-account-create-rn5zs" Oct 02 03:09:56 crc kubenswrapper[4775]: I1002 03:09:56.250483 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 02 03:09:56 crc kubenswrapper[4775]: I1002 03:09:56.264200 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c027-account-create-rn5zs"] Oct 02 03:09:56 crc kubenswrapper[4775]: I1002 03:09:56.341541 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l5rg\" (UniqueName: \"kubernetes.io/projected/fc0b72bd-79d3-4195-b494-d2c783115fd7-kube-api-access-8l5rg\") pod \"barbican-c027-account-create-rn5zs\" (UID: \"fc0b72bd-79d3-4195-b494-d2c783115fd7\") " pod="openstack/barbican-c027-account-create-rn5zs" Oct 02 03:09:56 crc kubenswrapper[4775]: I1002 03:09:56.443676 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l5rg\" (UniqueName: \"kubernetes.io/projected/fc0b72bd-79d3-4195-b494-d2c783115fd7-kube-api-access-8l5rg\") pod \"barbican-c027-account-create-rn5zs\" (UID: \"fc0b72bd-79d3-4195-b494-d2c783115fd7\") " pod="openstack/barbican-c027-account-create-rn5zs" Oct 02 03:09:56 crc kubenswrapper[4775]: I1002 03:09:56.476116 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l5rg\" (UniqueName: \"kubernetes.io/projected/fc0b72bd-79d3-4195-b494-d2c783115fd7-kube-api-access-8l5rg\") pod \"barbican-c027-account-create-rn5zs\" (UID: \"fc0b72bd-79d3-4195-b494-d2c783115fd7\") " pod="openstack/barbican-c027-account-create-rn5zs" Oct 02 03:09:56 crc kubenswrapper[4775]: I1002 03:09:56.575039 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c027-account-create-rn5zs" Oct 02 03:09:57 crc kubenswrapper[4775]: I1002 03:09:57.064271 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c027-account-create-rn5zs"] Oct 02 03:09:57 crc kubenswrapper[4775]: I1002 03:09:57.286363 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c027-account-create-rn5zs" event={"ID":"fc0b72bd-79d3-4195-b494-d2c783115fd7","Type":"ContainerStarted","Data":"bda9b3638a12955f1d1a462b8f97fd8d02dde1dce8c49dd422d50ab5bdc5b5d6"} Oct 02 03:09:57 crc kubenswrapper[4775]: I1002 03:09:57.286430 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c027-account-create-rn5zs" event={"ID":"fc0b72bd-79d3-4195-b494-d2c783115fd7","Type":"ContainerStarted","Data":"03788cb2de75b13b6fb31b724b68742175cbc89300c01300b5d8f1b7b9d42ab3"} Oct 02 03:09:57 crc kubenswrapper[4775]: I1002 03:09:57.307103 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-c027-account-create-rn5zs" podStartSLOduration=1.307069322 podStartE2EDuration="1.307069322s" podCreationTimestamp="2025-10-02 03:09:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:09:57.303160731 +0000 UTC m=+5334.469904801" watchObservedRunningTime="2025-10-02 03:09:57.307069322 +0000 UTC m=+5334.473813442" Oct 02 03:09:58 crc kubenswrapper[4775]: I1002 03:09:58.301694 4775 generic.go:334] "Generic (PLEG): container finished" podID="fc0b72bd-79d3-4195-b494-d2c783115fd7" containerID="bda9b3638a12955f1d1a462b8f97fd8d02dde1dce8c49dd422d50ab5bdc5b5d6" exitCode=0 Oct 02 03:09:58 crc kubenswrapper[4775]: I1002 03:09:58.301912 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c027-account-create-rn5zs" event={"ID":"fc0b72bd-79d3-4195-b494-d2c783115fd7","Type":"ContainerDied","Data":"bda9b3638a12955f1d1a462b8f97fd8d02dde1dce8c49dd422d50ab5bdc5b5d6"} Oct 02 03:09:59 crc kubenswrapper[4775]: I1002 03:09:59.727114 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c027-account-create-rn5zs" Oct 02 03:09:59 crc kubenswrapper[4775]: I1002 03:09:59.813734 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l5rg\" (UniqueName: \"kubernetes.io/projected/fc0b72bd-79d3-4195-b494-d2c783115fd7-kube-api-access-8l5rg\") pod \"fc0b72bd-79d3-4195-b494-d2c783115fd7\" (UID: \"fc0b72bd-79d3-4195-b494-d2c783115fd7\") " Oct 02 03:09:59 crc kubenswrapper[4775]: I1002 03:09:59.832192 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc0b72bd-79d3-4195-b494-d2c783115fd7-kube-api-access-8l5rg" (OuterVolumeSpecName: "kube-api-access-8l5rg") pod "fc0b72bd-79d3-4195-b494-d2c783115fd7" (UID: "fc0b72bd-79d3-4195-b494-d2c783115fd7"). InnerVolumeSpecName "kube-api-access-8l5rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:09:59 crc kubenswrapper[4775]: I1002 03:09:59.917199 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l5rg\" (UniqueName: \"kubernetes.io/projected/fc0b72bd-79d3-4195-b494-d2c783115fd7-kube-api-access-8l5rg\") on node \"crc\" DevicePath \"\"" Oct 02 03:10:00 crc kubenswrapper[4775]: I1002 03:10:00.322264 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c027-account-create-rn5zs" event={"ID":"fc0b72bd-79d3-4195-b494-d2c783115fd7","Type":"ContainerDied","Data":"03788cb2de75b13b6fb31b724b68742175cbc89300c01300b5d8f1b7b9d42ab3"} Oct 02 03:10:00 crc kubenswrapper[4775]: I1002 03:10:00.323134 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03788cb2de75b13b6fb31b724b68742175cbc89300c01300b5d8f1b7b9d42ab3" Oct 02 03:10:00 crc kubenswrapper[4775]: I1002 03:10:00.322539 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c027-account-create-rn5zs" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.451558 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-5758k"] Oct 02 03:10:01 crc kubenswrapper[4775]: E1002 03:10:01.452148 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc0b72bd-79d3-4195-b494-d2c783115fd7" containerName="mariadb-account-create" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.452160 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc0b72bd-79d3-4195-b494-d2c783115fd7" containerName="mariadb-account-create" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.452300 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc0b72bd-79d3-4195-b494-d2c783115fd7" containerName="mariadb-account-create" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.452798 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5758k" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.459603 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-4jr9c" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.459693 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.487373 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5758k"] Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.551347 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frh89\" (UniqueName: \"kubernetes.io/projected/f14ca3c4-6278-41a9-94b2-29045c809785-kube-api-access-frh89\") pod \"barbican-db-sync-5758k\" (UID: \"f14ca3c4-6278-41a9-94b2-29045c809785\") " pod="openstack/barbican-db-sync-5758k" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.551469 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f14ca3c4-6278-41a9-94b2-29045c809785-db-sync-config-data\") pod \"barbican-db-sync-5758k\" (UID: \"f14ca3c4-6278-41a9-94b2-29045c809785\") " pod="openstack/barbican-db-sync-5758k" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.551600 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14ca3c4-6278-41a9-94b2-29045c809785-combined-ca-bundle\") pod \"barbican-db-sync-5758k\" (UID: \"f14ca3c4-6278-41a9-94b2-29045c809785\") " pod="openstack/barbican-db-sync-5758k" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.653192 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14ca3c4-6278-41a9-94b2-29045c809785-combined-ca-bundle\") pod \"barbican-db-sync-5758k\" (UID: \"f14ca3c4-6278-41a9-94b2-29045c809785\") " pod="openstack/barbican-db-sync-5758k" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.653290 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frh89\" (UniqueName: \"kubernetes.io/projected/f14ca3c4-6278-41a9-94b2-29045c809785-kube-api-access-frh89\") pod \"barbican-db-sync-5758k\" (UID: \"f14ca3c4-6278-41a9-94b2-29045c809785\") " pod="openstack/barbican-db-sync-5758k" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.653363 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f14ca3c4-6278-41a9-94b2-29045c809785-db-sync-config-data\") pod \"barbican-db-sync-5758k\" (UID: \"f14ca3c4-6278-41a9-94b2-29045c809785\") " pod="openstack/barbican-db-sync-5758k" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.657441 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f14ca3c4-6278-41a9-94b2-29045c809785-db-sync-config-data\") pod \"barbican-db-sync-5758k\" (UID: \"f14ca3c4-6278-41a9-94b2-29045c809785\") " pod="openstack/barbican-db-sync-5758k" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.659393 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14ca3c4-6278-41a9-94b2-29045c809785-combined-ca-bundle\") pod \"barbican-db-sync-5758k\" (UID: \"f14ca3c4-6278-41a9-94b2-29045c809785\") " pod="openstack/barbican-db-sync-5758k" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.691590 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frh89\" (UniqueName: \"kubernetes.io/projected/f14ca3c4-6278-41a9-94b2-29045c809785-kube-api-access-frh89\") pod \"barbican-db-sync-5758k\" (UID: \"f14ca3c4-6278-41a9-94b2-29045c809785\") " pod="openstack/barbican-db-sync-5758k" Oct 02 03:10:01 crc kubenswrapper[4775]: I1002 03:10:01.775097 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5758k" Oct 02 03:10:02 crc kubenswrapper[4775]: I1002 03:10:02.026330 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-5758k"] Oct 02 03:10:02 crc kubenswrapper[4775]: I1002 03:10:02.343626 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5758k" event={"ID":"f14ca3c4-6278-41a9-94b2-29045c809785","Type":"ContainerStarted","Data":"9b3a5a3dbe93481567b2f28b1194c37ea150030e6f7f1589415a4fae7a391161"} Oct 02 03:10:02 crc kubenswrapper[4775]: I1002 03:10:02.344159 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5758k" event={"ID":"f14ca3c4-6278-41a9-94b2-29045c809785","Type":"ContainerStarted","Data":"65b9049ed4cf6eaf72bfa2690236e08aafc41986ca1612389ddfdaafc2dc5300"} Oct 02 03:10:02 crc kubenswrapper[4775]: I1002 03:10:02.362012 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-5758k" podStartSLOduration=1.361990573 podStartE2EDuration="1.361990573s" podCreationTimestamp="2025-10-02 03:10:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:10:02.359078428 +0000 UTC m=+5339.525822458" watchObservedRunningTime="2025-10-02 03:10:02.361990573 +0000 UTC m=+5339.528734613" Oct 02 03:10:04 crc kubenswrapper[4775]: I1002 03:10:04.368443 4775 generic.go:334] "Generic (PLEG): container finished" podID="f14ca3c4-6278-41a9-94b2-29045c809785" containerID="9b3a5a3dbe93481567b2f28b1194c37ea150030e6f7f1589415a4fae7a391161" exitCode=0 Oct 02 03:10:04 crc kubenswrapper[4775]: I1002 03:10:04.368526 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5758k" event={"ID":"f14ca3c4-6278-41a9-94b2-29045c809785","Type":"ContainerDied","Data":"9b3a5a3dbe93481567b2f28b1194c37ea150030e6f7f1589415a4fae7a391161"} Oct 02 03:10:05 crc kubenswrapper[4775]: I1002 03:10:05.763145 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5758k" Oct 02 03:10:05 crc kubenswrapper[4775]: I1002 03:10:05.929929 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frh89\" (UniqueName: \"kubernetes.io/projected/f14ca3c4-6278-41a9-94b2-29045c809785-kube-api-access-frh89\") pod \"f14ca3c4-6278-41a9-94b2-29045c809785\" (UID: \"f14ca3c4-6278-41a9-94b2-29045c809785\") " Oct 02 03:10:05 crc kubenswrapper[4775]: I1002 03:10:05.930699 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f14ca3c4-6278-41a9-94b2-29045c809785-db-sync-config-data\") pod \"f14ca3c4-6278-41a9-94b2-29045c809785\" (UID: \"f14ca3c4-6278-41a9-94b2-29045c809785\") " Oct 02 03:10:05 crc kubenswrapper[4775]: I1002 03:10:05.930757 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14ca3c4-6278-41a9-94b2-29045c809785-combined-ca-bundle\") pod \"f14ca3c4-6278-41a9-94b2-29045c809785\" (UID: \"f14ca3c4-6278-41a9-94b2-29045c809785\") " Oct 02 03:10:05 crc kubenswrapper[4775]: I1002 03:10:05.939457 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f14ca3c4-6278-41a9-94b2-29045c809785-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f14ca3c4-6278-41a9-94b2-29045c809785" (UID: "f14ca3c4-6278-41a9-94b2-29045c809785"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:10:05 crc kubenswrapper[4775]: I1002 03:10:05.939775 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f14ca3c4-6278-41a9-94b2-29045c809785-kube-api-access-frh89" (OuterVolumeSpecName: "kube-api-access-frh89") pod "f14ca3c4-6278-41a9-94b2-29045c809785" (UID: "f14ca3c4-6278-41a9-94b2-29045c809785"). InnerVolumeSpecName "kube-api-access-frh89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:10:05 crc kubenswrapper[4775]: I1002 03:10:05.957000 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f14ca3c4-6278-41a9-94b2-29045c809785-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f14ca3c4-6278-41a9-94b2-29045c809785" (UID: "f14ca3c4-6278-41a9-94b2-29045c809785"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.033203 4775 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f14ca3c4-6278-41a9-94b2-29045c809785-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.033300 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f14ca3c4-6278-41a9-94b2-29045c809785-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.033323 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frh89\" (UniqueName: \"kubernetes.io/projected/f14ca3c4-6278-41a9-94b2-29045c809785-kube-api-access-frh89\") on node \"crc\" DevicePath \"\"" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.397514 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-5758k" event={"ID":"f14ca3c4-6278-41a9-94b2-29045c809785","Type":"ContainerDied","Data":"65b9049ed4cf6eaf72bfa2690236e08aafc41986ca1612389ddfdaafc2dc5300"} Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.397579 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65b9049ed4cf6eaf72bfa2690236e08aafc41986ca1612389ddfdaafc2dc5300" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.397875 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-5758k" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.667381 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6b5587698d-j95nw"] Oct 02 03:10:06 crc kubenswrapper[4775]: E1002 03:10:06.667797 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f14ca3c4-6278-41a9-94b2-29045c809785" containerName="barbican-db-sync" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.667813 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f14ca3c4-6278-41a9-94b2-29045c809785" containerName="barbican-db-sync" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.668060 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f14ca3c4-6278-41a9-94b2-29045c809785" containerName="barbican-db-sync" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.669171 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.672883 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.673155 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-4jr9c" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.683563 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.683869 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6f755bf8ff-wb6kc"] Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.686235 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.691365 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6b5587698d-j95nw"] Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.693362 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.727683 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6f755bf8ff-wb6kc"] Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.746650 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42lg4\" (UniqueName: \"kubernetes.io/projected/445125c9-802a-4759-ab11-11845f420062-kube-api-access-42lg4\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.746731 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f069d67d-c773-4393-ac3d-ef520b69355b-combined-ca-bundle\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.746774 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4rh5\" (UniqueName: \"kubernetes.io/projected/f069d67d-c773-4393-ac3d-ef520b69355b-kube-api-access-k4rh5\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.746827 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f069d67d-c773-4393-ac3d-ef520b69355b-logs\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.746879 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/445125c9-802a-4759-ab11-11845f420062-combined-ca-bundle\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.747038 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f069d67d-c773-4393-ac3d-ef520b69355b-config-data-custom\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.747083 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/445125c9-802a-4759-ab11-11845f420062-logs\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.747100 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/445125c9-802a-4759-ab11-11845f420062-config-data-custom\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.747121 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f069d67d-c773-4393-ac3d-ef520b69355b-config-data\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.747152 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/445125c9-802a-4759-ab11-11845f420062-config-data\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.761172 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67476cd545-bdlxs"] Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.763005 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.770315 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67476cd545-bdlxs"] Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.833848 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5f5d675bc4-lw5dk"] Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.835091 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.839603 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.849600 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/445125c9-802a-4759-ab11-11845f420062-config-data\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.849647 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42lg4\" (UniqueName: \"kubernetes.io/projected/445125c9-802a-4759-ab11-11845f420062-kube-api-access-42lg4\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.849681 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f069d67d-c773-4393-ac3d-ef520b69355b-combined-ca-bundle\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.849702 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4rh5\" (UniqueName: \"kubernetes.io/projected/f069d67d-c773-4393-ac3d-ef520b69355b-kube-api-access-k4rh5\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.849731 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f069d67d-c773-4393-ac3d-ef520b69355b-logs\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.849757 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/445125c9-802a-4759-ab11-11845f420062-combined-ca-bundle\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.849799 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f069d67d-c773-4393-ac3d-ef520b69355b-config-data-custom\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.849819 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/445125c9-802a-4759-ab11-11845f420062-logs\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.849835 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/445125c9-802a-4759-ab11-11845f420062-config-data-custom\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.849850 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f069d67d-c773-4393-ac3d-ef520b69355b-config-data\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.851199 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f5d675bc4-lw5dk"] Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.851779 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f069d67d-c773-4393-ac3d-ef520b69355b-logs\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.852051 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/445125c9-802a-4759-ab11-11845f420062-logs\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.854463 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f069d67d-c773-4393-ac3d-ef520b69355b-config-data\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.856236 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/445125c9-802a-4759-ab11-11845f420062-config-data-custom\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.858186 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f069d67d-c773-4393-ac3d-ef520b69355b-config-data-custom\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.863319 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/445125c9-802a-4759-ab11-11845f420062-combined-ca-bundle\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.864298 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f069d67d-c773-4393-ac3d-ef520b69355b-combined-ca-bundle\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.864620 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/445125c9-802a-4759-ab11-11845f420062-config-data\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.875217 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4rh5\" (UniqueName: \"kubernetes.io/projected/f069d67d-c773-4393-ac3d-ef520b69355b-kube-api-access-k4rh5\") pod \"barbican-worker-6f755bf8ff-wb6kc\" (UID: \"f069d67d-c773-4393-ac3d-ef520b69355b\") " pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.878046 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42lg4\" (UniqueName: \"kubernetes.io/projected/445125c9-802a-4759-ab11-11845f420062-kube-api-access-42lg4\") pod \"barbican-keystone-listener-6b5587698d-j95nw\" (UID: \"445125c9-802a-4759-ab11-11845f420062\") " pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.951529 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-dns-svc\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.951579 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7463fd90-1ce1-484f-b3ca-499e93e46f8e-config-data-custom\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.951654 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-config\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.952263 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7463fd90-1ce1-484f-b3ca-499e93e46f8e-config-data\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.952281 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7463fd90-1ce1-484f-b3ca-499e93e46f8e-logs\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.952299 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkrxc\" (UniqueName: \"kubernetes.io/projected/7463fd90-1ce1-484f-b3ca-499e93e46f8e-kube-api-access-pkrxc\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.952408 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-ovsdbserver-nb\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.952445 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-ovsdbserver-sb\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.952576 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7463fd90-1ce1-484f-b3ca-499e93e46f8e-combined-ca-bundle\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:06 crc kubenswrapper[4775]: I1002 03:10:06.952598 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbzb4\" (UniqueName: \"kubernetes.io/projected/5125f6d1-9611-4d09-bebe-b4d7b3e99460-kube-api-access-bbzb4\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.022826 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.053823 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6f755bf8ff-wb6kc" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.053906 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7463fd90-1ce1-484f-b3ca-499e93e46f8e-combined-ca-bundle\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.054721 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbzb4\" (UniqueName: \"kubernetes.io/projected/5125f6d1-9611-4d09-bebe-b4d7b3e99460-kube-api-access-bbzb4\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.054801 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-dns-svc\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.054836 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7463fd90-1ce1-484f-b3ca-499e93e46f8e-config-data-custom\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.054872 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-config\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.054975 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7463fd90-1ce1-484f-b3ca-499e93e46f8e-config-data\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.054998 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7463fd90-1ce1-484f-b3ca-499e93e46f8e-logs\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.055020 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkrxc\" (UniqueName: \"kubernetes.io/projected/7463fd90-1ce1-484f-b3ca-499e93e46f8e-kube-api-access-pkrxc\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.055080 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-ovsdbserver-nb\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.055123 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-ovsdbserver-sb\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.055780 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-dns-svc\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.056078 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-ovsdbserver-sb\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.056442 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7463fd90-1ce1-484f-b3ca-499e93e46f8e-logs\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.056886 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-config\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.058448 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7463fd90-1ce1-484f-b3ca-499e93e46f8e-combined-ca-bundle\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.059054 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-ovsdbserver-nb\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.062325 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7463fd90-1ce1-484f-b3ca-499e93e46f8e-config-data-custom\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.065343 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7463fd90-1ce1-484f-b3ca-499e93e46f8e-config-data\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.073164 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbzb4\" (UniqueName: \"kubernetes.io/projected/5125f6d1-9611-4d09-bebe-b4d7b3e99460-kube-api-access-bbzb4\") pod \"dnsmasq-dns-67476cd545-bdlxs\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.073665 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkrxc\" (UniqueName: \"kubernetes.io/projected/7463fd90-1ce1-484f-b3ca-499e93e46f8e-kube-api-access-pkrxc\") pod \"barbican-api-5f5d675bc4-lw5dk\" (UID: \"7463fd90-1ce1-484f-b3ca-499e93e46f8e\") " pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.095514 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.162842 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.522780 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6f755bf8ff-wb6kc"] Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.598552 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6b5587698d-j95nw"] Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.643630 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67476cd545-bdlxs"] Oct 02 03:10:07 crc kubenswrapper[4775]: W1002 03:10:07.658088 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5125f6d1_9611_4d09_bebe_b4d7b3e99460.slice/crio-108eae153dd20ef6a08aa048f8f2290dd43c178233c815a57504cf55eb345e6b WatchSource:0}: Error finding container 108eae153dd20ef6a08aa048f8f2290dd43c178233c815a57504cf55eb345e6b: Status 404 returned error can't find the container with id 108eae153dd20ef6a08aa048f8f2290dd43c178233c815a57504cf55eb345e6b Oct 02 03:10:07 crc kubenswrapper[4775]: I1002 03:10:07.663463 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5f5d675bc4-lw5dk"] Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.417679 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f5d675bc4-lw5dk" event={"ID":"7463fd90-1ce1-484f-b3ca-499e93e46f8e","Type":"ContainerStarted","Data":"2f490d6a2ac68898d22caa882cef6ae5bffecc9b302ba16a6a8d5dfc5f505878"} Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.417720 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f5d675bc4-lw5dk" event={"ID":"7463fd90-1ce1-484f-b3ca-499e93e46f8e","Type":"ContainerStarted","Data":"1d4e5e9d9b389509d52c250043256b705e316a780e10219de2569010f0d40d06"} Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.417742 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5f5d675bc4-lw5dk" event={"ID":"7463fd90-1ce1-484f-b3ca-499e93e46f8e","Type":"ContainerStarted","Data":"14ca88b8774f2869a7d9ead3c6c13d98b72c1cab9d2406dff6020e8e1d9b7c86"} Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.417810 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.419512 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" event={"ID":"445125c9-802a-4759-ab11-11845f420062","Type":"ContainerStarted","Data":"31102e36c21af99b0d341d3fef511441a6086da2b3a2c09a39984a78a938fc47"} Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.419584 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" event={"ID":"445125c9-802a-4759-ab11-11845f420062","Type":"ContainerStarted","Data":"7b6f03133fe15e512230d34f08a430c5384eca7b5dd3e2c22680a080ae5deb07"} Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.419595 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" event={"ID":"445125c9-802a-4759-ab11-11845f420062","Type":"ContainerStarted","Data":"f93a807f75406fa3c555c6fa04ae99261c60f4b2ad8c8fe03e9ba95cbd62c6be"} Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.422093 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6f755bf8ff-wb6kc" event={"ID":"f069d67d-c773-4393-ac3d-ef520b69355b","Type":"ContainerStarted","Data":"686b01410564acf87acbff85883388224018ca0c19d70d0131a8beadabfdc4c9"} Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.422134 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6f755bf8ff-wb6kc" event={"ID":"f069d67d-c773-4393-ac3d-ef520b69355b","Type":"ContainerStarted","Data":"15265e181227231fa96dbfd801ddfdb457428d6be904f9b736d016b631ccbdcd"} Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.422144 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6f755bf8ff-wb6kc" event={"ID":"f069d67d-c773-4393-ac3d-ef520b69355b","Type":"ContainerStarted","Data":"38a3fdd819436d766479c1f6b036f8434f7a59d384049688e6022407f710a42e"} Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.424376 4775 generic.go:334] "Generic (PLEG): container finished" podID="5125f6d1-9611-4d09-bebe-b4d7b3e99460" containerID="21cd184b426ef67a8be2aafaac5325aa43b84823fcbfbd2e51d22adaebadbe26" exitCode=0 Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.424415 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67476cd545-bdlxs" event={"ID":"5125f6d1-9611-4d09-bebe-b4d7b3e99460","Type":"ContainerDied","Data":"21cd184b426ef67a8be2aafaac5325aa43b84823fcbfbd2e51d22adaebadbe26"} Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.424434 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67476cd545-bdlxs" event={"ID":"5125f6d1-9611-4d09-bebe-b4d7b3e99460","Type":"ContainerStarted","Data":"108eae153dd20ef6a08aa048f8f2290dd43c178233c815a57504cf55eb345e6b"} Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.441671 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5f5d675bc4-lw5dk" podStartSLOduration=2.44163847 podStartE2EDuration="2.44163847s" podCreationTimestamp="2025-10-02 03:10:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:10:08.43660168 +0000 UTC m=+5345.603345720" watchObservedRunningTime="2025-10-02 03:10:08.44163847 +0000 UTC m=+5345.608382510" Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.459218 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6b5587698d-j95nw" podStartSLOduration=2.459201533 podStartE2EDuration="2.459201533s" podCreationTimestamp="2025-10-02 03:10:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:10:08.451734231 +0000 UTC m=+5345.618478281" watchObservedRunningTime="2025-10-02 03:10:08.459201533 +0000 UTC m=+5345.625945573" Oct 02 03:10:08 crc kubenswrapper[4775]: I1002 03:10:08.497317 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6f755bf8ff-wb6kc" podStartSLOduration=2.497298065 podStartE2EDuration="2.497298065s" podCreationTimestamp="2025-10-02 03:10:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:10:08.493459056 +0000 UTC m=+5345.660203096" watchObservedRunningTime="2025-10-02 03:10:08.497298065 +0000 UTC m=+5345.664042105" Oct 02 03:10:09 crc kubenswrapper[4775]: I1002 03:10:09.441707 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67476cd545-bdlxs" event={"ID":"5125f6d1-9611-4d09-bebe-b4d7b3e99460","Type":"ContainerStarted","Data":"2329878b563daff78e069e7606b84295cf44f86a97265abb55cd587fac699ae7"} Oct 02 03:10:09 crc kubenswrapper[4775]: I1002 03:10:09.442576 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:09 crc kubenswrapper[4775]: I1002 03:10:09.477838 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67476cd545-bdlxs" podStartSLOduration=3.477774328 podStartE2EDuration="3.477774328s" podCreationTimestamp="2025-10-02 03:10:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:10:09.471751613 +0000 UTC m=+5346.638495693" watchObservedRunningTime="2025-10-02 03:10:09.477774328 +0000 UTC m=+5346.644518408" Oct 02 03:10:10 crc kubenswrapper[4775]: I1002 03:10:10.452011 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:13 crc kubenswrapper[4775]: I1002 03:10:13.563224 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:15 crc kubenswrapper[4775]: I1002 03:10:15.059467 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5f5d675bc4-lw5dk" Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.097259 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.216872 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f7dfc9995-xlkfd"] Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.217251 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" podUID="fc13c15d-c9db-4039-8526-1151e2c080cc" containerName="dnsmasq-dns" containerID="cri-o://11ba9dcd44e21f64181ec61b63a445c44883856a6d4441608fe04c5c93765def" gracePeriod=10 Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.562147 4775 generic.go:334] "Generic (PLEG): container finished" podID="fc13c15d-c9db-4039-8526-1151e2c080cc" containerID="11ba9dcd44e21f64181ec61b63a445c44883856a6d4441608fe04c5c93765def" exitCode=0 Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.562228 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" event={"ID":"fc13c15d-c9db-4039-8526-1151e2c080cc","Type":"ContainerDied","Data":"11ba9dcd44e21f64181ec61b63a445c44883856a6d4441608fe04c5c93765def"} Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.686135 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.784106 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-ovsdbserver-nb\") pod \"fc13c15d-c9db-4039-8526-1151e2c080cc\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.784220 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-ovsdbserver-sb\") pod \"fc13c15d-c9db-4039-8526-1151e2c080cc\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.784302 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-config\") pod \"fc13c15d-c9db-4039-8526-1151e2c080cc\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.784381 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gprcf\" (UniqueName: \"kubernetes.io/projected/fc13c15d-c9db-4039-8526-1151e2c080cc-kube-api-access-gprcf\") pod \"fc13c15d-c9db-4039-8526-1151e2c080cc\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.784468 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-dns-svc\") pod \"fc13c15d-c9db-4039-8526-1151e2c080cc\" (UID: \"fc13c15d-c9db-4039-8526-1151e2c080cc\") " Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.823193 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc13c15d-c9db-4039-8526-1151e2c080cc-kube-api-access-gprcf" (OuterVolumeSpecName: "kube-api-access-gprcf") pod "fc13c15d-c9db-4039-8526-1151e2c080cc" (UID: "fc13c15d-c9db-4039-8526-1151e2c080cc"). InnerVolumeSpecName "kube-api-access-gprcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.855999 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fc13c15d-c9db-4039-8526-1151e2c080cc" (UID: "fc13c15d-c9db-4039-8526-1151e2c080cc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.872457 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-config" (OuterVolumeSpecName: "config") pod "fc13c15d-c9db-4039-8526-1151e2c080cc" (UID: "fc13c15d-c9db-4039-8526-1151e2c080cc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.876484 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fc13c15d-c9db-4039-8526-1151e2c080cc" (UID: "fc13c15d-c9db-4039-8526-1151e2c080cc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.895510 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.895665 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gprcf\" (UniqueName: \"kubernetes.io/projected/fc13c15d-c9db-4039-8526-1151e2c080cc-kube-api-access-gprcf\") on node \"crc\" DevicePath \"\"" Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.895703 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.895721 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.899629 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fc13c15d-c9db-4039-8526-1151e2c080cc" (UID: "fc13c15d-c9db-4039-8526-1151e2c080cc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:10:17 crc kubenswrapper[4775]: I1002 03:10:17.997069 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fc13c15d-c9db-4039-8526-1151e2c080cc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 03:10:18 crc kubenswrapper[4775]: I1002 03:10:18.572419 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" event={"ID":"fc13c15d-c9db-4039-8526-1151e2c080cc","Type":"ContainerDied","Data":"18c28f488d8f74ab79ee1ea9299a7f4f009888b1031eece2b77ab0256a1a8ef3"} Oct 02 03:10:18 crc kubenswrapper[4775]: I1002 03:10:18.572491 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f7dfc9995-xlkfd" Oct 02 03:10:18 crc kubenswrapper[4775]: I1002 03:10:18.572507 4775 scope.go:117] "RemoveContainer" containerID="11ba9dcd44e21f64181ec61b63a445c44883856a6d4441608fe04c5c93765def" Oct 02 03:10:18 crc kubenswrapper[4775]: I1002 03:10:18.606112 4775 scope.go:117] "RemoveContainer" containerID="ff51368a5a8e78612a5228b0debf8f912ec18d1e713c7508b98c85b0670975de" Oct 02 03:10:18 crc kubenswrapper[4775]: I1002 03:10:18.607403 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f7dfc9995-xlkfd"] Oct 02 03:10:18 crc kubenswrapper[4775]: I1002 03:10:18.617284 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f7dfc9995-xlkfd"] Oct 02 03:10:19 crc kubenswrapper[4775]: I1002 03:10:19.776512 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc13c15d-c9db-4039-8526-1151e2c080cc" path="/var/lib/kubelet/pods/fc13c15d-c9db-4039-8526-1151e2c080cc/volumes" Oct 02 03:10:29 crc kubenswrapper[4775]: I1002 03:10:29.648280 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-66vnq"] Oct 02 03:10:29 crc kubenswrapper[4775]: E1002 03:10:29.648993 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc13c15d-c9db-4039-8526-1151e2c080cc" containerName="init" Oct 02 03:10:29 crc kubenswrapper[4775]: I1002 03:10:29.649004 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc13c15d-c9db-4039-8526-1151e2c080cc" containerName="init" Oct 02 03:10:29 crc kubenswrapper[4775]: E1002 03:10:29.649032 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc13c15d-c9db-4039-8526-1151e2c080cc" containerName="dnsmasq-dns" Oct 02 03:10:29 crc kubenswrapper[4775]: I1002 03:10:29.649037 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc13c15d-c9db-4039-8526-1151e2c080cc" containerName="dnsmasq-dns" Oct 02 03:10:29 crc kubenswrapper[4775]: I1002 03:10:29.649195 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc13c15d-c9db-4039-8526-1151e2c080cc" containerName="dnsmasq-dns" Oct 02 03:10:29 crc kubenswrapper[4775]: I1002 03:10:29.649746 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-66vnq" Oct 02 03:10:29 crc kubenswrapper[4775]: I1002 03:10:29.670197 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-66vnq"] Oct 02 03:10:29 crc kubenswrapper[4775]: I1002 03:10:29.721695 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zp6n\" (UniqueName: \"kubernetes.io/projected/3050a03c-11f3-416c-a630-a75c621e949e-kube-api-access-6zp6n\") pod \"neutron-db-create-66vnq\" (UID: \"3050a03c-11f3-416c-a630-a75c621e949e\") " pod="openstack/neutron-db-create-66vnq" Oct 02 03:10:29 crc kubenswrapper[4775]: I1002 03:10:29.823030 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zp6n\" (UniqueName: \"kubernetes.io/projected/3050a03c-11f3-416c-a630-a75c621e949e-kube-api-access-6zp6n\") pod \"neutron-db-create-66vnq\" (UID: \"3050a03c-11f3-416c-a630-a75c621e949e\") " pod="openstack/neutron-db-create-66vnq" Oct 02 03:10:29 crc kubenswrapper[4775]: I1002 03:10:29.841381 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zp6n\" (UniqueName: \"kubernetes.io/projected/3050a03c-11f3-416c-a630-a75c621e949e-kube-api-access-6zp6n\") pod \"neutron-db-create-66vnq\" (UID: \"3050a03c-11f3-416c-a630-a75c621e949e\") " pod="openstack/neutron-db-create-66vnq" Oct 02 03:10:29 crc kubenswrapper[4775]: I1002 03:10:29.966517 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-66vnq" Oct 02 03:10:30 crc kubenswrapper[4775]: I1002 03:10:30.429465 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-66vnq"] Oct 02 03:10:30 crc kubenswrapper[4775]: W1002 03:10:30.431363 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3050a03c_11f3_416c_a630_a75c621e949e.slice/crio-b5ab3d72a57cd3d297d8f639654a43721f04e509d6f1067a583cbc603b495930 WatchSource:0}: Error finding container b5ab3d72a57cd3d297d8f639654a43721f04e509d6f1067a583cbc603b495930: Status 404 returned error can't find the container with id b5ab3d72a57cd3d297d8f639654a43721f04e509d6f1067a583cbc603b495930 Oct 02 03:10:30 crc kubenswrapper[4775]: I1002 03:10:30.710689 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-66vnq" event={"ID":"3050a03c-11f3-416c-a630-a75c621e949e","Type":"ContainerStarted","Data":"e905bf68bfd9cdbb864c08467b80a04191c06dfad7ceaa3f15650d5c6ee2c5dd"} Oct 02 03:10:30 crc kubenswrapper[4775]: I1002 03:10:30.711121 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-66vnq" event={"ID":"3050a03c-11f3-416c-a630-a75c621e949e","Type":"ContainerStarted","Data":"b5ab3d72a57cd3d297d8f639654a43721f04e509d6f1067a583cbc603b495930"} Oct 02 03:10:30 crc kubenswrapper[4775]: I1002 03:10:30.733971 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-66vnq" podStartSLOduration=1.733936043 podStartE2EDuration="1.733936043s" podCreationTimestamp="2025-10-02 03:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:10:30.727946348 +0000 UTC m=+5367.894690398" watchObservedRunningTime="2025-10-02 03:10:30.733936043 +0000 UTC m=+5367.900680093" Oct 02 03:10:31 crc kubenswrapper[4775]: I1002 03:10:31.724803 4775 generic.go:334] "Generic (PLEG): container finished" podID="3050a03c-11f3-416c-a630-a75c621e949e" containerID="e905bf68bfd9cdbb864c08467b80a04191c06dfad7ceaa3f15650d5c6ee2c5dd" exitCode=0 Oct 02 03:10:31 crc kubenswrapper[4775]: I1002 03:10:31.724937 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-66vnq" event={"ID":"3050a03c-11f3-416c-a630-a75c621e949e","Type":"ContainerDied","Data":"e905bf68bfd9cdbb864c08467b80a04191c06dfad7ceaa3f15650d5c6ee2c5dd"} Oct 02 03:10:33 crc kubenswrapper[4775]: I1002 03:10:33.154910 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-66vnq" Oct 02 03:10:33 crc kubenswrapper[4775]: I1002 03:10:33.217389 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zp6n\" (UniqueName: \"kubernetes.io/projected/3050a03c-11f3-416c-a630-a75c621e949e-kube-api-access-6zp6n\") pod \"3050a03c-11f3-416c-a630-a75c621e949e\" (UID: \"3050a03c-11f3-416c-a630-a75c621e949e\") " Oct 02 03:10:33 crc kubenswrapper[4775]: I1002 03:10:33.226676 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3050a03c-11f3-416c-a630-a75c621e949e-kube-api-access-6zp6n" (OuterVolumeSpecName: "kube-api-access-6zp6n") pod "3050a03c-11f3-416c-a630-a75c621e949e" (UID: "3050a03c-11f3-416c-a630-a75c621e949e"). InnerVolumeSpecName "kube-api-access-6zp6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:10:33 crc kubenswrapper[4775]: I1002 03:10:33.320191 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zp6n\" (UniqueName: \"kubernetes.io/projected/3050a03c-11f3-416c-a630-a75c621e949e-kube-api-access-6zp6n\") on node \"crc\" DevicePath \"\"" Oct 02 03:10:33 crc kubenswrapper[4775]: I1002 03:10:33.754495 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-66vnq" event={"ID":"3050a03c-11f3-416c-a630-a75c621e949e","Type":"ContainerDied","Data":"b5ab3d72a57cd3d297d8f639654a43721f04e509d6f1067a583cbc603b495930"} Oct 02 03:10:33 crc kubenswrapper[4775]: I1002 03:10:33.754556 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5ab3d72a57cd3d297d8f639654a43721f04e509d6f1067a583cbc603b495930" Oct 02 03:10:33 crc kubenswrapper[4775]: I1002 03:10:33.754639 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-66vnq" Oct 02 03:10:39 crc kubenswrapper[4775]: I1002 03:10:39.677037 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6871-account-create-tprh9"] Oct 02 03:10:39 crc kubenswrapper[4775]: E1002 03:10:39.678535 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3050a03c-11f3-416c-a630-a75c621e949e" containerName="mariadb-database-create" Oct 02 03:10:39 crc kubenswrapper[4775]: I1002 03:10:39.678560 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3050a03c-11f3-416c-a630-a75c621e949e" containerName="mariadb-database-create" Oct 02 03:10:39 crc kubenswrapper[4775]: I1002 03:10:39.678892 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3050a03c-11f3-416c-a630-a75c621e949e" containerName="mariadb-database-create" Oct 02 03:10:39 crc kubenswrapper[4775]: I1002 03:10:39.680413 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6871-account-create-tprh9" Oct 02 03:10:39 crc kubenswrapper[4775]: I1002 03:10:39.683726 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 02 03:10:39 crc kubenswrapper[4775]: I1002 03:10:39.685330 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6871-account-create-tprh9"] Oct 02 03:10:39 crc kubenswrapper[4775]: I1002 03:10:39.757500 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gt9hd\" (UniqueName: \"kubernetes.io/projected/236c95d5-4c86-40e1-8be3-ddf52e119401-kube-api-access-gt9hd\") pod \"neutron-6871-account-create-tprh9\" (UID: \"236c95d5-4c86-40e1-8be3-ddf52e119401\") " pod="openstack/neutron-6871-account-create-tprh9" Oct 02 03:10:39 crc kubenswrapper[4775]: I1002 03:10:39.859431 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gt9hd\" (UniqueName: \"kubernetes.io/projected/236c95d5-4c86-40e1-8be3-ddf52e119401-kube-api-access-gt9hd\") pod \"neutron-6871-account-create-tprh9\" (UID: \"236c95d5-4c86-40e1-8be3-ddf52e119401\") " pod="openstack/neutron-6871-account-create-tprh9" Oct 02 03:10:39 crc kubenswrapper[4775]: I1002 03:10:39.902209 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gt9hd\" (UniqueName: \"kubernetes.io/projected/236c95d5-4c86-40e1-8be3-ddf52e119401-kube-api-access-gt9hd\") pod \"neutron-6871-account-create-tprh9\" (UID: \"236c95d5-4c86-40e1-8be3-ddf52e119401\") " pod="openstack/neutron-6871-account-create-tprh9" Oct 02 03:10:40 crc kubenswrapper[4775]: I1002 03:10:40.020807 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6871-account-create-tprh9" Oct 02 03:10:40 crc kubenswrapper[4775]: I1002 03:10:40.332112 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6871-account-create-tprh9"] Oct 02 03:10:40 crc kubenswrapper[4775]: I1002 03:10:40.844167 4775 generic.go:334] "Generic (PLEG): container finished" podID="236c95d5-4c86-40e1-8be3-ddf52e119401" containerID="2e7aa961aab766d93eaa70f5c14beaf43a65f60065d58a23ab6696ce743dfcb1" exitCode=0 Oct 02 03:10:40 crc kubenswrapper[4775]: I1002 03:10:40.844226 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6871-account-create-tprh9" event={"ID":"236c95d5-4c86-40e1-8be3-ddf52e119401","Type":"ContainerDied","Data":"2e7aa961aab766d93eaa70f5c14beaf43a65f60065d58a23ab6696ce743dfcb1"} Oct 02 03:10:40 crc kubenswrapper[4775]: I1002 03:10:40.844264 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6871-account-create-tprh9" event={"ID":"236c95d5-4c86-40e1-8be3-ddf52e119401","Type":"ContainerStarted","Data":"9882ee2b5af5837629364ff7d8cc2b6ab0c9fdb4eaedcfdcbd22f1addf773ef4"} Oct 02 03:10:42 crc kubenswrapper[4775]: I1002 03:10:42.303665 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6871-account-create-tprh9" Oct 02 03:10:42 crc kubenswrapper[4775]: I1002 03:10:42.433821 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gt9hd\" (UniqueName: \"kubernetes.io/projected/236c95d5-4c86-40e1-8be3-ddf52e119401-kube-api-access-gt9hd\") pod \"236c95d5-4c86-40e1-8be3-ddf52e119401\" (UID: \"236c95d5-4c86-40e1-8be3-ddf52e119401\") " Oct 02 03:10:42 crc kubenswrapper[4775]: I1002 03:10:42.455051 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/236c95d5-4c86-40e1-8be3-ddf52e119401-kube-api-access-gt9hd" (OuterVolumeSpecName: "kube-api-access-gt9hd") pod "236c95d5-4c86-40e1-8be3-ddf52e119401" (UID: "236c95d5-4c86-40e1-8be3-ddf52e119401"). InnerVolumeSpecName "kube-api-access-gt9hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:10:42 crc kubenswrapper[4775]: I1002 03:10:42.537740 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gt9hd\" (UniqueName: \"kubernetes.io/projected/236c95d5-4c86-40e1-8be3-ddf52e119401-kube-api-access-gt9hd\") on node \"crc\" DevicePath \"\"" Oct 02 03:10:42 crc kubenswrapper[4775]: I1002 03:10:42.865226 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6871-account-create-tprh9" event={"ID":"236c95d5-4c86-40e1-8be3-ddf52e119401","Type":"ContainerDied","Data":"9882ee2b5af5837629364ff7d8cc2b6ab0c9fdb4eaedcfdcbd22f1addf773ef4"} Oct 02 03:10:42 crc kubenswrapper[4775]: I1002 03:10:42.865572 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9882ee2b5af5837629364ff7d8cc2b6ab0c9fdb4eaedcfdcbd22f1addf773ef4" Oct 02 03:10:42 crc kubenswrapper[4775]: I1002 03:10:42.865383 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6871-account-create-tprh9" Oct 02 03:10:44 crc kubenswrapper[4775]: I1002 03:10:44.933466 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-q9nq4"] Oct 02 03:10:44 crc kubenswrapper[4775]: E1002 03:10:44.934160 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="236c95d5-4c86-40e1-8be3-ddf52e119401" containerName="mariadb-account-create" Oct 02 03:10:44 crc kubenswrapper[4775]: I1002 03:10:44.934177 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="236c95d5-4c86-40e1-8be3-ddf52e119401" containerName="mariadb-account-create" Oct 02 03:10:44 crc kubenswrapper[4775]: I1002 03:10:44.934370 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="236c95d5-4c86-40e1-8be3-ddf52e119401" containerName="mariadb-account-create" Oct 02 03:10:44 crc kubenswrapper[4775]: I1002 03:10:44.935270 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-q9nq4" Oct 02 03:10:44 crc kubenswrapper[4775]: I1002 03:10:44.945819 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-kxqj6" Oct 02 03:10:44 crc kubenswrapper[4775]: I1002 03:10:44.946583 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 03:10:44 crc kubenswrapper[4775]: I1002 03:10:44.946814 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 03:10:44 crc kubenswrapper[4775]: I1002 03:10:44.951573 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-q9nq4"] Oct 02 03:10:45 crc kubenswrapper[4775]: I1002 03:10:45.086818 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-combined-ca-bundle\") pod \"neutron-db-sync-q9nq4\" (UID: \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\") " pod="openstack/neutron-db-sync-q9nq4" Oct 02 03:10:45 crc kubenswrapper[4775]: I1002 03:10:45.087104 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqp8b\" (UniqueName: \"kubernetes.io/projected/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-kube-api-access-nqp8b\") pod \"neutron-db-sync-q9nq4\" (UID: \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\") " pod="openstack/neutron-db-sync-q9nq4" Oct 02 03:10:45 crc kubenswrapper[4775]: I1002 03:10:45.087199 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-config\") pod \"neutron-db-sync-q9nq4\" (UID: \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\") " pod="openstack/neutron-db-sync-q9nq4" Oct 02 03:10:45 crc kubenswrapper[4775]: I1002 03:10:45.189397 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqp8b\" (UniqueName: \"kubernetes.io/projected/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-kube-api-access-nqp8b\") pod \"neutron-db-sync-q9nq4\" (UID: \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\") " pod="openstack/neutron-db-sync-q9nq4" Oct 02 03:10:45 crc kubenswrapper[4775]: I1002 03:10:45.189470 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-config\") pod \"neutron-db-sync-q9nq4\" (UID: \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\") " pod="openstack/neutron-db-sync-q9nq4" Oct 02 03:10:45 crc kubenswrapper[4775]: I1002 03:10:45.189529 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-combined-ca-bundle\") pod \"neutron-db-sync-q9nq4\" (UID: \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\") " pod="openstack/neutron-db-sync-q9nq4" Oct 02 03:10:45 crc kubenswrapper[4775]: I1002 03:10:45.202254 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-config\") pod \"neutron-db-sync-q9nq4\" (UID: \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\") " pod="openstack/neutron-db-sync-q9nq4" Oct 02 03:10:45 crc kubenswrapper[4775]: I1002 03:10:45.207144 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-combined-ca-bundle\") pod \"neutron-db-sync-q9nq4\" (UID: \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\") " pod="openstack/neutron-db-sync-q9nq4" Oct 02 03:10:45 crc kubenswrapper[4775]: I1002 03:10:45.212232 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqp8b\" (UniqueName: \"kubernetes.io/projected/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-kube-api-access-nqp8b\") pod \"neutron-db-sync-q9nq4\" (UID: \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\") " pod="openstack/neutron-db-sync-q9nq4" Oct 02 03:10:45 crc kubenswrapper[4775]: I1002 03:10:45.294627 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-q9nq4" Oct 02 03:10:45 crc kubenswrapper[4775]: I1002 03:10:45.614406 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-q9nq4"] Oct 02 03:10:45 crc kubenswrapper[4775]: W1002 03:10:45.628717 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56c0f90d_dd66_4bfd_9db2_52fcefd3919d.slice/crio-83cd93e1772e99ac52b313eea814da856c626d58ef41919f4304b4ccdc2df7b4 WatchSource:0}: Error finding container 83cd93e1772e99ac52b313eea814da856c626d58ef41919f4304b4ccdc2df7b4: Status 404 returned error can't find the container with id 83cd93e1772e99ac52b313eea814da856c626d58ef41919f4304b4ccdc2df7b4 Oct 02 03:10:45 crc kubenswrapper[4775]: I1002 03:10:45.899316 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-q9nq4" event={"ID":"56c0f90d-dd66-4bfd-9db2-52fcefd3919d","Type":"ContainerStarted","Data":"f2ab1daf84a1043ef3c9cd1424fc25b5d2a9bb1398312ed18b0edd2a5951414d"} Oct 02 03:10:45 crc kubenswrapper[4775]: I1002 03:10:45.899725 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-q9nq4" event={"ID":"56c0f90d-dd66-4bfd-9db2-52fcefd3919d","Type":"ContainerStarted","Data":"83cd93e1772e99ac52b313eea814da856c626d58ef41919f4304b4ccdc2df7b4"} Oct 02 03:10:45 crc kubenswrapper[4775]: I1002 03:10:45.918884 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-q9nq4" podStartSLOduration=1.918863636 podStartE2EDuration="1.918863636s" podCreationTimestamp="2025-10-02 03:10:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:10:45.915985542 +0000 UTC m=+5383.082729582" watchObservedRunningTime="2025-10-02 03:10:45.918863636 +0000 UTC m=+5383.085607686" Oct 02 03:10:49 crc kubenswrapper[4775]: I1002 03:10:49.951162 4775 generic.go:334] "Generic (PLEG): container finished" podID="56c0f90d-dd66-4bfd-9db2-52fcefd3919d" containerID="f2ab1daf84a1043ef3c9cd1424fc25b5d2a9bb1398312ed18b0edd2a5951414d" exitCode=0 Oct 02 03:10:49 crc kubenswrapper[4775]: I1002 03:10:49.951228 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-q9nq4" event={"ID":"56c0f90d-dd66-4bfd-9db2-52fcefd3919d","Type":"ContainerDied","Data":"f2ab1daf84a1043ef3c9cd1424fc25b5d2a9bb1398312ed18b0edd2a5951414d"} Oct 02 03:10:51 crc kubenswrapper[4775]: I1002 03:10:51.378196 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-q9nq4" Oct 02 03:10:51 crc kubenswrapper[4775]: I1002 03:10:51.534536 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-config\") pod \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\" (UID: \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\") " Oct 02 03:10:51 crc kubenswrapper[4775]: I1002 03:10:51.534665 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqp8b\" (UniqueName: \"kubernetes.io/projected/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-kube-api-access-nqp8b\") pod \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\" (UID: \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\") " Oct 02 03:10:51 crc kubenswrapper[4775]: I1002 03:10:51.534722 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-combined-ca-bundle\") pod \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\" (UID: \"56c0f90d-dd66-4bfd-9db2-52fcefd3919d\") " Oct 02 03:10:51 crc kubenswrapper[4775]: I1002 03:10:51.540830 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-kube-api-access-nqp8b" (OuterVolumeSpecName: "kube-api-access-nqp8b") pod "56c0f90d-dd66-4bfd-9db2-52fcefd3919d" (UID: "56c0f90d-dd66-4bfd-9db2-52fcefd3919d"). InnerVolumeSpecName "kube-api-access-nqp8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:10:51 crc kubenswrapper[4775]: I1002 03:10:51.576107 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56c0f90d-dd66-4bfd-9db2-52fcefd3919d" (UID: "56c0f90d-dd66-4bfd-9db2-52fcefd3919d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:10:51 crc kubenswrapper[4775]: I1002 03:10:51.580232 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-config" (OuterVolumeSpecName: "config") pod "56c0f90d-dd66-4bfd-9db2-52fcefd3919d" (UID: "56c0f90d-dd66-4bfd-9db2-52fcefd3919d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:10:51 crc kubenswrapper[4775]: I1002 03:10:51.637482 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:10:51 crc kubenswrapper[4775]: I1002 03:10:51.637532 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:10:51 crc kubenswrapper[4775]: I1002 03:10:51.637555 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqp8b\" (UniqueName: \"kubernetes.io/projected/56c0f90d-dd66-4bfd-9db2-52fcefd3919d-kube-api-access-nqp8b\") on node \"crc\" DevicePath \"\"" Oct 02 03:10:51 crc kubenswrapper[4775]: I1002 03:10:51.974554 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-q9nq4" event={"ID":"56c0f90d-dd66-4bfd-9db2-52fcefd3919d","Type":"ContainerDied","Data":"83cd93e1772e99ac52b313eea814da856c626d58ef41919f4304b4ccdc2df7b4"} Oct 02 03:10:51 crc kubenswrapper[4775]: I1002 03:10:51.974598 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83cd93e1772e99ac52b313eea814da856c626d58ef41919f4304b4ccdc2df7b4" Oct 02 03:10:51 crc kubenswrapper[4775]: I1002 03:10:51.974666 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-q9nq4" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.286463 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6547599f55-q7x9m"] Oct 02 03:10:52 crc kubenswrapper[4775]: E1002 03:10:52.286799 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56c0f90d-dd66-4bfd-9db2-52fcefd3919d" containerName="neutron-db-sync" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.286820 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="56c0f90d-dd66-4bfd-9db2-52fcefd3919d" containerName="neutron-db-sync" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.287021 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="56c0f90d-dd66-4bfd-9db2-52fcefd3919d" containerName="neutron-db-sync" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.287901 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.308670 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6547599f55-q7x9m"] Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.338437 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-776c686f69-g7qtq"] Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.339896 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.341351 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-kxqj6" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.341639 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.343941 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-776c686f69-g7qtq"] Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.344982 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.354688 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-ovsdbserver-sb\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.354795 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-ovsdbserver-nb\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.354840 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-dns-svc\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.354945 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9jw6\" (UniqueName: \"kubernetes.io/projected/905df254-0cfa-4226-848e-926f450b2b18-kube-api-access-x9jw6\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.355344 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-config\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.456888 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbn7l\" (UniqueName: \"kubernetes.io/projected/aca10bd9-aa40-47a8-a461-c0bfc1fa4638-kube-api-access-sbn7l\") pod \"neutron-776c686f69-g7qtq\" (UID: \"aca10bd9-aa40-47a8-a461-c0bfc1fa4638\") " pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.456932 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-config\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.456974 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aca10bd9-aa40-47a8-a461-c0bfc1fa4638-httpd-config\") pod \"neutron-776c686f69-g7qtq\" (UID: \"aca10bd9-aa40-47a8-a461-c0bfc1fa4638\") " pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.456991 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aca10bd9-aa40-47a8-a461-c0bfc1fa4638-combined-ca-bundle\") pod \"neutron-776c686f69-g7qtq\" (UID: \"aca10bd9-aa40-47a8-a461-c0bfc1fa4638\") " pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.457013 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aca10bd9-aa40-47a8-a461-c0bfc1fa4638-config\") pod \"neutron-776c686f69-g7qtq\" (UID: \"aca10bd9-aa40-47a8-a461-c0bfc1fa4638\") " pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.457031 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-ovsdbserver-sb\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.457075 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-ovsdbserver-nb\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.457099 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-dns-svc\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.457206 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9jw6\" (UniqueName: \"kubernetes.io/projected/905df254-0cfa-4226-848e-926f450b2b18-kube-api-access-x9jw6\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.457979 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-dns-svc\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.458044 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-ovsdbserver-sb\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.458059 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-ovsdbserver-nb\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.458923 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-config\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.477764 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9jw6\" (UniqueName: \"kubernetes.io/projected/905df254-0cfa-4226-848e-926f450b2b18-kube-api-access-x9jw6\") pod \"dnsmasq-dns-6547599f55-q7x9m\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.558894 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbn7l\" (UniqueName: \"kubernetes.io/projected/aca10bd9-aa40-47a8-a461-c0bfc1fa4638-kube-api-access-sbn7l\") pod \"neutron-776c686f69-g7qtq\" (UID: \"aca10bd9-aa40-47a8-a461-c0bfc1fa4638\") " pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.559244 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aca10bd9-aa40-47a8-a461-c0bfc1fa4638-httpd-config\") pod \"neutron-776c686f69-g7qtq\" (UID: \"aca10bd9-aa40-47a8-a461-c0bfc1fa4638\") " pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.559264 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aca10bd9-aa40-47a8-a461-c0bfc1fa4638-combined-ca-bundle\") pod \"neutron-776c686f69-g7qtq\" (UID: \"aca10bd9-aa40-47a8-a461-c0bfc1fa4638\") " pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.559287 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aca10bd9-aa40-47a8-a461-c0bfc1fa4638-config\") pod \"neutron-776c686f69-g7qtq\" (UID: \"aca10bd9-aa40-47a8-a461-c0bfc1fa4638\") " pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.564531 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aca10bd9-aa40-47a8-a461-c0bfc1fa4638-httpd-config\") pod \"neutron-776c686f69-g7qtq\" (UID: \"aca10bd9-aa40-47a8-a461-c0bfc1fa4638\") " pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.564701 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aca10bd9-aa40-47a8-a461-c0bfc1fa4638-combined-ca-bundle\") pod \"neutron-776c686f69-g7qtq\" (UID: \"aca10bd9-aa40-47a8-a461-c0bfc1fa4638\") " pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.565181 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/aca10bd9-aa40-47a8-a461-c0bfc1fa4638-config\") pod \"neutron-776c686f69-g7qtq\" (UID: \"aca10bd9-aa40-47a8-a461-c0bfc1fa4638\") " pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.575280 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbn7l\" (UniqueName: \"kubernetes.io/projected/aca10bd9-aa40-47a8-a461-c0bfc1fa4638-kube-api-access-sbn7l\") pod \"neutron-776c686f69-g7qtq\" (UID: \"aca10bd9-aa40-47a8-a461-c0bfc1fa4638\") " pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.613833 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:52 crc kubenswrapper[4775]: I1002 03:10:52.653903 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:53 crc kubenswrapper[4775]: I1002 03:10:53.023789 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6547599f55-q7x9m"] Oct 02 03:10:53 crc kubenswrapper[4775]: W1002 03:10:53.027227 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod905df254_0cfa_4226_848e_926f450b2b18.slice/crio-292c497af8476ce67cb26d47af61358c0a6b5eca28b2cf534ec197d0079d7a55 WatchSource:0}: Error finding container 292c497af8476ce67cb26d47af61358c0a6b5eca28b2cf534ec197d0079d7a55: Status 404 returned error can't find the container with id 292c497af8476ce67cb26d47af61358c0a6b5eca28b2cf534ec197d0079d7a55 Oct 02 03:10:53 crc kubenswrapper[4775]: I1002 03:10:53.245115 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-776c686f69-g7qtq"] Oct 02 03:10:53 crc kubenswrapper[4775]: W1002 03:10:53.267772 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaca10bd9_aa40_47a8_a461_c0bfc1fa4638.slice/crio-9d7a116ec6f0e6b683ff88955246db3b1032edbb7be1c2bcb554affa5d94623a WatchSource:0}: Error finding container 9d7a116ec6f0e6b683ff88955246db3b1032edbb7be1c2bcb554affa5d94623a: Status 404 returned error can't find the container with id 9d7a116ec6f0e6b683ff88955246db3b1032edbb7be1c2bcb554affa5d94623a Oct 02 03:10:53 crc kubenswrapper[4775]: I1002 03:10:53.997941 4775 generic.go:334] "Generic (PLEG): container finished" podID="905df254-0cfa-4226-848e-926f450b2b18" containerID="ab5a1bbc3c5f941df91e61ef117563be7348761b32297b73a884cdadccccb2ac" exitCode=0 Oct 02 03:10:53 crc kubenswrapper[4775]: I1002 03:10:53.998114 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6547599f55-q7x9m" event={"ID":"905df254-0cfa-4226-848e-926f450b2b18","Type":"ContainerDied","Data":"ab5a1bbc3c5f941df91e61ef117563be7348761b32297b73a884cdadccccb2ac"} Oct 02 03:10:53 crc kubenswrapper[4775]: I1002 03:10:53.998457 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6547599f55-q7x9m" event={"ID":"905df254-0cfa-4226-848e-926f450b2b18","Type":"ContainerStarted","Data":"292c497af8476ce67cb26d47af61358c0a6b5eca28b2cf534ec197d0079d7a55"} Oct 02 03:10:54 crc kubenswrapper[4775]: I1002 03:10:54.001051 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-776c686f69-g7qtq" event={"ID":"aca10bd9-aa40-47a8-a461-c0bfc1fa4638","Type":"ContainerStarted","Data":"df831f65e98167a24a5bff0125f2b0fa4966fb3e26b5496a16c96199246252d3"} Oct 02 03:10:54 crc kubenswrapper[4775]: I1002 03:10:54.001108 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-776c686f69-g7qtq" event={"ID":"aca10bd9-aa40-47a8-a461-c0bfc1fa4638","Type":"ContainerStarted","Data":"096805c17dec5446e7298de2ff92967df1084c1198a37579437c9f593b4498ce"} Oct 02 03:10:54 crc kubenswrapper[4775]: I1002 03:10:54.001127 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-776c686f69-g7qtq" event={"ID":"aca10bd9-aa40-47a8-a461-c0bfc1fa4638","Type":"ContainerStarted","Data":"9d7a116ec6f0e6b683ff88955246db3b1032edbb7be1c2bcb554affa5d94623a"} Oct 02 03:10:54 crc kubenswrapper[4775]: I1002 03:10:54.001726 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:10:54 crc kubenswrapper[4775]: I1002 03:10:54.061933 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-776c686f69-g7qtq" podStartSLOduration=2.061905881 podStartE2EDuration="2.061905881s" podCreationTimestamp="2025-10-02 03:10:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:10:54.057657401 +0000 UTC m=+5391.224401471" watchObservedRunningTime="2025-10-02 03:10:54.061905881 +0000 UTC m=+5391.228649951" Oct 02 03:10:55 crc kubenswrapper[4775]: I1002 03:10:55.018764 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6547599f55-q7x9m" event={"ID":"905df254-0cfa-4226-848e-926f450b2b18","Type":"ContainerStarted","Data":"29e77c79f11cfd88f4a446e24b4a8d237ebabb48f891d76c26361df5d3cde925"} Oct 02 03:10:55 crc kubenswrapper[4775]: I1002 03:10:55.019797 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:10:55 crc kubenswrapper[4775]: I1002 03:10:55.044909 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6547599f55-q7x9m" podStartSLOduration=3.044883449 podStartE2EDuration="3.044883449s" podCreationTimestamp="2025-10-02 03:10:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:10:55.03683316 +0000 UTC m=+5392.203577210" watchObservedRunningTime="2025-10-02 03:10:55.044883449 +0000 UTC m=+5392.211627529" Oct 02 03:10:57 crc kubenswrapper[4775]: I1002 03:10:57.947711 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-d8lk9"] Oct 02 03:10:57 crc kubenswrapper[4775]: I1002 03:10:57.954427 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:10:57 crc kubenswrapper[4775]: I1002 03:10:57.967469 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d8lk9"] Oct 02 03:10:58 crc kubenswrapper[4775]: I1002 03:10:58.082104 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-utilities\") pod \"redhat-operators-d8lk9\" (UID: \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\") " pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:10:58 crc kubenswrapper[4775]: I1002 03:10:58.082311 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-catalog-content\") pod \"redhat-operators-d8lk9\" (UID: \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\") " pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:10:58 crc kubenswrapper[4775]: I1002 03:10:58.082375 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sknxh\" (UniqueName: \"kubernetes.io/projected/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-kube-api-access-sknxh\") pod \"redhat-operators-d8lk9\" (UID: \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\") " pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:10:58 crc kubenswrapper[4775]: I1002 03:10:58.184218 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-catalog-content\") pod \"redhat-operators-d8lk9\" (UID: \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\") " pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:10:58 crc kubenswrapper[4775]: I1002 03:10:58.184260 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sknxh\" (UniqueName: \"kubernetes.io/projected/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-kube-api-access-sknxh\") pod \"redhat-operators-d8lk9\" (UID: \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\") " pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:10:58 crc kubenswrapper[4775]: I1002 03:10:58.184384 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-utilities\") pod \"redhat-operators-d8lk9\" (UID: \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\") " pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:10:58 crc kubenswrapper[4775]: I1002 03:10:58.185003 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-utilities\") pod \"redhat-operators-d8lk9\" (UID: \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\") " pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:10:58 crc kubenswrapper[4775]: I1002 03:10:58.185055 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-catalog-content\") pod \"redhat-operators-d8lk9\" (UID: \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\") " pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:10:58 crc kubenswrapper[4775]: I1002 03:10:58.208741 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sknxh\" (UniqueName: \"kubernetes.io/projected/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-kube-api-access-sknxh\") pod \"redhat-operators-d8lk9\" (UID: \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\") " pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:10:58 crc kubenswrapper[4775]: I1002 03:10:58.290279 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:10:58 crc kubenswrapper[4775]: W1002 03:10:58.792862 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd696d60_d5b1_4aa3_9d4f_33752b432dc2.slice/crio-8b141752226b18cffbc33f17e807f504369a1403c350f94ee14cda2a1a5959a7 WatchSource:0}: Error finding container 8b141752226b18cffbc33f17e807f504369a1403c350f94ee14cda2a1a5959a7: Status 404 returned error can't find the container with id 8b141752226b18cffbc33f17e807f504369a1403c350f94ee14cda2a1a5959a7 Oct 02 03:10:58 crc kubenswrapper[4775]: I1002 03:10:58.798569 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-d8lk9"] Oct 02 03:10:59 crc kubenswrapper[4775]: I1002 03:10:59.061153 4775 generic.go:334] "Generic (PLEG): container finished" podID="fd696d60-d5b1-4aa3-9d4f-33752b432dc2" containerID="0662f4dfe166fc722a4aaa3287cc8d7d1c28028914aa0fba6b6ee09d831b2f45" exitCode=0 Oct 02 03:10:59 crc kubenswrapper[4775]: I1002 03:10:59.061250 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8lk9" event={"ID":"fd696d60-d5b1-4aa3-9d4f-33752b432dc2","Type":"ContainerDied","Data":"0662f4dfe166fc722a4aaa3287cc8d7d1c28028914aa0fba6b6ee09d831b2f45"} Oct 02 03:10:59 crc kubenswrapper[4775]: I1002 03:10:59.061401 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8lk9" event={"ID":"fd696d60-d5b1-4aa3-9d4f-33752b432dc2","Type":"ContainerStarted","Data":"8b141752226b18cffbc33f17e807f504369a1403c350f94ee14cda2a1a5959a7"} Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.087690 4775 generic.go:334] "Generic (PLEG): container finished" podID="fd696d60-d5b1-4aa3-9d4f-33752b432dc2" containerID="e4746a1954f98d9a5b166a43eb74a9580db49363a778f92bee593bdbe693404a" exitCode=0 Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.087750 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8lk9" event={"ID":"fd696d60-d5b1-4aa3-9d4f-33752b432dc2","Type":"ContainerDied","Data":"e4746a1954f98d9a5b166a43eb74a9580db49363a778f92bee593bdbe693404a"} Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.325438 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k9pmx"] Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.329660 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.352321 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k9pmx"] Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.447493 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czd2t\" (UniqueName: \"kubernetes.io/projected/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-kube-api-access-czd2t\") pod \"certified-operators-k9pmx\" (UID: \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\") " pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.447598 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-catalog-content\") pod \"certified-operators-k9pmx\" (UID: \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\") " pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.447745 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-utilities\") pod \"certified-operators-k9pmx\" (UID: \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\") " pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.551681 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-catalog-content\") pod \"certified-operators-k9pmx\" (UID: \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\") " pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.551769 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-utilities\") pod \"certified-operators-k9pmx\" (UID: \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\") " pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.552025 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czd2t\" (UniqueName: \"kubernetes.io/projected/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-kube-api-access-czd2t\") pod \"certified-operators-k9pmx\" (UID: \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\") " pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.552349 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-utilities\") pod \"certified-operators-k9pmx\" (UID: \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\") " pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.552625 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-catalog-content\") pod \"certified-operators-k9pmx\" (UID: \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\") " pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.604369 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czd2t\" (UniqueName: \"kubernetes.io/projected/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-kube-api-access-czd2t\") pod \"certified-operators-k9pmx\" (UID: \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\") " pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:01 crc kubenswrapper[4775]: I1002 03:11:01.670363 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:02 crc kubenswrapper[4775]: I1002 03:11:02.097206 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8lk9" event={"ID":"fd696d60-d5b1-4aa3-9d4f-33752b432dc2","Type":"ContainerStarted","Data":"988baecae766c3f92d19cb7b1c6dc5c37d8b4e6b72a43353add729f96ea4f3ee"} Oct 02 03:11:02 crc kubenswrapper[4775]: I1002 03:11:02.114050 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-d8lk9" podStartSLOduration=2.620334942 podStartE2EDuration="5.114025739s" podCreationTimestamp="2025-10-02 03:10:57 +0000 UTC" firstStartedPulling="2025-10-02 03:10:59.062741816 +0000 UTC m=+5396.229485856" lastFinishedPulling="2025-10-02 03:11:01.556432573 +0000 UTC m=+5398.723176653" observedRunningTime="2025-10-02 03:11:02.113394353 +0000 UTC m=+5399.280138403" watchObservedRunningTime="2025-10-02 03:11:02.114025739 +0000 UTC m=+5399.280769839" Oct 02 03:11:02 crc kubenswrapper[4775]: W1002 03:11:02.168209 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd8dedf2_9255_47fe_9297_67d78b6b3cb7.slice/crio-12a385501a55dbae3b38d126f7556220aeffdc66edce2f0272a86dbf8113a3da WatchSource:0}: Error finding container 12a385501a55dbae3b38d126f7556220aeffdc66edce2f0272a86dbf8113a3da: Status 404 returned error can't find the container with id 12a385501a55dbae3b38d126f7556220aeffdc66edce2f0272a86dbf8113a3da Oct 02 03:11:02 crc kubenswrapper[4775]: I1002 03:11:02.178912 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k9pmx"] Oct 02 03:11:02 crc kubenswrapper[4775]: I1002 03:11:02.615425 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:11:02 crc kubenswrapper[4775]: I1002 03:11:02.685668 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67476cd545-bdlxs"] Oct 02 03:11:02 crc kubenswrapper[4775]: I1002 03:11:02.686205 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67476cd545-bdlxs" podUID="5125f6d1-9611-4d09-bebe-b4d7b3e99460" containerName="dnsmasq-dns" containerID="cri-o://2329878b563daff78e069e7606b84295cf44f86a97265abb55cd587fac699ae7" gracePeriod=10 Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.106634 4775 generic.go:334] "Generic (PLEG): container finished" podID="5125f6d1-9611-4d09-bebe-b4d7b3e99460" containerID="2329878b563daff78e069e7606b84295cf44f86a97265abb55cd587fac699ae7" exitCode=0 Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.106688 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67476cd545-bdlxs" event={"ID":"5125f6d1-9611-4d09-bebe-b4d7b3e99460","Type":"ContainerDied","Data":"2329878b563daff78e069e7606b84295cf44f86a97265abb55cd587fac699ae7"} Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.106756 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67476cd545-bdlxs" event={"ID":"5125f6d1-9611-4d09-bebe-b4d7b3e99460","Type":"ContainerDied","Data":"108eae153dd20ef6a08aa048f8f2290dd43c178233c815a57504cf55eb345e6b"} Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.106781 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="108eae153dd20ef6a08aa048f8f2290dd43c178233c815a57504cf55eb345e6b" Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.108139 4775 generic.go:334] "Generic (PLEG): container finished" podID="cd8dedf2-9255-47fe-9297-67d78b6b3cb7" containerID="ac6be3cc82bcf6a256329e2c8d835c4b31c4a2c15898135fa593dae7178a1997" exitCode=0 Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.108201 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k9pmx" event={"ID":"cd8dedf2-9255-47fe-9297-67d78b6b3cb7","Type":"ContainerDied","Data":"ac6be3cc82bcf6a256329e2c8d835c4b31c4a2c15898135fa593dae7178a1997"} Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.108243 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k9pmx" event={"ID":"cd8dedf2-9255-47fe-9297-67d78b6b3cb7","Type":"ContainerStarted","Data":"12a385501a55dbae3b38d126f7556220aeffdc66edce2f0272a86dbf8113a3da"} Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.183415 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.293013 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-ovsdbserver-sb\") pod \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.293190 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-ovsdbserver-nb\") pod \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.293463 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbzb4\" (UniqueName: \"kubernetes.io/projected/5125f6d1-9611-4d09-bebe-b4d7b3e99460-kube-api-access-bbzb4\") pod \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.293564 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-dns-svc\") pod \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.293652 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-config\") pod \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\" (UID: \"5125f6d1-9611-4d09-bebe-b4d7b3e99460\") " Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.300777 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5125f6d1-9611-4d09-bebe-b4d7b3e99460-kube-api-access-bbzb4" (OuterVolumeSpecName: "kube-api-access-bbzb4") pod "5125f6d1-9611-4d09-bebe-b4d7b3e99460" (UID: "5125f6d1-9611-4d09-bebe-b4d7b3e99460"). InnerVolumeSpecName "kube-api-access-bbzb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.341341 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-config" (OuterVolumeSpecName: "config") pod "5125f6d1-9611-4d09-bebe-b4d7b3e99460" (UID: "5125f6d1-9611-4d09-bebe-b4d7b3e99460"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.349805 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5125f6d1-9611-4d09-bebe-b4d7b3e99460" (UID: "5125f6d1-9611-4d09-bebe-b4d7b3e99460"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.365986 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5125f6d1-9611-4d09-bebe-b4d7b3e99460" (UID: "5125f6d1-9611-4d09-bebe-b4d7b3e99460"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.367865 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5125f6d1-9611-4d09-bebe-b4d7b3e99460" (UID: "5125f6d1-9611-4d09-bebe-b4d7b3e99460"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.398580 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbzb4\" (UniqueName: \"kubernetes.io/projected/5125f6d1-9611-4d09-bebe-b4d7b3e99460-kube-api-access-bbzb4\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.398616 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.398625 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.398636 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:03 crc kubenswrapper[4775]: I1002 03:11:03.398644 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5125f6d1-9611-4d09-bebe-b4d7b3e99460-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:04 crc kubenswrapper[4775]: I1002 03:11:04.119909 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67476cd545-bdlxs" Oct 02 03:11:04 crc kubenswrapper[4775]: I1002 03:11:04.170052 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67476cd545-bdlxs"] Oct 02 03:11:04 crc kubenswrapper[4775]: I1002 03:11:04.176459 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67476cd545-bdlxs"] Oct 02 03:11:05 crc kubenswrapper[4775]: I1002 03:11:05.134103 4775 generic.go:334] "Generic (PLEG): container finished" podID="cd8dedf2-9255-47fe-9297-67d78b6b3cb7" containerID="cf7f8c349777bcf257506fc9effc67a1ee5bce6e7bda074376c415b21f009aeb" exitCode=0 Oct 02 03:11:05 crc kubenswrapper[4775]: I1002 03:11:05.134213 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k9pmx" event={"ID":"cd8dedf2-9255-47fe-9297-67d78b6b3cb7","Type":"ContainerDied","Data":"cf7f8c349777bcf257506fc9effc67a1ee5bce6e7bda074376c415b21f009aeb"} Oct 02 03:11:05 crc kubenswrapper[4775]: I1002 03:11:05.782909 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5125f6d1-9611-4d09-bebe-b4d7b3e99460" path="/var/lib/kubelet/pods/5125f6d1-9611-4d09-bebe-b4d7b3e99460/volumes" Oct 02 03:11:06 crc kubenswrapper[4775]: I1002 03:11:06.146986 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k9pmx" event={"ID":"cd8dedf2-9255-47fe-9297-67d78b6b3cb7","Type":"ContainerStarted","Data":"3bb8b91b9b0099be1ad0776e96808cd54c77945d978e0b3482667a399c42ae1a"} Oct 02 03:11:06 crc kubenswrapper[4775]: I1002 03:11:06.176082 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k9pmx" podStartSLOduration=2.618145958 podStartE2EDuration="5.176060941s" podCreationTimestamp="2025-10-02 03:11:01 +0000 UTC" firstStartedPulling="2025-10-02 03:11:03.10951198 +0000 UTC m=+5400.276256020" lastFinishedPulling="2025-10-02 03:11:05.667426923 +0000 UTC m=+5402.834171003" observedRunningTime="2025-10-02 03:11:06.173507894 +0000 UTC m=+5403.340251954" watchObservedRunningTime="2025-10-02 03:11:06.176060941 +0000 UTC m=+5403.342804991" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.290893 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.294164 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.368587 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.533344 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-489wk"] Oct 02 03:11:08 crc kubenswrapper[4775]: E1002 03:11:08.533941 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5125f6d1-9611-4d09-bebe-b4d7b3e99460" containerName="dnsmasq-dns" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.533989 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5125f6d1-9611-4d09-bebe-b4d7b3e99460" containerName="dnsmasq-dns" Oct 02 03:11:08 crc kubenswrapper[4775]: E1002 03:11:08.534037 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5125f6d1-9611-4d09-bebe-b4d7b3e99460" containerName="init" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.534049 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5125f6d1-9611-4d09-bebe-b4d7b3e99460" containerName="init" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.534335 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5125f6d1-9611-4d09-bebe-b4d7b3e99460" containerName="dnsmasq-dns" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.536397 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.548909 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-489wk"] Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.711315 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a25a6a36-11a5-4dad-8424-b3d207b0ef83-catalog-content\") pod \"community-operators-489wk\" (UID: \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\") " pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.711371 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a25a6a36-11a5-4dad-8424-b3d207b0ef83-utilities\") pod \"community-operators-489wk\" (UID: \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\") " pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.711395 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdlvm\" (UniqueName: \"kubernetes.io/projected/a25a6a36-11a5-4dad-8424-b3d207b0ef83-kube-api-access-vdlvm\") pod \"community-operators-489wk\" (UID: \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\") " pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.813698 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a25a6a36-11a5-4dad-8424-b3d207b0ef83-catalog-content\") pod \"community-operators-489wk\" (UID: \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\") " pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.813763 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a25a6a36-11a5-4dad-8424-b3d207b0ef83-utilities\") pod \"community-operators-489wk\" (UID: \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\") " pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.813782 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdlvm\" (UniqueName: \"kubernetes.io/projected/a25a6a36-11a5-4dad-8424-b3d207b0ef83-kube-api-access-vdlvm\") pod \"community-operators-489wk\" (UID: \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\") " pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.814487 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a25a6a36-11a5-4dad-8424-b3d207b0ef83-catalog-content\") pod \"community-operators-489wk\" (UID: \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\") " pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.814689 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a25a6a36-11a5-4dad-8424-b3d207b0ef83-utilities\") pod \"community-operators-489wk\" (UID: \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\") " pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.833567 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdlvm\" (UniqueName: \"kubernetes.io/projected/a25a6a36-11a5-4dad-8424-b3d207b0ef83-kube-api-access-vdlvm\") pod \"community-operators-489wk\" (UID: \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\") " pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:08 crc kubenswrapper[4775]: I1002 03:11:08.911577 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:09 crc kubenswrapper[4775]: I1002 03:11:09.208681 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-489wk"] Oct 02 03:11:09 crc kubenswrapper[4775]: I1002 03:11:09.259711 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:11:10 crc kubenswrapper[4775]: I1002 03:11:10.197558 4775 generic.go:334] "Generic (PLEG): container finished" podID="a25a6a36-11a5-4dad-8424-b3d207b0ef83" containerID="87077f2e5cb26481ba65bbcb57d518e705daac18b85aa4278ff15dea3d23266b" exitCode=0 Oct 02 03:11:10 crc kubenswrapper[4775]: I1002 03:11:10.197672 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-489wk" event={"ID":"a25a6a36-11a5-4dad-8424-b3d207b0ef83","Type":"ContainerDied","Data":"87077f2e5cb26481ba65bbcb57d518e705daac18b85aa4278ff15dea3d23266b"} Oct 02 03:11:10 crc kubenswrapper[4775]: I1002 03:11:10.198022 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-489wk" event={"ID":"a25a6a36-11a5-4dad-8424-b3d207b0ef83","Type":"ContainerStarted","Data":"25f6fa3144823b4d8f4f7f49d24deda4fefc3b9ab560e4737033b94bd9501277"} Oct 02 03:11:10 crc kubenswrapper[4775]: I1002 03:11:10.711232 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d8lk9"] Oct 02 03:11:11 crc kubenswrapper[4775]: I1002 03:11:11.212566 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-489wk" event={"ID":"a25a6a36-11a5-4dad-8424-b3d207b0ef83","Type":"ContainerStarted","Data":"53bde58ff63d60315f9d334a9d39b1ca5ed7f7d6597c55e081cbfa8013634145"} Oct 02 03:11:11 crc kubenswrapper[4775]: I1002 03:11:11.671357 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:11 crc kubenswrapper[4775]: I1002 03:11:11.671442 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:11 crc kubenswrapper[4775]: I1002 03:11:11.754353 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:12 crc kubenswrapper[4775]: I1002 03:11:12.228149 4775 generic.go:334] "Generic (PLEG): container finished" podID="a25a6a36-11a5-4dad-8424-b3d207b0ef83" containerID="53bde58ff63d60315f9d334a9d39b1ca5ed7f7d6597c55e081cbfa8013634145" exitCode=0 Oct 02 03:11:12 crc kubenswrapper[4775]: I1002 03:11:12.228292 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-489wk" event={"ID":"a25a6a36-11a5-4dad-8424-b3d207b0ef83","Type":"ContainerDied","Data":"53bde58ff63d60315f9d334a9d39b1ca5ed7f7d6597c55e081cbfa8013634145"} Oct 02 03:11:12 crc kubenswrapper[4775]: I1002 03:11:12.228772 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-d8lk9" podUID="fd696d60-d5b1-4aa3-9d4f-33752b432dc2" containerName="registry-server" containerID="cri-o://988baecae766c3f92d19cb7b1c6dc5c37d8b4e6b72a43353add729f96ea4f3ee" gracePeriod=2 Oct 02 03:11:12 crc kubenswrapper[4775]: I1002 03:11:12.316160 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:12 crc kubenswrapper[4775]: I1002 03:11:12.802216 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:11:12 crc kubenswrapper[4775]: I1002 03:11:12.900081 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-utilities\") pod \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\" (UID: \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\") " Oct 02 03:11:12 crc kubenswrapper[4775]: I1002 03:11:12.900247 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sknxh\" (UniqueName: \"kubernetes.io/projected/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-kube-api-access-sknxh\") pod \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\" (UID: \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\") " Oct 02 03:11:12 crc kubenswrapper[4775]: I1002 03:11:12.900347 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-catalog-content\") pod \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\" (UID: \"fd696d60-d5b1-4aa3-9d4f-33752b432dc2\") " Oct 02 03:11:12 crc kubenswrapper[4775]: I1002 03:11:12.901675 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-utilities" (OuterVolumeSpecName: "utilities") pod "fd696d60-d5b1-4aa3-9d4f-33752b432dc2" (UID: "fd696d60-d5b1-4aa3-9d4f-33752b432dc2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:11:12 crc kubenswrapper[4775]: I1002 03:11:12.909401 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-kube-api-access-sknxh" (OuterVolumeSpecName: "kube-api-access-sknxh") pod "fd696d60-d5b1-4aa3-9d4f-33752b432dc2" (UID: "fd696d60-d5b1-4aa3-9d4f-33752b432dc2"). InnerVolumeSpecName "kube-api-access-sknxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:11:12 crc kubenswrapper[4775]: I1002 03:11:12.972645 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fd696d60-d5b1-4aa3-9d4f-33752b432dc2" (UID: "fd696d60-d5b1-4aa3-9d4f-33752b432dc2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.002913 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sknxh\" (UniqueName: \"kubernetes.io/projected/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-kube-api-access-sknxh\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.002977 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.002990 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fd696d60-d5b1-4aa3-9d4f-33752b432dc2-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.240769 4775 generic.go:334] "Generic (PLEG): container finished" podID="fd696d60-d5b1-4aa3-9d4f-33752b432dc2" containerID="988baecae766c3f92d19cb7b1c6dc5c37d8b4e6b72a43353add729f96ea4f3ee" exitCode=0 Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.240832 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-d8lk9" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.240882 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8lk9" event={"ID":"fd696d60-d5b1-4aa3-9d4f-33752b432dc2","Type":"ContainerDied","Data":"988baecae766c3f92d19cb7b1c6dc5c37d8b4e6b72a43353add729f96ea4f3ee"} Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.240924 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-d8lk9" event={"ID":"fd696d60-d5b1-4aa3-9d4f-33752b432dc2","Type":"ContainerDied","Data":"8b141752226b18cffbc33f17e807f504369a1403c350f94ee14cda2a1a5959a7"} Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.240983 4775 scope.go:117] "RemoveContainer" containerID="988baecae766c3f92d19cb7b1c6dc5c37d8b4e6b72a43353add729f96ea4f3ee" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.245455 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-489wk" event={"ID":"a25a6a36-11a5-4dad-8424-b3d207b0ef83","Type":"ContainerStarted","Data":"dd2fa1a094b47898ce1cd48f48014a619761be90e27e56512f13a8fce9bd2caf"} Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.267760 4775 scope.go:117] "RemoveContainer" containerID="e4746a1954f98d9a5b166a43eb74a9580db49363a778f92bee593bdbe693404a" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.289495 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-489wk" podStartSLOduration=2.85101469 podStartE2EDuration="5.289470828s" podCreationTimestamp="2025-10-02 03:11:08 +0000 UTC" firstStartedPulling="2025-10-02 03:11:10.200409524 +0000 UTC m=+5407.367153604" lastFinishedPulling="2025-10-02 03:11:12.638865692 +0000 UTC m=+5409.805609742" observedRunningTime="2025-10-02 03:11:13.278559725 +0000 UTC m=+5410.445303805" watchObservedRunningTime="2025-10-02 03:11:13.289470828 +0000 UTC m=+5410.456214908" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.305779 4775 scope.go:117] "RemoveContainer" containerID="0662f4dfe166fc722a4aaa3287cc8d7d1c28028914aa0fba6b6ee09d831b2f45" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.308799 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-d8lk9"] Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.320595 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-d8lk9"] Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.360482 4775 scope.go:117] "RemoveContainer" containerID="988baecae766c3f92d19cb7b1c6dc5c37d8b4e6b72a43353add729f96ea4f3ee" Oct 02 03:11:13 crc kubenswrapper[4775]: E1002 03:11:13.361728 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"988baecae766c3f92d19cb7b1c6dc5c37d8b4e6b72a43353add729f96ea4f3ee\": container with ID starting with 988baecae766c3f92d19cb7b1c6dc5c37d8b4e6b72a43353add729f96ea4f3ee not found: ID does not exist" containerID="988baecae766c3f92d19cb7b1c6dc5c37d8b4e6b72a43353add729f96ea4f3ee" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.361795 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"988baecae766c3f92d19cb7b1c6dc5c37d8b4e6b72a43353add729f96ea4f3ee"} err="failed to get container status \"988baecae766c3f92d19cb7b1c6dc5c37d8b4e6b72a43353add729f96ea4f3ee\": rpc error: code = NotFound desc = could not find container \"988baecae766c3f92d19cb7b1c6dc5c37d8b4e6b72a43353add729f96ea4f3ee\": container with ID starting with 988baecae766c3f92d19cb7b1c6dc5c37d8b4e6b72a43353add729f96ea4f3ee not found: ID does not exist" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.361839 4775 scope.go:117] "RemoveContainer" containerID="e4746a1954f98d9a5b166a43eb74a9580db49363a778f92bee593bdbe693404a" Oct 02 03:11:13 crc kubenswrapper[4775]: E1002 03:11:13.362285 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4746a1954f98d9a5b166a43eb74a9580db49363a778f92bee593bdbe693404a\": container with ID starting with e4746a1954f98d9a5b166a43eb74a9580db49363a778f92bee593bdbe693404a not found: ID does not exist" containerID="e4746a1954f98d9a5b166a43eb74a9580db49363a778f92bee593bdbe693404a" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.362318 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4746a1954f98d9a5b166a43eb74a9580db49363a778f92bee593bdbe693404a"} err="failed to get container status \"e4746a1954f98d9a5b166a43eb74a9580db49363a778f92bee593bdbe693404a\": rpc error: code = NotFound desc = could not find container \"e4746a1954f98d9a5b166a43eb74a9580db49363a778f92bee593bdbe693404a\": container with ID starting with e4746a1954f98d9a5b166a43eb74a9580db49363a778f92bee593bdbe693404a not found: ID does not exist" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.362340 4775 scope.go:117] "RemoveContainer" containerID="0662f4dfe166fc722a4aaa3287cc8d7d1c28028914aa0fba6b6ee09d831b2f45" Oct 02 03:11:13 crc kubenswrapper[4775]: E1002 03:11:13.362580 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0662f4dfe166fc722a4aaa3287cc8d7d1c28028914aa0fba6b6ee09d831b2f45\": container with ID starting with 0662f4dfe166fc722a4aaa3287cc8d7d1c28028914aa0fba6b6ee09d831b2f45 not found: ID does not exist" containerID="0662f4dfe166fc722a4aaa3287cc8d7d1c28028914aa0fba6b6ee09d831b2f45" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.362623 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0662f4dfe166fc722a4aaa3287cc8d7d1c28028914aa0fba6b6ee09d831b2f45"} err="failed to get container status \"0662f4dfe166fc722a4aaa3287cc8d7d1c28028914aa0fba6b6ee09d831b2f45\": rpc error: code = NotFound desc = could not find container \"0662f4dfe166fc722a4aaa3287cc8d7d1c28028914aa0fba6b6ee09d831b2f45\": container with ID starting with 0662f4dfe166fc722a4aaa3287cc8d7d1c28028914aa0fba6b6ee09d831b2f45 not found: ID does not exist" Oct 02 03:11:13 crc kubenswrapper[4775]: E1002 03:11:13.563648 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5125f6d1_9611_4d09_bebe_b4d7b3e99460.slice/crio-108eae153dd20ef6a08aa048f8f2290dd43c178233c815a57504cf55eb345e6b\": RecentStats: unable to find data in memory cache]" Oct 02 03:11:13 crc kubenswrapper[4775]: I1002 03:11:13.800679 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd696d60-d5b1-4aa3-9d4f-33752b432dc2" path="/var/lib/kubelet/pods/fd696d60-d5b1-4aa3-9d4f-33752b432dc2/volumes" Oct 02 03:11:14 crc kubenswrapper[4775]: I1002 03:11:14.917771 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k9pmx"] Oct 02 03:11:14 crc kubenswrapper[4775]: I1002 03:11:14.918222 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k9pmx" podUID="cd8dedf2-9255-47fe-9297-67d78b6b3cb7" containerName="registry-server" containerID="cri-o://3bb8b91b9b0099be1ad0776e96808cd54c77945d978e0b3482667a399c42ae1a" gracePeriod=2 Oct 02 03:11:15 crc kubenswrapper[4775]: I1002 03:11:15.279768 4775 generic.go:334] "Generic (PLEG): container finished" podID="cd8dedf2-9255-47fe-9297-67d78b6b3cb7" containerID="3bb8b91b9b0099be1ad0776e96808cd54c77945d978e0b3482667a399c42ae1a" exitCode=0 Oct 02 03:11:15 crc kubenswrapper[4775]: I1002 03:11:15.279896 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k9pmx" event={"ID":"cd8dedf2-9255-47fe-9297-67d78b6b3cb7","Type":"ContainerDied","Data":"3bb8b91b9b0099be1ad0776e96808cd54c77945d978e0b3482667a399c42ae1a"} Oct 02 03:11:15 crc kubenswrapper[4775]: I1002 03:11:15.412322 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:15 crc kubenswrapper[4775]: I1002 03:11:15.565143 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-utilities\") pod \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\" (UID: \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\") " Oct 02 03:11:15 crc kubenswrapper[4775]: I1002 03:11:15.565217 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-catalog-content\") pod \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\" (UID: \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\") " Oct 02 03:11:15 crc kubenswrapper[4775]: I1002 03:11:15.565279 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czd2t\" (UniqueName: \"kubernetes.io/projected/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-kube-api-access-czd2t\") pod \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\" (UID: \"cd8dedf2-9255-47fe-9297-67d78b6b3cb7\") " Oct 02 03:11:15 crc kubenswrapper[4775]: I1002 03:11:15.566174 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-utilities" (OuterVolumeSpecName: "utilities") pod "cd8dedf2-9255-47fe-9297-67d78b6b3cb7" (UID: "cd8dedf2-9255-47fe-9297-67d78b6b3cb7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:11:15 crc kubenswrapper[4775]: I1002 03:11:15.572095 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-kube-api-access-czd2t" (OuterVolumeSpecName: "kube-api-access-czd2t") pod "cd8dedf2-9255-47fe-9297-67d78b6b3cb7" (UID: "cd8dedf2-9255-47fe-9297-67d78b6b3cb7"). InnerVolumeSpecName "kube-api-access-czd2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:11:15 crc kubenswrapper[4775]: I1002 03:11:15.622802 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd8dedf2-9255-47fe-9297-67d78b6b3cb7" (UID: "cd8dedf2-9255-47fe-9297-67d78b6b3cb7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:11:15 crc kubenswrapper[4775]: I1002 03:11:15.667552 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:15 crc kubenswrapper[4775]: I1002 03:11:15.667605 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czd2t\" (UniqueName: \"kubernetes.io/projected/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-kube-api-access-czd2t\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:15 crc kubenswrapper[4775]: I1002 03:11:15.667627 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd8dedf2-9255-47fe-9297-67d78b6b3cb7-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:16 crc kubenswrapper[4775]: I1002 03:11:16.295436 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k9pmx" event={"ID":"cd8dedf2-9255-47fe-9297-67d78b6b3cb7","Type":"ContainerDied","Data":"12a385501a55dbae3b38d126f7556220aeffdc66edce2f0272a86dbf8113a3da"} Oct 02 03:11:16 crc kubenswrapper[4775]: I1002 03:11:16.295494 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k9pmx" Oct 02 03:11:16 crc kubenswrapper[4775]: I1002 03:11:16.295515 4775 scope.go:117] "RemoveContainer" containerID="3bb8b91b9b0099be1ad0776e96808cd54c77945d978e0b3482667a399c42ae1a" Oct 02 03:11:16 crc kubenswrapper[4775]: I1002 03:11:16.335522 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k9pmx"] Oct 02 03:11:16 crc kubenswrapper[4775]: I1002 03:11:16.338310 4775 scope.go:117] "RemoveContainer" containerID="cf7f8c349777bcf257506fc9effc67a1ee5bce6e7bda074376c415b21f009aeb" Oct 02 03:11:16 crc kubenswrapper[4775]: I1002 03:11:16.346728 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k9pmx"] Oct 02 03:11:16 crc kubenswrapper[4775]: I1002 03:11:16.371039 4775 scope.go:117] "RemoveContainer" containerID="ac6be3cc82bcf6a256329e2c8d835c4b31c4a2c15898135fa593dae7178a1997" Oct 02 03:11:17 crc kubenswrapper[4775]: I1002 03:11:17.781377 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd8dedf2-9255-47fe-9297-67d78b6b3cb7" path="/var/lib/kubelet/pods/cd8dedf2-9255-47fe-9297-67d78b6b3cb7/volumes" Oct 02 03:11:18 crc kubenswrapper[4775]: I1002 03:11:18.912829 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:18 crc kubenswrapper[4775]: I1002 03:11:18.912881 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:18 crc kubenswrapper[4775]: I1002 03:11:18.977388 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:19 crc kubenswrapper[4775]: I1002 03:11:19.406488 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:20 crc kubenswrapper[4775]: I1002 03:11:20.312167 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-489wk"] Oct 02 03:11:21 crc kubenswrapper[4775]: I1002 03:11:21.355436 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-489wk" podUID="a25a6a36-11a5-4dad-8424-b3d207b0ef83" containerName="registry-server" containerID="cri-o://dd2fa1a094b47898ce1cd48f48014a619761be90e27e56512f13a8fce9bd2caf" gracePeriod=2 Oct 02 03:11:21 crc kubenswrapper[4775]: I1002 03:11:21.910505 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.092068 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a25a6a36-11a5-4dad-8424-b3d207b0ef83-catalog-content\") pod \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\" (UID: \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\") " Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.092256 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a25a6a36-11a5-4dad-8424-b3d207b0ef83-utilities\") pod \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\" (UID: \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\") " Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.092302 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdlvm\" (UniqueName: \"kubernetes.io/projected/a25a6a36-11a5-4dad-8424-b3d207b0ef83-kube-api-access-vdlvm\") pod \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\" (UID: \"a25a6a36-11a5-4dad-8424-b3d207b0ef83\") " Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.094875 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a25a6a36-11a5-4dad-8424-b3d207b0ef83-utilities" (OuterVolumeSpecName: "utilities") pod "a25a6a36-11a5-4dad-8424-b3d207b0ef83" (UID: "a25a6a36-11a5-4dad-8424-b3d207b0ef83"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.100527 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a25a6a36-11a5-4dad-8424-b3d207b0ef83-kube-api-access-vdlvm" (OuterVolumeSpecName: "kube-api-access-vdlvm") pod "a25a6a36-11a5-4dad-8424-b3d207b0ef83" (UID: "a25a6a36-11a5-4dad-8424-b3d207b0ef83"). InnerVolumeSpecName "kube-api-access-vdlvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.169049 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a25a6a36-11a5-4dad-8424-b3d207b0ef83-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a25a6a36-11a5-4dad-8424-b3d207b0ef83" (UID: "a25a6a36-11a5-4dad-8424-b3d207b0ef83"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.194274 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a25a6a36-11a5-4dad-8424-b3d207b0ef83-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.194308 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a25a6a36-11a5-4dad-8424-b3d207b0ef83-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.194318 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdlvm\" (UniqueName: \"kubernetes.io/projected/a25a6a36-11a5-4dad-8424-b3d207b0ef83-kube-api-access-vdlvm\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.369296 4775 generic.go:334] "Generic (PLEG): container finished" podID="a25a6a36-11a5-4dad-8424-b3d207b0ef83" containerID="dd2fa1a094b47898ce1cd48f48014a619761be90e27e56512f13a8fce9bd2caf" exitCode=0 Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.369339 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-489wk" event={"ID":"a25a6a36-11a5-4dad-8424-b3d207b0ef83","Type":"ContainerDied","Data":"dd2fa1a094b47898ce1cd48f48014a619761be90e27e56512f13a8fce9bd2caf"} Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.369364 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-489wk" event={"ID":"a25a6a36-11a5-4dad-8424-b3d207b0ef83","Type":"ContainerDied","Data":"25f6fa3144823b4d8f4f7f49d24deda4fefc3b9ab560e4737033b94bd9501277"} Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.369381 4775 scope.go:117] "RemoveContainer" containerID="dd2fa1a094b47898ce1cd48f48014a619761be90e27e56512f13a8fce9bd2caf" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.369574 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-489wk" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.403198 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-489wk"] Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.410754 4775 scope.go:117] "RemoveContainer" containerID="53bde58ff63d60315f9d334a9d39b1ca5ed7f7d6597c55e081cbfa8013634145" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.415591 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-489wk"] Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.441735 4775 scope.go:117] "RemoveContainer" containerID="87077f2e5cb26481ba65bbcb57d518e705daac18b85aa4278ff15dea3d23266b" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.475647 4775 scope.go:117] "RemoveContainer" containerID="dd2fa1a094b47898ce1cd48f48014a619761be90e27e56512f13a8fce9bd2caf" Oct 02 03:11:22 crc kubenswrapper[4775]: E1002 03:11:22.476319 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd2fa1a094b47898ce1cd48f48014a619761be90e27e56512f13a8fce9bd2caf\": container with ID starting with dd2fa1a094b47898ce1cd48f48014a619761be90e27e56512f13a8fce9bd2caf not found: ID does not exist" containerID="dd2fa1a094b47898ce1cd48f48014a619761be90e27e56512f13a8fce9bd2caf" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.476374 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd2fa1a094b47898ce1cd48f48014a619761be90e27e56512f13a8fce9bd2caf"} err="failed to get container status \"dd2fa1a094b47898ce1cd48f48014a619761be90e27e56512f13a8fce9bd2caf\": rpc error: code = NotFound desc = could not find container \"dd2fa1a094b47898ce1cd48f48014a619761be90e27e56512f13a8fce9bd2caf\": container with ID starting with dd2fa1a094b47898ce1cd48f48014a619761be90e27e56512f13a8fce9bd2caf not found: ID does not exist" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.476408 4775 scope.go:117] "RemoveContainer" containerID="53bde58ff63d60315f9d334a9d39b1ca5ed7f7d6597c55e081cbfa8013634145" Oct 02 03:11:22 crc kubenswrapper[4775]: E1002 03:11:22.476978 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53bde58ff63d60315f9d334a9d39b1ca5ed7f7d6597c55e081cbfa8013634145\": container with ID starting with 53bde58ff63d60315f9d334a9d39b1ca5ed7f7d6597c55e081cbfa8013634145 not found: ID does not exist" containerID="53bde58ff63d60315f9d334a9d39b1ca5ed7f7d6597c55e081cbfa8013634145" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.477015 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53bde58ff63d60315f9d334a9d39b1ca5ed7f7d6597c55e081cbfa8013634145"} err="failed to get container status \"53bde58ff63d60315f9d334a9d39b1ca5ed7f7d6597c55e081cbfa8013634145\": rpc error: code = NotFound desc = could not find container \"53bde58ff63d60315f9d334a9d39b1ca5ed7f7d6597c55e081cbfa8013634145\": container with ID starting with 53bde58ff63d60315f9d334a9d39b1ca5ed7f7d6597c55e081cbfa8013634145 not found: ID does not exist" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.477046 4775 scope.go:117] "RemoveContainer" containerID="87077f2e5cb26481ba65bbcb57d518e705daac18b85aa4278ff15dea3d23266b" Oct 02 03:11:22 crc kubenswrapper[4775]: E1002 03:11:22.477456 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87077f2e5cb26481ba65bbcb57d518e705daac18b85aa4278ff15dea3d23266b\": container with ID starting with 87077f2e5cb26481ba65bbcb57d518e705daac18b85aa4278ff15dea3d23266b not found: ID does not exist" containerID="87077f2e5cb26481ba65bbcb57d518e705daac18b85aa4278ff15dea3d23266b" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.477486 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87077f2e5cb26481ba65bbcb57d518e705daac18b85aa4278ff15dea3d23266b"} err="failed to get container status \"87077f2e5cb26481ba65bbcb57d518e705daac18b85aa4278ff15dea3d23266b\": rpc error: code = NotFound desc = could not find container \"87077f2e5cb26481ba65bbcb57d518e705daac18b85aa4278ff15dea3d23266b\": container with ID starting with 87077f2e5cb26481ba65bbcb57d518e705daac18b85aa4278ff15dea3d23266b not found: ID does not exist" Oct 02 03:11:22 crc kubenswrapper[4775]: I1002 03:11:22.672442 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-776c686f69-g7qtq" Oct 02 03:11:23 crc kubenswrapper[4775]: I1002 03:11:23.789223 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a25a6a36-11a5-4dad-8424-b3d207b0ef83" path="/var/lib/kubelet/pods/a25a6a36-11a5-4dad-8424-b3d207b0ef83/volumes" Oct 02 03:11:23 crc kubenswrapper[4775]: E1002 03:11:23.856538 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5125f6d1_9611_4d09_bebe_b4d7b3e99460.slice/crio-108eae153dd20ef6a08aa048f8f2290dd43c178233c815a57504cf55eb345e6b\": RecentStats: unable to find data in memory cache]" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.792523 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-8n4dw"] Oct 02 03:11:30 crc kubenswrapper[4775]: E1002 03:11:30.793381 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd696d60-d5b1-4aa3-9d4f-33752b432dc2" containerName="extract-content" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.793393 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd696d60-d5b1-4aa3-9d4f-33752b432dc2" containerName="extract-content" Oct 02 03:11:30 crc kubenswrapper[4775]: E1002 03:11:30.793402 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd8dedf2-9255-47fe-9297-67d78b6b3cb7" containerName="extract-content" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.793408 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd8dedf2-9255-47fe-9297-67d78b6b3cb7" containerName="extract-content" Oct 02 03:11:30 crc kubenswrapper[4775]: E1002 03:11:30.793426 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd8dedf2-9255-47fe-9297-67d78b6b3cb7" containerName="extract-utilities" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.793432 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd8dedf2-9255-47fe-9297-67d78b6b3cb7" containerName="extract-utilities" Oct 02 03:11:30 crc kubenswrapper[4775]: E1002 03:11:30.793442 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd696d60-d5b1-4aa3-9d4f-33752b432dc2" containerName="registry-server" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.793447 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd696d60-d5b1-4aa3-9d4f-33752b432dc2" containerName="registry-server" Oct 02 03:11:30 crc kubenswrapper[4775]: E1002 03:11:30.793460 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a25a6a36-11a5-4dad-8424-b3d207b0ef83" containerName="extract-utilities" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.793465 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a25a6a36-11a5-4dad-8424-b3d207b0ef83" containerName="extract-utilities" Oct 02 03:11:30 crc kubenswrapper[4775]: E1002 03:11:30.793480 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a25a6a36-11a5-4dad-8424-b3d207b0ef83" containerName="extract-content" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.793486 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a25a6a36-11a5-4dad-8424-b3d207b0ef83" containerName="extract-content" Oct 02 03:11:30 crc kubenswrapper[4775]: E1002 03:11:30.793496 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a25a6a36-11a5-4dad-8424-b3d207b0ef83" containerName="registry-server" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.793509 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a25a6a36-11a5-4dad-8424-b3d207b0ef83" containerName="registry-server" Oct 02 03:11:30 crc kubenswrapper[4775]: E1002 03:11:30.793518 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd8dedf2-9255-47fe-9297-67d78b6b3cb7" containerName="registry-server" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.793524 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd8dedf2-9255-47fe-9297-67d78b6b3cb7" containerName="registry-server" Oct 02 03:11:30 crc kubenswrapper[4775]: E1002 03:11:30.793533 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd696d60-d5b1-4aa3-9d4f-33752b432dc2" containerName="extract-utilities" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.793539 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd696d60-d5b1-4aa3-9d4f-33752b432dc2" containerName="extract-utilities" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.793701 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd696d60-d5b1-4aa3-9d4f-33752b432dc2" containerName="registry-server" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.793715 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd8dedf2-9255-47fe-9297-67d78b6b3cb7" containerName="registry-server" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.793735 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a25a6a36-11a5-4dad-8424-b3d207b0ef83" containerName="registry-server" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.794276 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8n4dw" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.805261 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8n4dw"] Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.866808 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96r7h\" (UniqueName: \"kubernetes.io/projected/28461030-4574-48b2-bc21-b782267b22fd-kube-api-access-96r7h\") pod \"glance-db-create-8n4dw\" (UID: \"28461030-4574-48b2-bc21-b782267b22fd\") " pod="openstack/glance-db-create-8n4dw" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.968548 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96r7h\" (UniqueName: \"kubernetes.io/projected/28461030-4574-48b2-bc21-b782267b22fd-kube-api-access-96r7h\") pod \"glance-db-create-8n4dw\" (UID: \"28461030-4574-48b2-bc21-b782267b22fd\") " pod="openstack/glance-db-create-8n4dw" Oct 02 03:11:30 crc kubenswrapper[4775]: I1002 03:11:30.998315 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96r7h\" (UniqueName: \"kubernetes.io/projected/28461030-4574-48b2-bc21-b782267b22fd-kube-api-access-96r7h\") pod \"glance-db-create-8n4dw\" (UID: \"28461030-4574-48b2-bc21-b782267b22fd\") " pod="openstack/glance-db-create-8n4dw" Oct 02 03:11:31 crc kubenswrapper[4775]: I1002 03:11:31.114749 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8n4dw" Oct 02 03:11:31 crc kubenswrapper[4775]: I1002 03:11:31.558257 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-8n4dw"] Oct 02 03:11:32 crc kubenswrapper[4775]: I1002 03:11:32.516346 4775 generic.go:334] "Generic (PLEG): container finished" podID="28461030-4574-48b2-bc21-b782267b22fd" containerID="0f868e320aa80c1ea6a82a0589215ec5bbc223f3e0310275dc59bf61e150b317" exitCode=0 Oct 02 03:11:32 crc kubenswrapper[4775]: I1002 03:11:32.516421 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8n4dw" event={"ID":"28461030-4574-48b2-bc21-b782267b22fd","Type":"ContainerDied","Data":"0f868e320aa80c1ea6a82a0589215ec5bbc223f3e0310275dc59bf61e150b317"} Oct 02 03:11:32 crc kubenswrapper[4775]: I1002 03:11:32.516740 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8n4dw" event={"ID":"28461030-4574-48b2-bc21-b782267b22fd","Type":"ContainerStarted","Data":"bf4b128cfa5d0aff289944abef2fb38c290564cb3d54878c5a75c226c29c2fc4"} Oct 02 03:11:33 crc kubenswrapper[4775]: I1002 03:11:33.960461 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8n4dw" Oct 02 03:11:34 crc kubenswrapper[4775]: I1002 03:11:34.028024 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96r7h\" (UniqueName: \"kubernetes.io/projected/28461030-4574-48b2-bc21-b782267b22fd-kube-api-access-96r7h\") pod \"28461030-4574-48b2-bc21-b782267b22fd\" (UID: \"28461030-4574-48b2-bc21-b782267b22fd\") " Oct 02 03:11:34 crc kubenswrapper[4775]: I1002 03:11:34.035280 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28461030-4574-48b2-bc21-b782267b22fd-kube-api-access-96r7h" (OuterVolumeSpecName: "kube-api-access-96r7h") pod "28461030-4574-48b2-bc21-b782267b22fd" (UID: "28461030-4574-48b2-bc21-b782267b22fd"). InnerVolumeSpecName "kube-api-access-96r7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:11:34 crc kubenswrapper[4775]: E1002 03:11:34.092088 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5125f6d1_9611_4d09_bebe_b4d7b3e99460.slice/crio-108eae153dd20ef6a08aa048f8f2290dd43c178233c815a57504cf55eb345e6b\": RecentStats: unable to find data in memory cache]" Oct 02 03:11:34 crc kubenswrapper[4775]: I1002 03:11:34.129758 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96r7h\" (UniqueName: \"kubernetes.io/projected/28461030-4574-48b2-bc21-b782267b22fd-kube-api-access-96r7h\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:34 crc kubenswrapper[4775]: I1002 03:11:34.541298 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-8n4dw" event={"ID":"28461030-4574-48b2-bc21-b782267b22fd","Type":"ContainerDied","Data":"bf4b128cfa5d0aff289944abef2fb38c290564cb3d54878c5a75c226c29c2fc4"} Oct 02 03:11:34 crc kubenswrapper[4775]: I1002 03:11:34.541627 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf4b128cfa5d0aff289944abef2fb38c290564cb3d54878c5a75c226c29c2fc4" Oct 02 03:11:34 crc kubenswrapper[4775]: I1002 03:11:34.541381 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-8n4dw" Oct 02 03:11:37 crc kubenswrapper[4775]: I1002 03:11:37.234020 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:11:37 crc kubenswrapper[4775]: I1002 03:11:37.234790 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:11:40 crc kubenswrapper[4775]: I1002 03:11:40.841817 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-78b4-account-create-qprsk"] Oct 02 03:11:40 crc kubenswrapper[4775]: E1002 03:11:40.843037 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28461030-4574-48b2-bc21-b782267b22fd" containerName="mariadb-database-create" Oct 02 03:11:40 crc kubenswrapper[4775]: I1002 03:11:40.843068 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="28461030-4574-48b2-bc21-b782267b22fd" containerName="mariadb-database-create" Oct 02 03:11:40 crc kubenswrapper[4775]: I1002 03:11:40.843499 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="28461030-4574-48b2-bc21-b782267b22fd" containerName="mariadb-database-create" Oct 02 03:11:40 crc kubenswrapper[4775]: I1002 03:11:40.844609 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-78b4-account-create-qprsk" Oct 02 03:11:40 crc kubenswrapper[4775]: I1002 03:11:40.847716 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 02 03:11:40 crc kubenswrapper[4775]: I1002 03:11:40.854876 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-78b4-account-create-qprsk"] Oct 02 03:11:40 crc kubenswrapper[4775]: I1002 03:11:40.959350 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brkf6\" (UniqueName: \"kubernetes.io/projected/ae0119d1-e4f7-4bcf-b47c-2650f169a4e7-kube-api-access-brkf6\") pod \"glance-78b4-account-create-qprsk\" (UID: \"ae0119d1-e4f7-4bcf-b47c-2650f169a4e7\") " pod="openstack/glance-78b4-account-create-qprsk" Oct 02 03:11:41 crc kubenswrapper[4775]: I1002 03:11:41.061878 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brkf6\" (UniqueName: \"kubernetes.io/projected/ae0119d1-e4f7-4bcf-b47c-2650f169a4e7-kube-api-access-brkf6\") pod \"glance-78b4-account-create-qprsk\" (UID: \"ae0119d1-e4f7-4bcf-b47c-2650f169a4e7\") " pod="openstack/glance-78b4-account-create-qprsk" Oct 02 03:11:41 crc kubenswrapper[4775]: I1002 03:11:41.094613 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brkf6\" (UniqueName: \"kubernetes.io/projected/ae0119d1-e4f7-4bcf-b47c-2650f169a4e7-kube-api-access-brkf6\") pod \"glance-78b4-account-create-qprsk\" (UID: \"ae0119d1-e4f7-4bcf-b47c-2650f169a4e7\") " pod="openstack/glance-78b4-account-create-qprsk" Oct 02 03:11:41 crc kubenswrapper[4775]: I1002 03:11:41.183263 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-78b4-account-create-qprsk" Oct 02 03:11:41 crc kubenswrapper[4775]: I1002 03:11:41.667520 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-78b4-account-create-qprsk"] Oct 02 03:11:41 crc kubenswrapper[4775]: W1002 03:11:41.672442 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae0119d1_e4f7_4bcf_b47c_2650f169a4e7.slice/crio-f17940204a154455b3ee044e52d0a6662c8069e4d3fbf11ad80570a64016890f WatchSource:0}: Error finding container f17940204a154455b3ee044e52d0a6662c8069e4d3fbf11ad80570a64016890f: Status 404 returned error can't find the container with id f17940204a154455b3ee044e52d0a6662c8069e4d3fbf11ad80570a64016890f Oct 02 03:11:42 crc kubenswrapper[4775]: I1002 03:11:42.634815 4775 generic.go:334] "Generic (PLEG): container finished" podID="ae0119d1-e4f7-4bcf-b47c-2650f169a4e7" containerID="f21a152c6724d033fb12f1a28fba8be83d85d2152ccf9e7ec09d12eb09e5ddd2" exitCode=0 Oct 02 03:11:42 crc kubenswrapper[4775]: I1002 03:11:42.635063 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-78b4-account-create-qprsk" event={"ID":"ae0119d1-e4f7-4bcf-b47c-2650f169a4e7","Type":"ContainerDied","Data":"f21a152c6724d033fb12f1a28fba8be83d85d2152ccf9e7ec09d12eb09e5ddd2"} Oct 02 03:11:42 crc kubenswrapper[4775]: I1002 03:11:42.635497 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-78b4-account-create-qprsk" event={"ID":"ae0119d1-e4f7-4bcf-b47c-2650f169a4e7","Type":"ContainerStarted","Data":"f17940204a154455b3ee044e52d0a6662c8069e4d3fbf11ad80570a64016890f"} Oct 02 03:11:44 crc kubenswrapper[4775]: I1002 03:11:44.032683 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-78b4-account-create-qprsk" Oct 02 03:11:44 crc kubenswrapper[4775]: I1002 03:11:44.121692 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brkf6\" (UniqueName: \"kubernetes.io/projected/ae0119d1-e4f7-4bcf-b47c-2650f169a4e7-kube-api-access-brkf6\") pod \"ae0119d1-e4f7-4bcf-b47c-2650f169a4e7\" (UID: \"ae0119d1-e4f7-4bcf-b47c-2650f169a4e7\") " Oct 02 03:11:44 crc kubenswrapper[4775]: I1002 03:11:44.128412 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae0119d1-e4f7-4bcf-b47c-2650f169a4e7-kube-api-access-brkf6" (OuterVolumeSpecName: "kube-api-access-brkf6") pod "ae0119d1-e4f7-4bcf-b47c-2650f169a4e7" (UID: "ae0119d1-e4f7-4bcf-b47c-2650f169a4e7"). InnerVolumeSpecName "kube-api-access-brkf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:11:44 crc kubenswrapper[4775]: I1002 03:11:44.223945 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brkf6\" (UniqueName: \"kubernetes.io/projected/ae0119d1-e4f7-4bcf-b47c-2650f169a4e7-kube-api-access-brkf6\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:44 crc kubenswrapper[4775]: E1002 03:11:44.275726 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5125f6d1_9611_4d09_bebe_b4d7b3e99460.slice/crio-108eae153dd20ef6a08aa048f8f2290dd43c178233c815a57504cf55eb345e6b\": RecentStats: unable to find data in memory cache]" Oct 02 03:11:44 crc kubenswrapper[4775]: I1002 03:11:44.663934 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-78b4-account-create-qprsk" event={"ID":"ae0119d1-e4f7-4bcf-b47c-2650f169a4e7","Type":"ContainerDied","Data":"f17940204a154455b3ee044e52d0a6662c8069e4d3fbf11ad80570a64016890f"} Oct 02 03:11:44 crc kubenswrapper[4775]: I1002 03:11:44.664038 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f17940204a154455b3ee044e52d0a6662c8069e4d3fbf11ad80570a64016890f" Oct 02 03:11:44 crc kubenswrapper[4775]: I1002 03:11:44.664083 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-78b4-account-create-qprsk" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.061515 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-sgm76"] Oct 02 03:11:46 crc kubenswrapper[4775]: E1002 03:11:46.062081 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae0119d1-e4f7-4bcf-b47c-2650f169a4e7" containerName="mariadb-account-create" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.062093 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae0119d1-e4f7-4bcf-b47c-2650f169a4e7" containerName="mariadb-account-create" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.062232 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae0119d1-e4f7-4bcf-b47c-2650f169a4e7" containerName="mariadb-account-create" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.062721 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.066037 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.066373 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ngf6p" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.090707 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-sgm76"] Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.259219 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhhlr\" (UniqueName: \"kubernetes.io/projected/a599a786-6215-46c8-8628-34f2cd3aba4f-kube-api-access-dhhlr\") pod \"glance-db-sync-sgm76\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.259511 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-db-sync-config-data\") pod \"glance-db-sync-sgm76\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.259628 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-combined-ca-bundle\") pod \"glance-db-sync-sgm76\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.259743 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-config-data\") pod \"glance-db-sync-sgm76\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.381099 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhhlr\" (UniqueName: \"kubernetes.io/projected/a599a786-6215-46c8-8628-34f2cd3aba4f-kube-api-access-dhhlr\") pod \"glance-db-sync-sgm76\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.381584 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-db-sync-config-data\") pod \"glance-db-sync-sgm76\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.381841 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-combined-ca-bundle\") pod \"glance-db-sync-sgm76\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.382094 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-config-data\") pod \"glance-db-sync-sgm76\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.388595 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-combined-ca-bundle\") pod \"glance-db-sync-sgm76\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.389469 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-config-data\") pod \"glance-db-sync-sgm76\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.399691 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-db-sync-config-data\") pod \"glance-db-sync-sgm76\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.403917 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhhlr\" (UniqueName: \"kubernetes.io/projected/a599a786-6215-46c8-8628-34f2cd3aba4f-kube-api-access-dhhlr\") pod \"glance-db-sync-sgm76\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:46 crc kubenswrapper[4775]: I1002 03:11:46.691809 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:47 crc kubenswrapper[4775]: I1002 03:11:47.256069 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-sgm76"] Oct 02 03:11:47 crc kubenswrapper[4775]: W1002 03:11:47.258047 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda599a786_6215_46c8_8628_34f2cd3aba4f.slice/crio-3f621fccda7d63e454f22132ab1a14013e4316fb6064e3c85b3cd76f4477e4ac WatchSource:0}: Error finding container 3f621fccda7d63e454f22132ab1a14013e4316fb6064e3c85b3cd76f4477e4ac: Status 404 returned error can't find the container with id 3f621fccda7d63e454f22132ab1a14013e4316fb6064e3c85b3cd76f4477e4ac Oct 02 03:11:47 crc kubenswrapper[4775]: I1002 03:11:47.700052 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sgm76" event={"ID":"a599a786-6215-46c8-8628-34f2cd3aba4f","Type":"ContainerStarted","Data":"3f621fccda7d63e454f22132ab1a14013e4316fb6064e3c85b3cd76f4477e4ac"} Oct 02 03:11:48 crc kubenswrapper[4775]: I1002 03:11:48.710669 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sgm76" event={"ID":"a599a786-6215-46c8-8628-34f2cd3aba4f","Type":"ContainerStarted","Data":"114977fc7487ad08c77de37183e952cd3c0a1f3d5556a334a3ab56b30ef5e73b"} Oct 02 03:11:48 crc kubenswrapper[4775]: I1002 03:11:48.736423 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-sgm76" podStartSLOduration=2.736404154 podStartE2EDuration="2.736404154s" podCreationTimestamp="2025-10-02 03:11:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:11:48.731019985 +0000 UTC m=+5445.897764025" watchObservedRunningTime="2025-10-02 03:11:48.736404154 +0000 UTC m=+5445.903148194" Oct 02 03:11:51 crc kubenswrapper[4775]: I1002 03:11:51.750101 4775 generic.go:334] "Generic (PLEG): container finished" podID="a599a786-6215-46c8-8628-34f2cd3aba4f" containerID="114977fc7487ad08c77de37183e952cd3c0a1f3d5556a334a3ab56b30ef5e73b" exitCode=0 Oct 02 03:11:51 crc kubenswrapper[4775]: I1002 03:11:51.750186 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sgm76" event={"ID":"a599a786-6215-46c8-8628-34f2cd3aba4f","Type":"ContainerDied","Data":"114977fc7487ad08c77de37183e952cd3c0a1f3d5556a334a3ab56b30ef5e73b"} Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.272937 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.414066 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-db-sync-config-data\") pod \"a599a786-6215-46c8-8628-34f2cd3aba4f\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.414161 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhhlr\" (UniqueName: \"kubernetes.io/projected/a599a786-6215-46c8-8628-34f2cd3aba4f-kube-api-access-dhhlr\") pod \"a599a786-6215-46c8-8628-34f2cd3aba4f\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.414248 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-config-data\") pod \"a599a786-6215-46c8-8628-34f2cd3aba4f\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.414427 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-combined-ca-bundle\") pod \"a599a786-6215-46c8-8628-34f2cd3aba4f\" (UID: \"a599a786-6215-46c8-8628-34f2cd3aba4f\") " Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.420672 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a599a786-6215-46c8-8628-34f2cd3aba4f-kube-api-access-dhhlr" (OuterVolumeSpecName: "kube-api-access-dhhlr") pod "a599a786-6215-46c8-8628-34f2cd3aba4f" (UID: "a599a786-6215-46c8-8628-34f2cd3aba4f"). InnerVolumeSpecName "kube-api-access-dhhlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.422295 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "a599a786-6215-46c8-8628-34f2cd3aba4f" (UID: "a599a786-6215-46c8-8628-34f2cd3aba4f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.463894 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a599a786-6215-46c8-8628-34f2cd3aba4f" (UID: "a599a786-6215-46c8-8628-34f2cd3aba4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.488754 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-config-data" (OuterVolumeSpecName: "config-data") pod "a599a786-6215-46c8-8628-34f2cd3aba4f" (UID: "a599a786-6215-46c8-8628-34f2cd3aba4f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.518114 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.518186 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.518203 4775 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/a599a786-6215-46c8-8628-34f2cd3aba4f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.518218 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhhlr\" (UniqueName: \"kubernetes.io/projected/a599a786-6215-46c8-8628-34f2cd3aba4f-kube-api-access-dhhlr\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.781094 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-sgm76" Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.788157 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-sgm76" event={"ID":"a599a786-6215-46c8-8628-34f2cd3aba4f","Type":"ContainerDied","Data":"3f621fccda7d63e454f22132ab1a14013e4316fb6064e3c85b3cd76f4477e4ac"} Oct 02 03:11:53 crc kubenswrapper[4775]: I1002 03:11:53.788230 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f621fccda7d63e454f22132ab1a14013e4316fb6064e3c85b3cd76f4477e4ac" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.256397 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-858b587f45-rzn5s"] Oct 02 03:11:54 crc kubenswrapper[4775]: E1002 03:11:54.257116 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a599a786-6215-46c8-8628-34f2cd3aba4f" containerName="glance-db-sync" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.257138 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a599a786-6215-46c8-8628-34f2cd3aba4f" containerName="glance-db-sync" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.257353 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a599a786-6215-46c8-8628-34f2cd3aba4f" containerName="glance-db-sync" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.258473 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.273135 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-858b587f45-rzn5s"] Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.288085 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.289532 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.292412 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.292632 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.292737 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.294468 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ngf6p" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.321777 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.336819 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-scripts\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.336875 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-ovsdbserver-nb\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.336901 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67b27c3e-44bb-4925-bf91-6e021f019547-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.336920 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-config-data\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.336952 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-662b6\" (UniqueName: \"kubernetes.io/projected/8a35f34a-2e54-4107-a630-523cc5d05dd5-kube-api-access-662b6\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.336990 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c2vx\" (UniqueName: \"kubernetes.io/projected/67b27c3e-44bb-4925-bf91-6e021f019547-kube-api-access-4c2vx\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.337012 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-ovsdbserver-sb\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.337026 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67b27c3e-44bb-4925-bf91-6e021f019547-ceph\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.337140 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-config\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.337214 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.337233 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67b27c3e-44bb-4925-bf91-6e021f019547-logs\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.337336 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-dns-svc\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.373502 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.382953 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.386785 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.390293 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.440531 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-662b6\" (UniqueName: \"kubernetes.io/projected/8a35f34a-2e54-4107-a630-523cc5d05dd5-kube-api-access-662b6\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.440580 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c2vx\" (UniqueName: \"kubernetes.io/projected/67b27c3e-44bb-4925-bf91-6e021f019547-kube-api-access-4c2vx\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.440605 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-ovsdbserver-sb\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.440622 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67b27c3e-44bb-4925-bf91-6e021f019547-ceph\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.440648 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-config\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.440674 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.440690 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67b27c3e-44bb-4925-bf91-6e021f019547-logs\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.440734 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-dns-svc\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.440759 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-scripts\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.440790 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-ovsdbserver-nb\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.440817 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67b27c3e-44bb-4925-bf91-6e021f019547-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.440838 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-config-data\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.444016 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-ovsdbserver-sb\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.446029 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67b27c3e-44bb-4925-bf91-6e021f019547-logs\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.445101 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-ovsdbserver-nb\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.446519 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-dns-svc\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.446682 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-config\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.447205 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67b27c3e-44bb-4925-bf91-6e021f019547-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.450142 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.450420 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67b27c3e-44bb-4925-bf91-6e021f019547-ceph\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.456143 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-config-data\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.457516 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-scripts\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.461642 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c2vx\" (UniqueName: \"kubernetes.io/projected/67b27c3e-44bb-4925-bf91-6e021f019547-kube-api-access-4c2vx\") pod \"glance-default-external-api-0\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.462239 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-662b6\" (UniqueName: \"kubernetes.io/projected/8a35f34a-2e54-4107-a630-523cc5d05dd5-kube-api-access-662b6\") pod \"dnsmasq-dns-858b587f45-rzn5s\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.541838 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c515ceea-8609-4f0e-b1b6-321003d57c5a-logs\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.541885 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c515ceea-8609-4f0e-b1b6-321003d57c5a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.541909 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.541927 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.542047 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64mz8\" (UniqueName: \"kubernetes.io/projected/c515ceea-8609-4f0e-b1b6-321003d57c5a-kube-api-access-64mz8\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.542087 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c515ceea-8609-4f0e-b1b6-321003d57c5a-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.542141 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: E1002 03:11:54.562512 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5125f6d1_9611_4d09_bebe_b4d7b3e99460.slice/crio-108eae153dd20ef6a08aa048f8f2290dd43c178233c815a57504cf55eb345e6b\": RecentStats: unable to find data in memory cache]" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.590557 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.606240 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.644253 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c515ceea-8609-4f0e-b1b6-321003d57c5a-logs\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.644320 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c515ceea-8609-4f0e-b1b6-321003d57c5a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.644345 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.644385 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.644405 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64mz8\" (UniqueName: \"kubernetes.io/projected/c515ceea-8609-4f0e-b1b6-321003d57c5a-kube-api-access-64mz8\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.644423 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c515ceea-8609-4f0e-b1b6-321003d57c5a-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.644478 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.645337 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c515ceea-8609-4f0e-b1b6-321003d57c5a-logs\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.645700 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c515ceea-8609-4f0e-b1b6-321003d57c5a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.651196 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.651688 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.653195 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.658083 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c515ceea-8609-4f0e-b1b6-321003d57c5a-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.665698 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64mz8\" (UniqueName: \"kubernetes.io/projected/c515ceea-8609-4f0e-b1b6-321003d57c5a-kube-api-access-64mz8\") pod \"glance-default-internal-api-0\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.706851 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 03:11:54 crc kubenswrapper[4775]: I1002 03:11:54.996941 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:11:55 crc kubenswrapper[4775]: I1002 03:11:55.078085 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-858b587f45-rzn5s"] Oct 02 03:11:55 crc kubenswrapper[4775]: W1002 03:11:55.081727 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a35f34a_2e54_4107_a630_523cc5d05dd5.slice/crio-4efa6d8f2aeb39a4420cba16b22e5d382fa0177a38c5aec856ceee2fa63b1013 WatchSource:0}: Error finding container 4efa6d8f2aeb39a4420cba16b22e5d382fa0177a38c5aec856ceee2fa63b1013: Status 404 returned error can't find the container with id 4efa6d8f2aeb39a4420cba16b22e5d382fa0177a38c5aec856ceee2fa63b1013 Oct 02 03:11:55 crc kubenswrapper[4775]: I1002 03:11:55.258837 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:11:55 crc kubenswrapper[4775]: W1002 03:11:55.268457 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc515ceea_8609_4f0e_b1b6_321003d57c5a.slice/crio-880106299e5badd82bd928935123761e3370aeaeff74877b91d6135c77e48f95 WatchSource:0}: Error finding container 880106299e5badd82bd928935123761e3370aeaeff74877b91d6135c77e48f95: Status 404 returned error can't find the container with id 880106299e5badd82bd928935123761e3370aeaeff74877b91d6135c77e48f95 Oct 02 03:11:55 crc kubenswrapper[4775]: I1002 03:11:55.451371 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:11:55 crc kubenswrapper[4775]: I1002 03:11:55.821178 4775 generic.go:334] "Generic (PLEG): container finished" podID="8a35f34a-2e54-4107-a630-523cc5d05dd5" containerID="cca83b4c1e79415dd1ea5766a9d77c4836869632bfd78c7bbe84c07ba87a8ee0" exitCode=0 Oct 02 03:11:55 crc kubenswrapper[4775]: I1002 03:11:55.821277 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858b587f45-rzn5s" event={"ID":"8a35f34a-2e54-4107-a630-523cc5d05dd5","Type":"ContainerDied","Data":"cca83b4c1e79415dd1ea5766a9d77c4836869632bfd78c7bbe84c07ba87a8ee0"} Oct 02 03:11:55 crc kubenswrapper[4775]: I1002 03:11:55.821520 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858b587f45-rzn5s" event={"ID":"8a35f34a-2e54-4107-a630-523cc5d05dd5","Type":"ContainerStarted","Data":"4efa6d8f2aeb39a4420cba16b22e5d382fa0177a38c5aec856ceee2fa63b1013"} Oct 02 03:11:55 crc kubenswrapper[4775]: I1002 03:11:55.824618 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67b27c3e-44bb-4925-bf91-6e021f019547","Type":"ContainerStarted","Data":"be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c"} Oct 02 03:11:55 crc kubenswrapper[4775]: I1002 03:11:55.824658 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67b27c3e-44bb-4925-bf91-6e021f019547","Type":"ContainerStarted","Data":"067010c742c9e1d0f04fa89da5d863f0dd78b9b733d2af5b4897b0f60ef05238"} Oct 02 03:11:55 crc kubenswrapper[4775]: I1002 03:11:55.826587 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c515ceea-8609-4f0e-b1b6-321003d57c5a","Type":"ContainerStarted","Data":"880106299e5badd82bd928935123761e3370aeaeff74877b91d6135c77e48f95"} Oct 02 03:11:56 crc kubenswrapper[4775]: I1002 03:11:56.843210 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c515ceea-8609-4f0e-b1b6-321003d57c5a","Type":"ContainerStarted","Data":"073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf"} Oct 02 03:11:56 crc kubenswrapper[4775]: I1002 03:11:56.843555 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c515ceea-8609-4f0e-b1b6-321003d57c5a","Type":"ContainerStarted","Data":"911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b"} Oct 02 03:11:56 crc kubenswrapper[4775]: I1002 03:11:56.848732 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858b587f45-rzn5s" event={"ID":"8a35f34a-2e54-4107-a630-523cc5d05dd5","Type":"ContainerStarted","Data":"275edac11af78849528ab6096409da999cc96af821ec8254da67b07da64a1845"} Oct 02 03:11:56 crc kubenswrapper[4775]: I1002 03:11:56.848997 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:11:56 crc kubenswrapper[4775]: I1002 03:11:56.851747 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67b27c3e-44bb-4925-bf91-6e021f019547","Type":"ContainerStarted","Data":"b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19"} Oct 02 03:11:56 crc kubenswrapper[4775]: I1002 03:11:56.852020 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="67b27c3e-44bb-4925-bf91-6e021f019547" containerName="glance-log" containerID="cri-o://be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c" gracePeriod=30 Oct 02 03:11:56 crc kubenswrapper[4775]: I1002 03:11:56.852188 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="67b27c3e-44bb-4925-bf91-6e021f019547" containerName="glance-httpd" containerID="cri-o://b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19" gracePeriod=30 Oct 02 03:11:56 crc kubenswrapper[4775]: I1002 03:11:56.865990 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.865932297 podStartE2EDuration="2.865932297s" podCreationTimestamp="2025-10-02 03:11:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:11:56.863108024 +0000 UTC m=+5454.029852074" watchObservedRunningTime="2025-10-02 03:11:56.865932297 +0000 UTC m=+5454.032676367" Oct 02 03:11:56 crc kubenswrapper[4775]: I1002 03:11:56.893670 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-858b587f45-rzn5s" podStartSLOduration=2.893651715 podStartE2EDuration="2.893651715s" podCreationTimestamp="2025-10-02 03:11:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:11:56.88921747 +0000 UTC m=+5454.055961560" watchObservedRunningTime="2025-10-02 03:11:56.893651715 +0000 UTC m=+5454.060395745" Oct 02 03:11:56 crc kubenswrapper[4775]: I1002 03:11:56.922236 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.922217405 podStartE2EDuration="2.922217405s" podCreationTimestamp="2025-10-02 03:11:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:11:56.914111115 +0000 UTC m=+5454.080855225" watchObservedRunningTime="2025-10-02 03:11:56.922217405 +0000 UTC m=+5454.088961445" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.570600 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.694352 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67b27c3e-44bb-4925-bf91-6e021f019547-logs\") pod \"67b27c3e-44bb-4925-bf91-6e021f019547\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.694418 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-scripts\") pod \"67b27c3e-44bb-4925-bf91-6e021f019547\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.694474 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-config-data\") pod \"67b27c3e-44bb-4925-bf91-6e021f019547\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.694536 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4c2vx\" (UniqueName: \"kubernetes.io/projected/67b27c3e-44bb-4925-bf91-6e021f019547-kube-api-access-4c2vx\") pod \"67b27c3e-44bb-4925-bf91-6e021f019547\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.694554 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-combined-ca-bundle\") pod \"67b27c3e-44bb-4925-bf91-6e021f019547\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.694588 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67b27c3e-44bb-4925-bf91-6e021f019547-httpd-run\") pod \"67b27c3e-44bb-4925-bf91-6e021f019547\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.694671 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67b27c3e-44bb-4925-bf91-6e021f019547-ceph\") pod \"67b27c3e-44bb-4925-bf91-6e021f019547\" (UID: \"67b27c3e-44bb-4925-bf91-6e021f019547\") " Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.695645 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67b27c3e-44bb-4925-bf91-6e021f019547-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "67b27c3e-44bb-4925-bf91-6e021f019547" (UID: "67b27c3e-44bb-4925-bf91-6e021f019547"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.695693 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67b27c3e-44bb-4925-bf91-6e021f019547-logs" (OuterVolumeSpecName: "logs") pod "67b27c3e-44bb-4925-bf91-6e021f019547" (UID: "67b27c3e-44bb-4925-bf91-6e021f019547"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.700387 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67b27c3e-44bb-4925-bf91-6e021f019547-kube-api-access-4c2vx" (OuterVolumeSpecName: "kube-api-access-4c2vx") pod "67b27c3e-44bb-4925-bf91-6e021f019547" (UID: "67b27c3e-44bb-4925-bf91-6e021f019547"). InnerVolumeSpecName "kube-api-access-4c2vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.703017 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67b27c3e-44bb-4925-bf91-6e021f019547-ceph" (OuterVolumeSpecName: "ceph") pod "67b27c3e-44bb-4925-bf91-6e021f019547" (UID: "67b27c3e-44bb-4925-bf91-6e021f019547"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.703421 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-scripts" (OuterVolumeSpecName: "scripts") pod "67b27c3e-44bb-4925-bf91-6e021f019547" (UID: "67b27c3e-44bb-4925-bf91-6e021f019547"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.757588 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-config-data" (OuterVolumeSpecName: "config-data") pod "67b27c3e-44bb-4925-bf91-6e021f019547" (UID: "67b27c3e-44bb-4925-bf91-6e021f019547"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.762977 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.771148 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67b27c3e-44bb-4925-bf91-6e021f019547" (UID: "67b27c3e-44bb-4925-bf91-6e021f019547"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.798947 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/67b27c3e-44bb-4925-bf91-6e021f019547-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.799007 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67b27c3e-44bb-4925-bf91-6e021f019547-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.799016 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.799024 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.799034 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4c2vx\" (UniqueName: \"kubernetes.io/projected/67b27c3e-44bb-4925-bf91-6e021f019547-kube-api-access-4c2vx\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.799046 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b27c3e-44bb-4925-bf91-6e021f019547-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.799054 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/67b27c3e-44bb-4925-bf91-6e021f019547-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.860664 4775 generic.go:334] "Generic (PLEG): container finished" podID="67b27c3e-44bb-4925-bf91-6e021f019547" containerID="b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19" exitCode=0 Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.860696 4775 generic.go:334] "Generic (PLEG): container finished" podID="67b27c3e-44bb-4925-bf91-6e021f019547" containerID="be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c" exitCode=143 Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.861352 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.861770 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67b27c3e-44bb-4925-bf91-6e021f019547","Type":"ContainerDied","Data":"b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19"} Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.861794 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67b27c3e-44bb-4925-bf91-6e021f019547","Type":"ContainerDied","Data":"be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c"} Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.861804 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"67b27c3e-44bb-4925-bf91-6e021f019547","Type":"ContainerDied","Data":"067010c742c9e1d0f04fa89da5d863f0dd78b9b733d2af5b4897b0f60ef05238"} Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.861819 4775 scope.go:117] "RemoveContainer" containerID="b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.916344 4775 scope.go:117] "RemoveContainer" containerID="be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.923789 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.932376 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.937492 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:11:57 crc kubenswrapper[4775]: E1002 03:11:57.937852 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67b27c3e-44bb-4925-bf91-6e021f019547" containerName="glance-log" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.937868 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="67b27c3e-44bb-4925-bf91-6e021f019547" containerName="glance-log" Oct 02 03:11:57 crc kubenswrapper[4775]: E1002 03:11:57.937886 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67b27c3e-44bb-4925-bf91-6e021f019547" containerName="glance-httpd" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.937892 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="67b27c3e-44bb-4925-bf91-6e021f019547" containerName="glance-httpd" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.938094 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="67b27c3e-44bb-4925-bf91-6e021f019547" containerName="glance-httpd" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.938106 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="67b27c3e-44bb-4925-bf91-6e021f019547" containerName="glance-log" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.938990 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.940672 4775 scope.go:117] "RemoveContainer" containerID="b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.942586 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:11:57 crc kubenswrapper[4775]: E1002 03:11:57.945428 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19\": container with ID starting with b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19 not found: ID does not exist" containerID="b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.945647 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19"} err="failed to get container status \"b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19\": rpc error: code = NotFound desc = could not find container \"b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19\": container with ID starting with b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19 not found: ID does not exist" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.945790 4775 scope.go:117] "RemoveContainer" containerID="be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.945962 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 03:11:57 crc kubenswrapper[4775]: E1002 03:11:57.946437 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c\": container with ID starting with be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c not found: ID does not exist" containerID="be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.946471 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c"} err="failed to get container status \"be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c\": rpc error: code = NotFound desc = could not find container \"be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c\": container with ID starting with be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c not found: ID does not exist" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.946495 4775 scope.go:117] "RemoveContainer" containerID="b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.947009 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19"} err="failed to get container status \"b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19\": rpc error: code = NotFound desc = could not find container \"b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19\": container with ID starting with b2378dba25e54fd3be2ad7a57ed3e14e985110fc5031bc36e1ecca4a9da7fd19 not found: ID does not exist" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.947132 4775 scope.go:117] "RemoveContainer" containerID="be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c" Oct 02 03:11:57 crc kubenswrapper[4775]: I1002 03:11:57.947463 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c"} err="failed to get container status \"be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c\": rpc error: code = NotFound desc = could not find container \"be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c\": container with ID starting with be014f69cc61ec99587858904a362585da3de8e8f6b3c00b65766d2a5097452c not found: ID does not exist" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.011845 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/74fd9de0-3d01-420e-bdc7-35223a06a96d-ceph\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.011907 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qkvj\" (UniqueName: \"kubernetes.io/projected/74fd9de0-3d01-420e-bdc7-35223a06a96d-kube-api-access-2qkvj\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.011978 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-config-data\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.011999 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74fd9de0-3d01-420e-bdc7-35223a06a96d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.012019 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.012069 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74fd9de0-3d01-420e-bdc7-35223a06a96d-logs\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.012185 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-scripts\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.113162 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74fd9de0-3d01-420e-bdc7-35223a06a96d-logs\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.113227 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-scripts\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.113278 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/74fd9de0-3d01-420e-bdc7-35223a06a96d-ceph\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.113304 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qkvj\" (UniqueName: \"kubernetes.io/projected/74fd9de0-3d01-420e-bdc7-35223a06a96d-kube-api-access-2qkvj\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.113351 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-config-data\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.113373 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74fd9de0-3d01-420e-bdc7-35223a06a96d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.113392 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.114002 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74fd9de0-3d01-420e-bdc7-35223a06a96d-logs\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.114306 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74fd9de0-3d01-420e-bdc7-35223a06a96d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.118387 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-config-data\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.118411 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.118973 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/74fd9de0-3d01-420e-bdc7-35223a06a96d-ceph\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.119797 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-scripts\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.131660 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qkvj\" (UniqueName: \"kubernetes.io/projected/74fd9de0-3d01-420e-bdc7-35223a06a96d-kube-api-access-2qkvj\") pod \"glance-default-external-api-0\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.257519 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.873670 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c515ceea-8609-4f0e-b1b6-321003d57c5a" containerName="glance-log" containerID="cri-o://073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf" gracePeriod=30 Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.874018 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="c515ceea-8609-4f0e-b1b6-321003d57c5a" containerName="glance-httpd" containerID="cri-o://911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b" gracePeriod=30 Oct 02 03:11:58 crc kubenswrapper[4775]: I1002 03:11:58.904727 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:11:58 crc kubenswrapper[4775]: W1002 03:11:58.921880 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74fd9de0_3d01_420e_bdc7_35223a06a96d.slice/crio-e88874dd460a17bf43c17f31b2a7dd929bbac257764cc69446615aaf9aa4176a WatchSource:0}: Error finding container e88874dd460a17bf43c17f31b2a7dd929bbac257764cc69446615aaf9aa4176a: Status 404 returned error can't find the container with id e88874dd460a17bf43c17f31b2a7dd929bbac257764cc69446615aaf9aa4176a Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.456387 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.646239 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-combined-ca-bundle\") pod \"c515ceea-8609-4f0e-b1b6-321003d57c5a\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.646685 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c515ceea-8609-4f0e-b1b6-321003d57c5a-ceph\") pod \"c515ceea-8609-4f0e-b1b6-321003d57c5a\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.646915 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64mz8\" (UniqueName: \"kubernetes.io/projected/c515ceea-8609-4f0e-b1b6-321003d57c5a-kube-api-access-64mz8\") pod \"c515ceea-8609-4f0e-b1b6-321003d57c5a\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.646974 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-config-data\") pod \"c515ceea-8609-4f0e-b1b6-321003d57c5a\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.647040 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c515ceea-8609-4f0e-b1b6-321003d57c5a-httpd-run\") pod \"c515ceea-8609-4f0e-b1b6-321003d57c5a\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.647102 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-scripts\") pod \"c515ceea-8609-4f0e-b1b6-321003d57c5a\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.647132 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c515ceea-8609-4f0e-b1b6-321003d57c5a-logs\") pod \"c515ceea-8609-4f0e-b1b6-321003d57c5a\" (UID: \"c515ceea-8609-4f0e-b1b6-321003d57c5a\") " Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.647763 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c515ceea-8609-4f0e-b1b6-321003d57c5a-logs" (OuterVolumeSpecName: "logs") pod "c515ceea-8609-4f0e-b1b6-321003d57c5a" (UID: "c515ceea-8609-4f0e-b1b6-321003d57c5a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.648150 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c515ceea-8609-4f0e-b1b6-321003d57c5a-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.648186 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c515ceea-8609-4f0e-b1b6-321003d57c5a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c515ceea-8609-4f0e-b1b6-321003d57c5a" (UID: "c515ceea-8609-4f0e-b1b6-321003d57c5a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.653183 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c515ceea-8609-4f0e-b1b6-321003d57c5a-kube-api-access-64mz8" (OuterVolumeSpecName: "kube-api-access-64mz8") pod "c515ceea-8609-4f0e-b1b6-321003d57c5a" (UID: "c515ceea-8609-4f0e-b1b6-321003d57c5a"). InnerVolumeSpecName "kube-api-access-64mz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.655161 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c515ceea-8609-4f0e-b1b6-321003d57c5a-ceph" (OuterVolumeSpecName: "ceph") pod "c515ceea-8609-4f0e-b1b6-321003d57c5a" (UID: "c515ceea-8609-4f0e-b1b6-321003d57c5a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.679150 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-scripts" (OuterVolumeSpecName: "scripts") pod "c515ceea-8609-4f0e-b1b6-321003d57c5a" (UID: "c515ceea-8609-4f0e-b1b6-321003d57c5a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.697205 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c515ceea-8609-4f0e-b1b6-321003d57c5a" (UID: "c515ceea-8609-4f0e-b1b6-321003d57c5a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.700205 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-config-data" (OuterVolumeSpecName: "config-data") pod "c515ceea-8609-4f0e-b1b6-321003d57c5a" (UID: "c515ceea-8609-4f0e-b1b6-321003d57c5a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.750078 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.750117 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c515ceea-8609-4f0e-b1b6-321003d57c5a-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.750130 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64mz8\" (UniqueName: \"kubernetes.io/projected/c515ceea-8609-4f0e-b1b6-321003d57c5a-kube-api-access-64mz8\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.750142 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.750153 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c515ceea-8609-4f0e-b1b6-321003d57c5a-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.750162 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c515ceea-8609-4f0e-b1b6-321003d57c5a-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.775243 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67b27c3e-44bb-4925-bf91-6e021f019547" path="/var/lib/kubelet/pods/67b27c3e-44bb-4925-bf91-6e021f019547/volumes" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.886452 4775 generic.go:334] "Generic (PLEG): container finished" podID="c515ceea-8609-4f0e-b1b6-321003d57c5a" containerID="911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b" exitCode=0 Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.886482 4775 generic.go:334] "Generic (PLEG): container finished" podID="c515ceea-8609-4f0e-b1b6-321003d57c5a" containerID="073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf" exitCode=143 Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.886504 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.886525 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c515ceea-8609-4f0e-b1b6-321003d57c5a","Type":"ContainerDied","Data":"911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b"} Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.886573 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c515ceea-8609-4f0e-b1b6-321003d57c5a","Type":"ContainerDied","Data":"073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf"} Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.886587 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c515ceea-8609-4f0e-b1b6-321003d57c5a","Type":"ContainerDied","Data":"880106299e5badd82bd928935123761e3370aeaeff74877b91d6135c77e48f95"} Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.886605 4775 scope.go:117] "RemoveContainer" containerID="911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.894148 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74fd9de0-3d01-420e-bdc7-35223a06a96d","Type":"ContainerStarted","Data":"f1fc960cdde4c98e2edafe7b23cba75f4c05e9d0a8ce499568d09c54eb0a2aeb"} Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.894183 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74fd9de0-3d01-420e-bdc7-35223a06a96d","Type":"ContainerStarted","Data":"e88874dd460a17bf43c17f31b2a7dd929bbac257764cc69446615aaf9aa4176a"} Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.904273 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.914764 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.945578 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:11:59 crc kubenswrapper[4775]: E1002 03:11:59.946066 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c515ceea-8609-4f0e-b1b6-321003d57c5a" containerName="glance-httpd" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.946090 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c515ceea-8609-4f0e-b1b6-321003d57c5a" containerName="glance-httpd" Oct 02 03:11:59 crc kubenswrapper[4775]: E1002 03:11:59.946118 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c515ceea-8609-4f0e-b1b6-321003d57c5a" containerName="glance-log" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.946127 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c515ceea-8609-4f0e-b1b6-321003d57c5a" containerName="glance-log" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.946335 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c515ceea-8609-4f0e-b1b6-321003d57c5a" containerName="glance-log" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.946362 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c515ceea-8609-4f0e-b1b6-321003d57c5a" containerName="glance-httpd" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.947496 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.949069 4775 scope.go:117] "RemoveContainer" containerID="073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.952644 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.954578 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.955052 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a00190d2-bf43-401c-b86c-e52bcbfefcd9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.955096 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a00190d2-bf43-401c-b86c-e52bcbfefcd9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.955116 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a00190d2-bf43-401c-b86c-e52bcbfefcd9-logs\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.955135 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.955154 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jm6b\" (UniqueName: \"kubernetes.io/projected/a00190d2-bf43-401c-b86c-e52bcbfefcd9-kube-api-access-9jm6b\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.955235 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.955272 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.987732 4775 scope.go:117] "RemoveContainer" containerID="911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b" Oct 02 03:11:59 crc kubenswrapper[4775]: E1002 03:11:59.988196 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b\": container with ID starting with 911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b not found: ID does not exist" containerID="911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.988250 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b"} err="failed to get container status \"911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b\": rpc error: code = NotFound desc = could not find container \"911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b\": container with ID starting with 911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b not found: ID does not exist" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.988281 4775 scope.go:117] "RemoveContainer" containerID="073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf" Oct 02 03:11:59 crc kubenswrapper[4775]: E1002 03:11:59.988609 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf\": container with ID starting with 073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf not found: ID does not exist" containerID="073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.988668 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf"} err="failed to get container status \"073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf\": rpc error: code = NotFound desc = could not find container \"073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf\": container with ID starting with 073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf not found: ID does not exist" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.988690 4775 scope.go:117] "RemoveContainer" containerID="911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.989183 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b"} err="failed to get container status \"911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b\": rpc error: code = NotFound desc = could not find container \"911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b\": container with ID starting with 911cbc9eafbd028f7794f4c0f143fc0bb435045c5d8a08d5ed6067ade98e0a2b not found: ID does not exist" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.989209 4775 scope.go:117] "RemoveContainer" containerID="073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf" Oct 02 03:11:59 crc kubenswrapper[4775]: I1002 03:11:59.989468 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf"} err="failed to get container status \"073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf\": rpc error: code = NotFound desc = could not find container \"073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf\": container with ID starting with 073e99b990c2eb85aff35a66b4173cb8a278b013e4b6c38241be5f76fbd09eaf not found: ID does not exist" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.056744 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a00190d2-bf43-401c-b86c-e52bcbfefcd9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.056829 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a00190d2-bf43-401c-b86c-e52bcbfefcd9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.056869 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a00190d2-bf43-401c-b86c-e52bcbfefcd9-logs\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.056904 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.056946 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jm6b\" (UniqueName: \"kubernetes.io/projected/a00190d2-bf43-401c-b86c-e52bcbfefcd9-kube-api-access-9jm6b\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.057059 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.057119 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.058443 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a00190d2-bf43-401c-b86c-e52bcbfefcd9-logs\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.058606 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a00190d2-bf43-401c-b86c-e52bcbfefcd9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.063972 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a00190d2-bf43-401c-b86c-e52bcbfefcd9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.064288 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.064190 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.065597 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.077399 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jm6b\" (UniqueName: \"kubernetes.io/projected/a00190d2-bf43-401c-b86c-e52bcbfefcd9-kube-api-access-9jm6b\") pod \"glance-default-internal-api-0\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.265192 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.867775 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:12:00 crc kubenswrapper[4775]: W1002 03:12:00.879276 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda00190d2_bf43_401c_b86c_e52bcbfefcd9.slice/crio-348e3915a85af1aa181aac83792d404f0f198309b9fde5cdf586f1e1913d7498 WatchSource:0}: Error finding container 348e3915a85af1aa181aac83792d404f0f198309b9fde5cdf586f1e1913d7498: Status 404 returned error can't find the container with id 348e3915a85af1aa181aac83792d404f0f198309b9fde5cdf586f1e1913d7498 Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.906473 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74fd9de0-3d01-420e-bdc7-35223a06a96d","Type":"ContainerStarted","Data":"96f0f4427f4e8c46c7165cd632656c506d1272458cf0f0318c2ab2d30d34c098"} Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.908241 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a00190d2-bf43-401c-b86c-e52bcbfefcd9","Type":"ContainerStarted","Data":"348e3915a85af1aa181aac83792d404f0f198309b9fde5cdf586f1e1913d7498"} Oct 02 03:12:00 crc kubenswrapper[4775]: I1002 03:12:00.944586 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.944563569 podStartE2EDuration="3.944563569s" podCreationTimestamp="2025-10-02 03:11:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:12:00.934323933 +0000 UTC m=+5458.101068013" watchObservedRunningTime="2025-10-02 03:12:00.944563569 +0000 UTC m=+5458.111307619" Oct 02 03:12:01 crc kubenswrapper[4775]: I1002 03:12:01.780104 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c515ceea-8609-4f0e-b1b6-321003d57c5a" path="/var/lib/kubelet/pods/c515ceea-8609-4f0e-b1b6-321003d57c5a/volumes" Oct 02 03:12:01 crc kubenswrapper[4775]: I1002 03:12:01.922788 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a00190d2-bf43-401c-b86c-e52bcbfefcd9","Type":"ContainerStarted","Data":"04ee5d869350eecf5f6e6c74e5d7903373815ccb3da1ac919cf2e49347f2eb65"} Oct 02 03:12:02 crc kubenswrapper[4775]: I1002 03:12:02.936247 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a00190d2-bf43-401c-b86c-e52bcbfefcd9","Type":"ContainerStarted","Data":"a3547ba71c1286ae422447403c12040e5231d0ada533c72ef9f64a4e03f8f587"} Oct 02 03:12:02 crc kubenswrapper[4775]: I1002 03:12:02.968935 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.968909907 podStartE2EDuration="3.968909907s" podCreationTimestamp="2025-10-02 03:11:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:12:02.96668577 +0000 UTC m=+5460.133429810" watchObservedRunningTime="2025-10-02 03:12:02.968909907 +0000 UTC m=+5460.135653977" Oct 02 03:12:04 crc kubenswrapper[4775]: I1002 03:12:04.592233 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:12:04 crc kubenswrapper[4775]: I1002 03:12:04.703078 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6547599f55-q7x9m"] Oct 02 03:12:04 crc kubenswrapper[4775]: I1002 03:12:04.703389 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6547599f55-q7x9m" podUID="905df254-0cfa-4226-848e-926f450b2b18" containerName="dnsmasq-dns" containerID="cri-o://29e77c79f11cfd88f4a446e24b4a8d237ebabb48f891d76c26361df5d3cde925" gracePeriod=10 Oct 02 03:12:04 crc kubenswrapper[4775]: E1002 03:12:04.863608 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod905df254_0cfa_4226_848e_926f450b2b18.slice/crio-29e77c79f11cfd88f4a446e24b4a8d237ebabb48f891d76c26361df5d3cde925.scope\": RecentStats: unable to find data in memory cache]" Oct 02 03:12:04 crc kubenswrapper[4775]: I1002 03:12:04.955429 4775 generic.go:334] "Generic (PLEG): container finished" podID="905df254-0cfa-4226-848e-926f450b2b18" containerID="29e77c79f11cfd88f4a446e24b4a8d237ebabb48f891d76c26361df5d3cde925" exitCode=0 Oct 02 03:12:04 crc kubenswrapper[4775]: I1002 03:12:04.955670 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6547599f55-q7x9m" event={"ID":"905df254-0cfa-4226-848e-926f450b2b18","Type":"ContainerDied","Data":"29e77c79f11cfd88f4a446e24b4a8d237ebabb48f891d76c26361df5d3cde925"} Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.191843 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.317701 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-ovsdbserver-sb\") pod \"905df254-0cfa-4226-848e-926f450b2b18\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.317787 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9jw6\" (UniqueName: \"kubernetes.io/projected/905df254-0cfa-4226-848e-926f450b2b18-kube-api-access-x9jw6\") pod \"905df254-0cfa-4226-848e-926f450b2b18\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.317849 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-dns-svc\") pod \"905df254-0cfa-4226-848e-926f450b2b18\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.318048 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-ovsdbserver-nb\") pod \"905df254-0cfa-4226-848e-926f450b2b18\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.318125 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-config\") pod \"905df254-0cfa-4226-848e-926f450b2b18\" (UID: \"905df254-0cfa-4226-848e-926f450b2b18\") " Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.324830 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/905df254-0cfa-4226-848e-926f450b2b18-kube-api-access-x9jw6" (OuterVolumeSpecName: "kube-api-access-x9jw6") pod "905df254-0cfa-4226-848e-926f450b2b18" (UID: "905df254-0cfa-4226-848e-926f450b2b18"). InnerVolumeSpecName "kube-api-access-x9jw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.379640 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "905df254-0cfa-4226-848e-926f450b2b18" (UID: "905df254-0cfa-4226-848e-926f450b2b18"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.382146 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "905df254-0cfa-4226-848e-926f450b2b18" (UID: "905df254-0cfa-4226-848e-926f450b2b18"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.390312 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "905df254-0cfa-4226-848e-926f450b2b18" (UID: "905df254-0cfa-4226-848e-926f450b2b18"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.394779 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-config" (OuterVolumeSpecName: "config") pod "905df254-0cfa-4226-848e-926f450b2b18" (UID: "905df254-0cfa-4226-848e-926f450b2b18"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.419837 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.419867 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9jw6\" (UniqueName: \"kubernetes.io/projected/905df254-0cfa-4226-848e-926f450b2b18-kube-api-access-x9jw6\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.419878 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.419887 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.419896 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/905df254-0cfa-4226-848e-926f450b2b18-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.973771 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6547599f55-q7x9m" event={"ID":"905df254-0cfa-4226-848e-926f450b2b18","Type":"ContainerDied","Data":"292c497af8476ce67cb26d47af61358c0a6b5eca28b2cf534ec197d0079d7a55"} Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.974040 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6547599f55-q7x9m" Oct 02 03:12:05 crc kubenswrapper[4775]: I1002 03:12:05.974073 4775 scope.go:117] "RemoveContainer" containerID="29e77c79f11cfd88f4a446e24b4a8d237ebabb48f891d76c26361df5d3cde925" Oct 02 03:12:06 crc kubenswrapper[4775]: I1002 03:12:05.999986 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6547599f55-q7x9m"] Oct 02 03:12:06 crc kubenswrapper[4775]: I1002 03:12:06.007716 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6547599f55-q7x9m"] Oct 02 03:12:06 crc kubenswrapper[4775]: I1002 03:12:06.022429 4775 scope.go:117] "RemoveContainer" containerID="ab5a1bbc3c5f941df91e61ef117563be7348761b32297b73a884cdadccccb2ac" Oct 02 03:12:07 crc kubenswrapper[4775]: I1002 03:12:07.234125 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:12:07 crc kubenswrapper[4775]: I1002 03:12:07.234197 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:12:07 crc kubenswrapper[4775]: I1002 03:12:07.785270 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="905df254-0cfa-4226-848e-926f450b2b18" path="/var/lib/kubelet/pods/905df254-0cfa-4226-848e-926f450b2b18/volumes" Oct 02 03:12:08 crc kubenswrapper[4775]: I1002 03:12:08.257827 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 03:12:08 crc kubenswrapper[4775]: I1002 03:12:08.257897 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 03:12:08 crc kubenswrapper[4775]: I1002 03:12:08.314770 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 03:12:08 crc kubenswrapper[4775]: I1002 03:12:08.332925 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 03:12:09 crc kubenswrapper[4775]: I1002 03:12:09.009859 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 03:12:09 crc kubenswrapper[4775]: I1002 03:12:09.009948 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 03:12:10 crc kubenswrapper[4775]: I1002 03:12:10.265816 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 03:12:10 crc kubenswrapper[4775]: I1002 03:12:10.267407 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 03:12:10 crc kubenswrapper[4775]: I1002 03:12:10.337634 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 03:12:10 crc kubenswrapper[4775]: I1002 03:12:10.350393 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 03:12:10 crc kubenswrapper[4775]: I1002 03:12:10.806277 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 03:12:10 crc kubenswrapper[4775]: I1002 03:12:10.808586 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 03:12:11 crc kubenswrapper[4775]: I1002 03:12:11.030601 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 03:12:11 crc kubenswrapper[4775]: I1002 03:12:11.030641 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 03:12:11 crc kubenswrapper[4775]: I1002 03:12:11.124657 4775 scope.go:117] "RemoveContainer" containerID="d2fa9f54cc55f33effa2e0feb0aebe2f4a5da2cfa68b3365931a5c57a4eb0ca8" Oct 02 03:12:11 crc kubenswrapper[4775]: I1002 03:12:11.177812 4775 scope.go:117] "RemoveContainer" containerID="a130041296f0f400ecb06409082211457f73f30e50e1c5ea06c45ba9df9350ea" Oct 02 03:12:12 crc kubenswrapper[4775]: I1002 03:12:12.753473 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 03:12:12 crc kubenswrapper[4775]: I1002 03:12:12.789930 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 03:12:19 crc kubenswrapper[4775]: I1002 03:12:19.334998 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-gzf2f"] Oct 02 03:12:19 crc kubenswrapper[4775]: E1002 03:12:19.336294 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="905df254-0cfa-4226-848e-926f450b2b18" containerName="dnsmasq-dns" Oct 02 03:12:19 crc kubenswrapper[4775]: I1002 03:12:19.336318 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="905df254-0cfa-4226-848e-926f450b2b18" containerName="dnsmasq-dns" Oct 02 03:12:19 crc kubenswrapper[4775]: E1002 03:12:19.336340 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="905df254-0cfa-4226-848e-926f450b2b18" containerName="init" Oct 02 03:12:19 crc kubenswrapper[4775]: I1002 03:12:19.336352 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="905df254-0cfa-4226-848e-926f450b2b18" containerName="init" Oct 02 03:12:19 crc kubenswrapper[4775]: I1002 03:12:19.336710 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="905df254-0cfa-4226-848e-926f450b2b18" containerName="dnsmasq-dns" Oct 02 03:12:19 crc kubenswrapper[4775]: I1002 03:12:19.337750 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gzf2f" Oct 02 03:12:19 crc kubenswrapper[4775]: I1002 03:12:19.343183 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-gzf2f"] Oct 02 03:12:19 crc kubenswrapper[4775]: I1002 03:12:19.407902 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klj6s\" (UniqueName: \"kubernetes.io/projected/436596a4-5708-4e75-9e51-cf45a2600dc4-kube-api-access-klj6s\") pod \"placement-db-create-gzf2f\" (UID: \"436596a4-5708-4e75-9e51-cf45a2600dc4\") " pod="openstack/placement-db-create-gzf2f" Oct 02 03:12:19 crc kubenswrapper[4775]: I1002 03:12:19.509790 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klj6s\" (UniqueName: \"kubernetes.io/projected/436596a4-5708-4e75-9e51-cf45a2600dc4-kube-api-access-klj6s\") pod \"placement-db-create-gzf2f\" (UID: \"436596a4-5708-4e75-9e51-cf45a2600dc4\") " pod="openstack/placement-db-create-gzf2f" Oct 02 03:12:19 crc kubenswrapper[4775]: I1002 03:12:19.533730 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klj6s\" (UniqueName: \"kubernetes.io/projected/436596a4-5708-4e75-9e51-cf45a2600dc4-kube-api-access-klj6s\") pod \"placement-db-create-gzf2f\" (UID: \"436596a4-5708-4e75-9e51-cf45a2600dc4\") " pod="openstack/placement-db-create-gzf2f" Oct 02 03:12:19 crc kubenswrapper[4775]: I1002 03:12:19.666281 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gzf2f" Oct 02 03:12:20 crc kubenswrapper[4775]: W1002 03:12:20.133513 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod436596a4_5708_4e75_9e51_cf45a2600dc4.slice/crio-09b4dc5f724458f418ec022a21ca3bbba19d876da3d31b99bc5e3017be150b10 WatchSource:0}: Error finding container 09b4dc5f724458f418ec022a21ca3bbba19d876da3d31b99bc5e3017be150b10: Status 404 returned error can't find the container with id 09b4dc5f724458f418ec022a21ca3bbba19d876da3d31b99bc5e3017be150b10 Oct 02 03:12:20 crc kubenswrapper[4775]: I1002 03:12:20.133836 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-gzf2f"] Oct 02 03:12:21 crc kubenswrapper[4775]: I1002 03:12:21.138178 4775 generic.go:334] "Generic (PLEG): container finished" podID="436596a4-5708-4e75-9e51-cf45a2600dc4" containerID="bdf0be3d019ed7273b9673dd8adf2a9d0a9e8dda92f31132581db793bcad9217" exitCode=0 Oct 02 03:12:21 crc kubenswrapper[4775]: I1002 03:12:21.138255 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-gzf2f" event={"ID":"436596a4-5708-4e75-9e51-cf45a2600dc4","Type":"ContainerDied","Data":"bdf0be3d019ed7273b9673dd8adf2a9d0a9e8dda92f31132581db793bcad9217"} Oct 02 03:12:21 crc kubenswrapper[4775]: I1002 03:12:21.138317 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-gzf2f" event={"ID":"436596a4-5708-4e75-9e51-cf45a2600dc4","Type":"ContainerStarted","Data":"09b4dc5f724458f418ec022a21ca3bbba19d876da3d31b99bc5e3017be150b10"} Oct 02 03:12:22 crc kubenswrapper[4775]: I1002 03:12:22.605391 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gzf2f" Oct 02 03:12:22 crc kubenswrapper[4775]: I1002 03:12:22.666549 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klj6s\" (UniqueName: \"kubernetes.io/projected/436596a4-5708-4e75-9e51-cf45a2600dc4-kube-api-access-klj6s\") pod \"436596a4-5708-4e75-9e51-cf45a2600dc4\" (UID: \"436596a4-5708-4e75-9e51-cf45a2600dc4\") " Oct 02 03:12:22 crc kubenswrapper[4775]: I1002 03:12:22.674639 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/436596a4-5708-4e75-9e51-cf45a2600dc4-kube-api-access-klj6s" (OuterVolumeSpecName: "kube-api-access-klj6s") pod "436596a4-5708-4e75-9e51-cf45a2600dc4" (UID: "436596a4-5708-4e75-9e51-cf45a2600dc4"). InnerVolumeSpecName "kube-api-access-klj6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:12:22 crc kubenswrapper[4775]: I1002 03:12:22.767673 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klj6s\" (UniqueName: \"kubernetes.io/projected/436596a4-5708-4e75-9e51-cf45a2600dc4-kube-api-access-klj6s\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:23 crc kubenswrapper[4775]: I1002 03:12:23.165045 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-gzf2f" Oct 02 03:12:23 crc kubenswrapper[4775]: I1002 03:12:23.164947 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-gzf2f" event={"ID":"436596a4-5708-4e75-9e51-cf45a2600dc4","Type":"ContainerDied","Data":"09b4dc5f724458f418ec022a21ca3bbba19d876da3d31b99bc5e3017be150b10"} Oct 02 03:12:23 crc kubenswrapper[4775]: I1002 03:12:23.165902 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09b4dc5f724458f418ec022a21ca3bbba19d876da3d31b99bc5e3017be150b10" Oct 02 03:12:29 crc kubenswrapper[4775]: I1002 03:12:29.419814 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-8804-account-create-vkrcs"] Oct 02 03:12:29 crc kubenswrapper[4775]: E1002 03:12:29.421040 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436596a4-5708-4e75-9e51-cf45a2600dc4" containerName="mariadb-database-create" Oct 02 03:12:29 crc kubenswrapper[4775]: I1002 03:12:29.421069 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="436596a4-5708-4e75-9e51-cf45a2600dc4" containerName="mariadb-database-create" Oct 02 03:12:29 crc kubenswrapper[4775]: I1002 03:12:29.421416 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="436596a4-5708-4e75-9e51-cf45a2600dc4" containerName="mariadb-database-create" Oct 02 03:12:29 crc kubenswrapper[4775]: I1002 03:12:29.422448 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8804-account-create-vkrcs" Oct 02 03:12:29 crc kubenswrapper[4775]: I1002 03:12:29.425442 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 02 03:12:29 crc kubenswrapper[4775]: I1002 03:12:29.430385 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8804-account-create-vkrcs"] Oct 02 03:12:29 crc kubenswrapper[4775]: I1002 03:12:29.508226 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvm9w\" (UniqueName: \"kubernetes.io/projected/7e9ed2ad-06c2-4e53-9359-2c26438ae139-kube-api-access-qvm9w\") pod \"placement-8804-account-create-vkrcs\" (UID: \"7e9ed2ad-06c2-4e53-9359-2c26438ae139\") " pod="openstack/placement-8804-account-create-vkrcs" Oct 02 03:12:29 crc kubenswrapper[4775]: I1002 03:12:29.609948 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvm9w\" (UniqueName: \"kubernetes.io/projected/7e9ed2ad-06c2-4e53-9359-2c26438ae139-kube-api-access-qvm9w\") pod \"placement-8804-account-create-vkrcs\" (UID: \"7e9ed2ad-06c2-4e53-9359-2c26438ae139\") " pod="openstack/placement-8804-account-create-vkrcs" Oct 02 03:12:29 crc kubenswrapper[4775]: I1002 03:12:29.634856 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvm9w\" (UniqueName: \"kubernetes.io/projected/7e9ed2ad-06c2-4e53-9359-2c26438ae139-kube-api-access-qvm9w\") pod \"placement-8804-account-create-vkrcs\" (UID: \"7e9ed2ad-06c2-4e53-9359-2c26438ae139\") " pod="openstack/placement-8804-account-create-vkrcs" Oct 02 03:12:29 crc kubenswrapper[4775]: I1002 03:12:29.767244 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8804-account-create-vkrcs" Oct 02 03:12:30 crc kubenswrapper[4775]: I1002 03:12:30.098201 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-8804-account-create-vkrcs"] Oct 02 03:12:30 crc kubenswrapper[4775]: I1002 03:12:30.244747 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8804-account-create-vkrcs" event={"ID":"7e9ed2ad-06c2-4e53-9359-2c26438ae139","Type":"ContainerStarted","Data":"ec523e857780e146dda3a9217be7c766eeecccb61020d32c5b35587ccbffcdfc"} Oct 02 03:12:31 crc kubenswrapper[4775]: I1002 03:12:31.257819 4775 generic.go:334] "Generic (PLEG): container finished" podID="7e9ed2ad-06c2-4e53-9359-2c26438ae139" containerID="8437f8b75a0f92054fb112385564c062db0176d5340decff9fefa4b465aeb915" exitCode=0 Oct 02 03:12:31 crc kubenswrapper[4775]: I1002 03:12:31.257859 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8804-account-create-vkrcs" event={"ID":"7e9ed2ad-06c2-4e53-9359-2c26438ae139","Type":"ContainerDied","Data":"8437f8b75a0f92054fb112385564c062db0176d5340decff9fefa4b465aeb915"} Oct 02 03:12:32 crc kubenswrapper[4775]: I1002 03:12:32.680174 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8804-account-create-vkrcs" Oct 02 03:12:32 crc kubenswrapper[4775]: I1002 03:12:32.788543 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvm9w\" (UniqueName: \"kubernetes.io/projected/7e9ed2ad-06c2-4e53-9359-2c26438ae139-kube-api-access-qvm9w\") pod \"7e9ed2ad-06c2-4e53-9359-2c26438ae139\" (UID: \"7e9ed2ad-06c2-4e53-9359-2c26438ae139\") " Oct 02 03:12:32 crc kubenswrapper[4775]: I1002 03:12:32.795661 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e9ed2ad-06c2-4e53-9359-2c26438ae139-kube-api-access-qvm9w" (OuterVolumeSpecName: "kube-api-access-qvm9w") pod "7e9ed2ad-06c2-4e53-9359-2c26438ae139" (UID: "7e9ed2ad-06c2-4e53-9359-2c26438ae139"). InnerVolumeSpecName "kube-api-access-qvm9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:12:32 crc kubenswrapper[4775]: I1002 03:12:32.890883 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvm9w\" (UniqueName: \"kubernetes.io/projected/7e9ed2ad-06c2-4e53-9359-2c26438ae139-kube-api-access-qvm9w\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:33 crc kubenswrapper[4775]: I1002 03:12:33.285926 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-8804-account-create-vkrcs" event={"ID":"7e9ed2ad-06c2-4e53-9359-2c26438ae139","Type":"ContainerDied","Data":"ec523e857780e146dda3a9217be7c766eeecccb61020d32c5b35587ccbffcdfc"} Oct 02 03:12:33 crc kubenswrapper[4775]: I1002 03:12:33.286018 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ec523e857780e146dda3a9217be7c766eeecccb61020d32c5b35587ccbffcdfc" Oct 02 03:12:33 crc kubenswrapper[4775]: I1002 03:12:33.286044 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-8804-account-create-vkrcs" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.641761 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-db697c7cc-w4mc6"] Oct 02 03:12:34 crc kubenswrapper[4775]: E1002 03:12:34.642324 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e9ed2ad-06c2-4e53-9359-2c26438ae139" containerName="mariadb-account-create" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.642336 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e9ed2ad-06c2-4e53-9359-2c26438ae139" containerName="mariadb-account-create" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.642501 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e9ed2ad-06c2-4e53-9359-2c26438ae139" containerName="mariadb-account-create" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.643370 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.664098 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db697c7cc-w4mc6"] Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.740101 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-config\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.740160 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-ovsdbserver-nb\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.740180 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-dns-svc\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.740210 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-ovsdbserver-sb\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.740247 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tnsh\" (UniqueName: \"kubernetes.io/projected/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-kube-api-access-4tnsh\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.761076 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-v4zst"] Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.762188 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.767305 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.767858 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-b8kp5" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.767985 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.780428 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-v4zst"] Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.841804 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-config\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.841869 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-ovsdbserver-nb\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.841893 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-dns-svc\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.841925 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-ovsdbserver-sb\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.841971 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-scripts\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.841989 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tnsh\" (UniqueName: \"kubernetes.io/projected/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-kube-api-access-4tnsh\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.842012 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afdda688-8e5c-49e4-83fc-1ba569359404-logs\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.842030 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-config-data\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.842067 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-combined-ca-bundle\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.842084 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dd2tg\" (UniqueName: \"kubernetes.io/projected/afdda688-8e5c-49e4-83fc-1ba569359404-kube-api-access-dd2tg\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.842922 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-config\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.843496 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-dns-svc\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.846296 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-ovsdbserver-nb\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.848120 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-ovsdbserver-sb\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.867254 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tnsh\" (UniqueName: \"kubernetes.io/projected/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-kube-api-access-4tnsh\") pod \"dnsmasq-dns-db697c7cc-w4mc6\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.942873 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-combined-ca-bundle\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.943170 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dd2tg\" (UniqueName: \"kubernetes.io/projected/afdda688-8e5c-49e4-83fc-1ba569359404-kube-api-access-dd2tg\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.943321 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-scripts\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.943359 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afdda688-8e5c-49e4-83fc-1ba569359404-logs\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.943382 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-config-data\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.943848 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afdda688-8e5c-49e4-83fc-1ba569359404-logs\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.946681 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-combined-ca-bundle\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.947102 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-config-data\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.953234 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-scripts\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:34 crc kubenswrapper[4775]: I1002 03:12:34.958728 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dd2tg\" (UniqueName: \"kubernetes.io/projected/afdda688-8e5c-49e4-83fc-1ba569359404-kube-api-access-dd2tg\") pod \"placement-db-sync-v4zst\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:35 crc kubenswrapper[4775]: I1002 03:12:35.007038 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:35 crc kubenswrapper[4775]: I1002 03:12:35.118174 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:35 crc kubenswrapper[4775]: I1002 03:12:35.443766 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db697c7cc-w4mc6"] Oct 02 03:12:35 crc kubenswrapper[4775]: W1002 03:12:35.444845 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ae17bef_94ba_4e40_ae25_fbcf8e4dab79.slice/crio-979c422b0cda2e52b370b2a9d7568c0035c14b0ace99efeb3a5e8967ddc29bf3 WatchSource:0}: Error finding container 979c422b0cda2e52b370b2a9d7568c0035c14b0ace99efeb3a5e8967ddc29bf3: Status 404 returned error can't find the container with id 979c422b0cda2e52b370b2a9d7568c0035c14b0ace99efeb3a5e8967ddc29bf3 Oct 02 03:12:35 crc kubenswrapper[4775]: I1002 03:12:35.597162 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-v4zst"] Oct 02 03:12:35 crc kubenswrapper[4775]: W1002 03:12:35.602971 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafdda688_8e5c_49e4_83fc_1ba569359404.slice/crio-f3b31474e8e594b34bc570e9123f0339c83d259f9be12c687759302f55129cae WatchSource:0}: Error finding container f3b31474e8e594b34bc570e9123f0339c83d259f9be12c687759302f55129cae: Status 404 returned error can't find the container with id f3b31474e8e594b34bc570e9123f0339c83d259f9be12c687759302f55129cae Oct 02 03:12:36 crc kubenswrapper[4775]: I1002 03:12:36.320470 4775 generic.go:334] "Generic (PLEG): container finished" podID="2ae17bef-94ba-4e40-ae25-fbcf8e4dab79" containerID="423c8d898cad7dc57d306509070a670cac226b66e0bcc654e0f7dea88faf5cc9" exitCode=0 Oct 02 03:12:36 crc kubenswrapper[4775]: I1002 03:12:36.320565 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" event={"ID":"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79","Type":"ContainerDied","Data":"423c8d898cad7dc57d306509070a670cac226b66e0bcc654e0f7dea88faf5cc9"} Oct 02 03:12:36 crc kubenswrapper[4775]: I1002 03:12:36.321210 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" event={"ID":"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79","Type":"ContainerStarted","Data":"979c422b0cda2e52b370b2a9d7568c0035c14b0ace99efeb3a5e8967ddc29bf3"} Oct 02 03:12:36 crc kubenswrapper[4775]: I1002 03:12:36.324439 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-v4zst" event={"ID":"afdda688-8e5c-49e4-83fc-1ba569359404","Type":"ContainerStarted","Data":"d2c7bfa8a50cd48145fbcb8de04dd2aafa3e5db017948be1c67e31d1957bc6c3"} Oct 02 03:12:36 crc kubenswrapper[4775]: I1002 03:12:36.324507 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-v4zst" event={"ID":"afdda688-8e5c-49e4-83fc-1ba569359404","Type":"ContainerStarted","Data":"f3b31474e8e594b34bc570e9123f0339c83d259f9be12c687759302f55129cae"} Oct 02 03:12:36 crc kubenswrapper[4775]: I1002 03:12:36.388471 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-v4zst" podStartSLOduration=2.388446197 podStartE2EDuration="2.388446197s" podCreationTimestamp="2025-10-02 03:12:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:12:36.381711463 +0000 UTC m=+5493.548455543" watchObservedRunningTime="2025-10-02 03:12:36.388446197 +0000 UTC m=+5493.555190267" Oct 02 03:12:37 crc kubenswrapper[4775]: I1002 03:12:37.233588 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:12:37 crc kubenswrapper[4775]: I1002 03:12:37.233680 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:12:37 crc kubenswrapper[4775]: I1002 03:12:37.233745 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 03:12:37 crc kubenswrapper[4775]: I1002 03:12:37.234942 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5c0c35a295e1473ae177f82875e76ff0ae056f4db6af4a094e5ee23221bb6584"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 03:12:37 crc kubenswrapper[4775]: I1002 03:12:37.235071 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://5c0c35a295e1473ae177f82875e76ff0ae056f4db6af4a094e5ee23221bb6584" gracePeriod=600 Oct 02 03:12:37 crc kubenswrapper[4775]: I1002 03:12:37.340033 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" event={"ID":"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79","Type":"ContainerStarted","Data":"28edfbd7b63632af67936b1c47cd2e2be628c7e79cca126a5cf1a38f1e3ce795"} Oct 02 03:12:37 crc kubenswrapper[4775]: I1002 03:12:37.340371 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:37 crc kubenswrapper[4775]: I1002 03:12:37.342093 4775 generic.go:334] "Generic (PLEG): container finished" podID="afdda688-8e5c-49e4-83fc-1ba569359404" containerID="d2c7bfa8a50cd48145fbcb8de04dd2aafa3e5db017948be1c67e31d1957bc6c3" exitCode=0 Oct 02 03:12:37 crc kubenswrapper[4775]: I1002 03:12:37.342141 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-v4zst" event={"ID":"afdda688-8e5c-49e4-83fc-1ba569359404","Type":"ContainerDied","Data":"d2c7bfa8a50cd48145fbcb8de04dd2aafa3e5db017948be1c67e31d1957bc6c3"} Oct 02 03:12:37 crc kubenswrapper[4775]: I1002 03:12:37.378177 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" podStartSLOduration=3.378146159 podStartE2EDuration="3.378146159s" podCreationTimestamp="2025-10-02 03:12:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:12:37.369658129 +0000 UTC m=+5494.536402239" watchObservedRunningTime="2025-10-02 03:12:37.378146159 +0000 UTC m=+5494.544890229" Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.360432 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="5c0c35a295e1473ae177f82875e76ff0ae056f4db6af4a094e5ee23221bb6584" exitCode=0 Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.360684 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"5c0c35a295e1473ae177f82875e76ff0ae056f4db6af4a094e5ee23221bb6584"} Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.361364 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6"} Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.361414 4775 scope.go:117] "RemoveContainer" containerID="fa9bc7c4ef51dbfac8466157f314a4348dabd156a8da29518b154c0d78e42636" Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.739582 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.826903 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-config-data\") pod \"afdda688-8e5c-49e4-83fc-1ba569359404\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.827073 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-combined-ca-bundle\") pod \"afdda688-8e5c-49e4-83fc-1ba569359404\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.827184 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afdda688-8e5c-49e4-83fc-1ba569359404-logs\") pod \"afdda688-8e5c-49e4-83fc-1ba569359404\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.827217 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-scripts\") pod \"afdda688-8e5c-49e4-83fc-1ba569359404\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.827257 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dd2tg\" (UniqueName: \"kubernetes.io/projected/afdda688-8e5c-49e4-83fc-1ba569359404-kube-api-access-dd2tg\") pod \"afdda688-8e5c-49e4-83fc-1ba569359404\" (UID: \"afdda688-8e5c-49e4-83fc-1ba569359404\") " Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.828138 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afdda688-8e5c-49e4-83fc-1ba569359404-logs" (OuterVolumeSpecName: "logs") pod "afdda688-8e5c-49e4-83fc-1ba569359404" (UID: "afdda688-8e5c-49e4-83fc-1ba569359404"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.833424 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afdda688-8e5c-49e4-83fc-1ba569359404-kube-api-access-dd2tg" (OuterVolumeSpecName: "kube-api-access-dd2tg") pod "afdda688-8e5c-49e4-83fc-1ba569359404" (UID: "afdda688-8e5c-49e4-83fc-1ba569359404"). InnerVolumeSpecName "kube-api-access-dd2tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.836297 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-scripts" (OuterVolumeSpecName: "scripts") pod "afdda688-8e5c-49e4-83fc-1ba569359404" (UID: "afdda688-8e5c-49e4-83fc-1ba569359404"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.849907 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "afdda688-8e5c-49e4-83fc-1ba569359404" (UID: "afdda688-8e5c-49e4-83fc-1ba569359404"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.870024 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-config-data" (OuterVolumeSpecName: "config-data") pod "afdda688-8e5c-49e4-83fc-1ba569359404" (UID: "afdda688-8e5c-49e4-83fc-1ba569359404"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.928735 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afdda688-8e5c-49e4-83fc-1ba569359404-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.928784 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.928803 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dd2tg\" (UniqueName: \"kubernetes.io/projected/afdda688-8e5c-49e4-83fc-1ba569359404-kube-api-access-dd2tg\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.928827 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:38 crc kubenswrapper[4775]: I1002 03:12:38.928845 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afdda688-8e5c-49e4-83fc-1ba569359404-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.375772 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-v4zst" event={"ID":"afdda688-8e5c-49e4-83fc-1ba569359404","Type":"ContainerDied","Data":"f3b31474e8e594b34bc570e9123f0339c83d259f9be12c687759302f55129cae"} Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.375827 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-v4zst" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.375830 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3b31474e8e594b34bc570e9123f0339c83d259f9be12c687759302f55129cae" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.592594 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7575468d44-vzv5w"] Oct 02 03:12:39 crc kubenswrapper[4775]: E1002 03:12:39.593169 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afdda688-8e5c-49e4-83fc-1ba569359404" containerName="placement-db-sync" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.593200 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="afdda688-8e5c-49e4-83fc-1ba569359404" containerName="placement-db-sync" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.593525 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="afdda688-8e5c-49e4-83fc-1ba569359404" containerName="placement-db-sync" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.595161 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.599589 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-b8kp5" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.599988 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.600838 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.612264 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7575468d44-vzv5w"] Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.655209 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-combined-ca-bundle\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.655255 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-logs\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.655279 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct672\" (UniqueName: \"kubernetes.io/projected/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-kube-api-access-ct672\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.655319 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-config-data\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.655397 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-scripts\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.757353 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-config-data\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.757466 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-scripts\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.757516 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-combined-ca-bundle\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.757537 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-logs\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.757579 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct672\" (UniqueName: \"kubernetes.io/projected/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-kube-api-access-ct672\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.762307 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-logs\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.765178 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-combined-ca-bundle\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.766504 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-config-data\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.775944 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-scripts\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.777294 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct672\" (UniqueName: \"kubernetes.io/projected/ecdfed0f-2bee-4334-9e88-ea6ad912a1e4-kube-api-access-ct672\") pod \"placement-7575468d44-vzv5w\" (UID: \"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4\") " pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:39 crc kubenswrapper[4775]: I1002 03:12:39.964254 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:41 crc kubenswrapper[4775]: I1002 03:12:41.310038 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7575468d44-vzv5w"] Oct 02 03:12:41 crc kubenswrapper[4775]: I1002 03:12:41.405547 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7575468d44-vzv5w" event={"ID":"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4","Type":"ContainerStarted","Data":"616827dc7e07b3d4942c492cad659ff66868ce38f401dc5bdac88a3e6a0acb32"} Oct 02 03:12:42 crc kubenswrapper[4775]: I1002 03:12:42.422190 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7575468d44-vzv5w" event={"ID":"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4","Type":"ContainerStarted","Data":"fdd27632bbf18990b1185c35d411a9180cfd65a64537f86a6ebd258f37f8415f"} Oct 02 03:12:42 crc kubenswrapper[4775]: I1002 03:12:42.422816 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:42 crc kubenswrapper[4775]: I1002 03:12:42.422837 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7575468d44-vzv5w" event={"ID":"ecdfed0f-2bee-4334-9e88-ea6ad912a1e4","Type":"ContainerStarted","Data":"ce807734a6335e52fc2fcd46a6ae2ff43964373919c7c4142847f2d861177736"} Oct 02 03:12:42 crc kubenswrapper[4775]: I1002 03:12:42.471354 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7575468d44-vzv5w" podStartSLOduration=3.471311225 podStartE2EDuration="3.471311225s" podCreationTimestamp="2025-10-02 03:12:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:12:42.456536262 +0000 UTC m=+5499.623280372" watchObservedRunningTime="2025-10-02 03:12:42.471311225 +0000 UTC m=+5499.638055295" Oct 02 03:12:43 crc kubenswrapper[4775]: I1002 03:12:43.433930 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.008286 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.087425 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-858b587f45-rzn5s"] Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.087780 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-858b587f45-rzn5s" podUID="8a35f34a-2e54-4107-a630-523cc5d05dd5" containerName="dnsmasq-dns" containerID="cri-o://275edac11af78849528ab6096409da999cc96af821ec8254da67b07da64a1845" gracePeriod=10 Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.482082 4775 generic.go:334] "Generic (PLEG): container finished" podID="8a35f34a-2e54-4107-a630-523cc5d05dd5" containerID="275edac11af78849528ab6096409da999cc96af821ec8254da67b07da64a1845" exitCode=0 Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.482321 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858b587f45-rzn5s" event={"ID":"8a35f34a-2e54-4107-a630-523cc5d05dd5","Type":"ContainerDied","Data":"275edac11af78849528ab6096409da999cc96af821ec8254da67b07da64a1845"} Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.592455 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.681365 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-ovsdbserver-nb\") pod \"8a35f34a-2e54-4107-a630-523cc5d05dd5\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.681458 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-662b6\" (UniqueName: \"kubernetes.io/projected/8a35f34a-2e54-4107-a630-523cc5d05dd5-kube-api-access-662b6\") pod \"8a35f34a-2e54-4107-a630-523cc5d05dd5\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.681536 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-config\") pod \"8a35f34a-2e54-4107-a630-523cc5d05dd5\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.681668 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-dns-svc\") pod \"8a35f34a-2e54-4107-a630-523cc5d05dd5\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.681748 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-ovsdbserver-sb\") pod \"8a35f34a-2e54-4107-a630-523cc5d05dd5\" (UID: \"8a35f34a-2e54-4107-a630-523cc5d05dd5\") " Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.686868 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a35f34a-2e54-4107-a630-523cc5d05dd5-kube-api-access-662b6" (OuterVolumeSpecName: "kube-api-access-662b6") pod "8a35f34a-2e54-4107-a630-523cc5d05dd5" (UID: "8a35f34a-2e54-4107-a630-523cc5d05dd5"). InnerVolumeSpecName "kube-api-access-662b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.722054 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-config" (OuterVolumeSpecName: "config") pod "8a35f34a-2e54-4107-a630-523cc5d05dd5" (UID: "8a35f34a-2e54-4107-a630-523cc5d05dd5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.727595 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8a35f34a-2e54-4107-a630-523cc5d05dd5" (UID: "8a35f34a-2e54-4107-a630-523cc5d05dd5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.738789 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8a35f34a-2e54-4107-a630-523cc5d05dd5" (UID: "8a35f34a-2e54-4107-a630-523cc5d05dd5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.739059 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8a35f34a-2e54-4107-a630-523cc5d05dd5" (UID: "8a35f34a-2e54-4107-a630-523cc5d05dd5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.784396 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.784428 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.784445 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-662b6\" (UniqueName: \"kubernetes.io/projected/8a35f34a-2e54-4107-a630-523cc5d05dd5-kube-api-access-662b6\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.784459 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:45 crc kubenswrapper[4775]: I1002 03:12:45.784472 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a35f34a-2e54-4107-a630-523cc5d05dd5-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 03:12:45 crc kubenswrapper[4775]: E1002 03:12:45.925627 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a35f34a_2e54_4107_a630_523cc5d05dd5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a35f34a_2e54_4107_a630_523cc5d05dd5.slice/crio-4efa6d8f2aeb39a4420cba16b22e5d382fa0177a38c5aec856ceee2fa63b1013\": RecentStats: unable to find data in memory cache]" Oct 02 03:12:46 crc kubenswrapper[4775]: I1002 03:12:46.499116 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-858b587f45-rzn5s" event={"ID":"8a35f34a-2e54-4107-a630-523cc5d05dd5","Type":"ContainerDied","Data":"4efa6d8f2aeb39a4420cba16b22e5d382fa0177a38c5aec856ceee2fa63b1013"} Oct 02 03:12:46 crc kubenswrapper[4775]: I1002 03:12:46.499399 4775 scope.go:117] "RemoveContainer" containerID="275edac11af78849528ab6096409da999cc96af821ec8254da67b07da64a1845" Oct 02 03:12:46 crc kubenswrapper[4775]: I1002 03:12:46.499232 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-858b587f45-rzn5s" Oct 02 03:12:46 crc kubenswrapper[4775]: I1002 03:12:46.543882 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-858b587f45-rzn5s"] Oct 02 03:12:46 crc kubenswrapper[4775]: I1002 03:12:46.544275 4775 scope.go:117] "RemoveContainer" containerID="cca83b4c1e79415dd1ea5766a9d77c4836869632bfd78c7bbe84c07ba87a8ee0" Oct 02 03:12:46 crc kubenswrapper[4775]: I1002 03:12:46.556237 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-858b587f45-rzn5s"] Oct 02 03:12:47 crc kubenswrapper[4775]: I1002 03:12:47.810451 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a35f34a-2e54-4107-a630-523cc5d05dd5" path="/var/lib/kubelet/pods/8a35f34a-2e54-4107-a630-523cc5d05dd5/volumes" Oct 02 03:13:10 crc kubenswrapper[4775]: I1002 03:13:10.937802 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:13:10 crc kubenswrapper[4775]: I1002 03:13:10.938931 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7575468d44-vzv5w" Oct 02 03:13:33 crc kubenswrapper[4775]: I1002 03:13:33.699684 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-srnpv"] Oct 02 03:13:33 crc kubenswrapper[4775]: E1002 03:13:33.700486 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a35f34a-2e54-4107-a630-523cc5d05dd5" containerName="dnsmasq-dns" Oct 02 03:13:33 crc kubenswrapper[4775]: I1002 03:13:33.700499 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a35f34a-2e54-4107-a630-523cc5d05dd5" containerName="dnsmasq-dns" Oct 02 03:13:33 crc kubenswrapper[4775]: E1002 03:13:33.700513 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a35f34a-2e54-4107-a630-523cc5d05dd5" containerName="init" Oct 02 03:13:33 crc kubenswrapper[4775]: I1002 03:13:33.700519 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a35f34a-2e54-4107-a630-523cc5d05dd5" containerName="init" Oct 02 03:13:33 crc kubenswrapper[4775]: I1002 03:13:33.700681 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a35f34a-2e54-4107-a630-523cc5d05dd5" containerName="dnsmasq-dns" Oct 02 03:13:33 crc kubenswrapper[4775]: I1002 03:13:33.701301 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-srnpv" Oct 02 03:13:33 crc kubenswrapper[4775]: I1002 03:13:33.709233 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-srnpv"] Oct 02 03:13:33 crc kubenswrapper[4775]: I1002 03:13:33.801898 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-kzdtw"] Oct 02 03:13:33 crc kubenswrapper[4775]: I1002 03:13:33.803077 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kzdtw" Oct 02 03:13:33 crc kubenswrapper[4775]: I1002 03:13:33.826641 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-kzdtw"] Oct 02 03:13:33 crc kubenswrapper[4775]: I1002 03:13:33.862313 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs8p4\" (UniqueName: \"kubernetes.io/projected/b563b0be-d1f0-431b-9e23-39b5a55ffc12-kube-api-access-bs8p4\") pod \"nova-api-db-create-srnpv\" (UID: \"b563b0be-d1f0-431b-9e23-39b5a55ffc12\") " pod="openstack/nova-api-db-create-srnpv" Oct 02 03:13:33 crc kubenswrapper[4775]: I1002 03:13:33.964105 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs8p4\" (UniqueName: \"kubernetes.io/projected/b563b0be-d1f0-431b-9e23-39b5a55ffc12-kube-api-access-bs8p4\") pod \"nova-api-db-create-srnpv\" (UID: \"b563b0be-d1f0-431b-9e23-39b5a55ffc12\") " pod="openstack/nova-api-db-create-srnpv" Oct 02 03:13:33 crc kubenswrapper[4775]: I1002 03:13:33.964206 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qssmj\" (UniqueName: \"kubernetes.io/projected/6223610a-fc09-44d4-9306-2b20384affeb-kube-api-access-qssmj\") pod \"nova-cell0-db-create-kzdtw\" (UID: \"6223610a-fc09-44d4-9306-2b20384affeb\") " pod="openstack/nova-cell0-db-create-kzdtw" Oct 02 03:13:33 crc kubenswrapper[4775]: I1002 03:13:33.983487 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs8p4\" (UniqueName: \"kubernetes.io/projected/b563b0be-d1f0-431b-9e23-39b5a55ffc12-kube-api-access-bs8p4\") pod \"nova-api-db-create-srnpv\" (UID: \"b563b0be-d1f0-431b-9e23-39b5a55ffc12\") " pod="openstack/nova-api-db-create-srnpv" Oct 02 03:13:34 crc kubenswrapper[4775]: I1002 03:13:34.016103 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-4zbzp"] Oct 02 03:13:34 crc kubenswrapper[4775]: I1002 03:13:34.018118 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4zbzp" Oct 02 03:13:34 crc kubenswrapper[4775]: I1002 03:13:34.023294 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4zbzp"] Oct 02 03:13:34 crc kubenswrapper[4775]: I1002 03:13:34.026322 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-srnpv" Oct 02 03:13:34 crc kubenswrapper[4775]: I1002 03:13:34.066420 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qssmj\" (UniqueName: \"kubernetes.io/projected/6223610a-fc09-44d4-9306-2b20384affeb-kube-api-access-qssmj\") pod \"nova-cell0-db-create-kzdtw\" (UID: \"6223610a-fc09-44d4-9306-2b20384affeb\") " pod="openstack/nova-cell0-db-create-kzdtw" Oct 02 03:13:34 crc kubenswrapper[4775]: I1002 03:13:34.083092 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qssmj\" (UniqueName: \"kubernetes.io/projected/6223610a-fc09-44d4-9306-2b20384affeb-kube-api-access-qssmj\") pod \"nova-cell0-db-create-kzdtw\" (UID: \"6223610a-fc09-44d4-9306-2b20384affeb\") " pod="openstack/nova-cell0-db-create-kzdtw" Oct 02 03:13:34 crc kubenswrapper[4775]: I1002 03:13:34.151628 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kzdtw" Oct 02 03:13:34 crc kubenswrapper[4775]: I1002 03:13:34.167979 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k7fs\" (UniqueName: \"kubernetes.io/projected/6960b657-431b-433a-b1b1-175d8fa098a7-kube-api-access-5k7fs\") pod \"nova-cell1-db-create-4zbzp\" (UID: \"6960b657-431b-433a-b1b1-175d8fa098a7\") " pod="openstack/nova-cell1-db-create-4zbzp" Oct 02 03:13:34 crc kubenswrapper[4775]: I1002 03:13:34.276728 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k7fs\" (UniqueName: \"kubernetes.io/projected/6960b657-431b-433a-b1b1-175d8fa098a7-kube-api-access-5k7fs\") pod \"nova-cell1-db-create-4zbzp\" (UID: \"6960b657-431b-433a-b1b1-175d8fa098a7\") " pod="openstack/nova-cell1-db-create-4zbzp" Oct 02 03:13:34 crc kubenswrapper[4775]: I1002 03:13:34.297232 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k7fs\" (UniqueName: \"kubernetes.io/projected/6960b657-431b-433a-b1b1-175d8fa098a7-kube-api-access-5k7fs\") pod \"nova-cell1-db-create-4zbzp\" (UID: \"6960b657-431b-433a-b1b1-175d8fa098a7\") " pod="openstack/nova-cell1-db-create-4zbzp" Oct 02 03:13:34 crc kubenswrapper[4775]: I1002 03:13:34.335409 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4zbzp" Oct 02 03:13:34 crc kubenswrapper[4775]: I1002 03:13:34.493990 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-srnpv"] Oct 02 03:13:34 crc kubenswrapper[4775]: I1002 03:13:34.593314 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-kzdtw"] Oct 02 03:13:34 crc kubenswrapper[4775]: W1002 03:13:34.617180 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6223610a_fc09_44d4_9306_2b20384affeb.slice/crio-c4d4e642ddd616928215ca88616625ff58ce0d5d74429c6a1e9e1a08995fec54 WatchSource:0}: Error finding container c4d4e642ddd616928215ca88616625ff58ce0d5d74429c6a1e9e1a08995fec54: Status 404 returned error can't find the container with id c4d4e642ddd616928215ca88616625ff58ce0d5d74429c6a1e9e1a08995fec54 Oct 02 03:13:34 crc kubenswrapper[4775]: I1002 03:13:34.880297 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-4zbzp"] Oct 02 03:13:34 crc kubenswrapper[4775]: W1002 03:13:34.911452 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6960b657_431b_433a_b1b1_175d8fa098a7.slice/crio-0f3c9ae96b856cfa26f22225e9197e4b9f2ddb3e4727e40c0740a49e0e733077 WatchSource:0}: Error finding container 0f3c9ae96b856cfa26f22225e9197e4b9f2ddb3e4727e40c0740a49e0e733077: Status 404 returned error can't find the container with id 0f3c9ae96b856cfa26f22225e9197e4b9f2ddb3e4727e40c0740a49e0e733077 Oct 02 03:13:35 crc kubenswrapper[4775]: I1002 03:13:35.022524 4775 generic.go:334] "Generic (PLEG): container finished" podID="6223610a-fc09-44d4-9306-2b20384affeb" containerID="ce1ed864de6e669bebfce7170ddb56905622149fb28c1c5d1adf537e0f06c83c" exitCode=0 Oct 02 03:13:35 crc kubenswrapper[4775]: I1002 03:13:35.022634 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-kzdtw" event={"ID":"6223610a-fc09-44d4-9306-2b20384affeb","Type":"ContainerDied","Data":"ce1ed864de6e669bebfce7170ddb56905622149fb28c1c5d1adf537e0f06c83c"} Oct 02 03:13:35 crc kubenswrapper[4775]: I1002 03:13:35.022675 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-kzdtw" event={"ID":"6223610a-fc09-44d4-9306-2b20384affeb","Type":"ContainerStarted","Data":"c4d4e642ddd616928215ca88616625ff58ce0d5d74429c6a1e9e1a08995fec54"} Oct 02 03:13:35 crc kubenswrapper[4775]: I1002 03:13:35.024468 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4zbzp" event={"ID":"6960b657-431b-433a-b1b1-175d8fa098a7","Type":"ContainerStarted","Data":"0f3c9ae96b856cfa26f22225e9197e4b9f2ddb3e4727e40c0740a49e0e733077"} Oct 02 03:13:35 crc kubenswrapper[4775]: I1002 03:13:35.027413 4775 generic.go:334] "Generic (PLEG): container finished" podID="b563b0be-d1f0-431b-9e23-39b5a55ffc12" containerID="fbe8b70ea6913b6eabce06048c0b6162a6cd2cd0250406c9ab4709c54ad75d8c" exitCode=0 Oct 02 03:13:35 crc kubenswrapper[4775]: I1002 03:13:35.027463 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-srnpv" event={"ID":"b563b0be-d1f0-431b-9e23-39b5a55ffc12","Type":"ContainerDied","Data":"fbe8b70ea6913b6eabce06048c0b6162a6cd2cd0250406c9ab4709c54ad75d8c"} Oct 02 03:13:35 crc kubenswrapper[4775]: I1002 03:13:35.027515 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-srnpv" event={"ID":"b563b0be-d1f0-431b-9e23-39b5a55ffc12","Type":"ContainerStarted","Data":"371ad0ba535e0e78310d6a29459468db473452c8d4af4ad03158dc5231bb1ba8"} Oct 02 03:13:36 crc kubenswrapper[4775]: I1002 03:13:36.040229 4775 generic.go:334] "Generic (PLEG): container finished" podID="6960b657-431b-433a-b1b1-175d8fa098a7" containerID="f79193914f690f7de1e103b446483f0daf8e40b21386f110c8f0d4824bff35c3" exitCode=0 Oct 02 03:13:36 crc kubenswrapper[4775]: I1002 03:13:36.040314 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4zbzp" event={"ID":"6960b657-431b-433a-b1b1-175d8fa098a7","Type":"ContainerDied","Data":"f79193914f690f7de1e103b446483f0daf8e40b21386f110c8f0d4824bff35c3"} Oct 02 03:13:36 crc kubenswrapper[4775]: I1002 03:13:36.467323 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kzdtw" Oct 02 03:13:36 crc kubenswrapper[4775]: I1002 03:13:36.478392 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-srnpv" Oct 02 03:13:36 crc kubenswrapper[4775]: I1002 03:13:36.646228 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs8p4\" (UniqueName: \"kubernetes.io/projected/b563b0be-d1f0-431b-9e23-39b5a55ffc12-kube-api-access-bs8p4\") pod \"b563b0be-d1f0-431b-9e23-39b5a55ffc12\" (UID: \"b563b0be-d1f0-431b-9e23-39b5a55ffc12\") " Oct 02 03:13:36 crc kubenswrapper[4775]: I1002 03:13:36.646736 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qssmj\" (UniqueName: \"kubernetes.io/projected/6223610a-fc09-44d4-9306-2b20384affeb-kube-api-access-qssmj\") pod \"6223610a-fc09-44d4-9306-2b20384affeb\" (UID: \"6223610a-fc09-44d4-9306-2b20384affeb\") " Oct 02 03:13:36 crc kubenswrapper[4775]: I1002 03:13:36.652343 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b563b0be-d1f0-431b-9e23-39b5a55ffc12-kube-api-access-bs8p4" (OuterVolumeSpecName: "kube-api-access-bs8p4") pod "b563b0be-d1f0-431b-9e23-39b5a55ffc12" (UID: "b563b0be-d1f0-431b-9e23-39b5a55ffc12"). InnerVolumeSpecName "kube-api-access-bs8p4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:13:36 crc kubenswrapper[4775]: I1002 03:13:36.652588 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6223610a-fc09-44d4-9306-2b20384affeb-kube-api-access-qssmj" (OuterVolumeSpecName: "kube-api-access-qssmj") pod "6223610a-fc09-44d4-9306-2b20384affeb" (UID: "6223610a-fc09-44d4-9306-2b20384affeb"). InnerVolumeSpecName "kube-api-access-qssmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:13:36 crc kubenswrapper[4775]: I1002 03:13:36.748137 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qssmj\" (UniqueName: \"kubernetes.io/projected/6223610a-fc09-44d4-9306-2b20384affeb-kube-api-access-qssmj\") on node \"crc\" DevicePath \"\"" Oct 02 03:13:36 crc kubenswrapper[4775]: I1002 03:13:36.748187 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs8p4\" (UniqueName: \"kubernetes.io/projected/b563b0be-d1f0-431b-9e23-39b5a55ffc12-kube-api-access-bs8p4\") on node \"crc\" DevicePath \"\"" Oct 02 03:13:37 crc kubenswrapper[4775]: I1002 03:13:37.059384 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-srnpv" Oct 02 03:13:37 crc kubenswrapper[4775]: I1002 03:13:37.060191 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-srnpv" event={"ID":"b563b0be-d1f0-431b-9e23-39b5a55ffc12","Type":"ContainerDied","Data":"371ad0ba535e0e78310d6a29459468db473452c8d4af4ad03158dc5231bb1ba8"} Oct 02 03:13:37 crc kubenswrapper[4775]: I1002 03:13:37.060306 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="371ad0ba535e0e78310d6a29459468db473452c8d4af4ad03158dc5231bb1ba8" Oct 02 03:13:37 crc kubenswrapper[4775]: I1002 03:13:37.061805 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-kzdtw" event={"ID":"6223610a-fc09-44d4-9306-2b20384affeb","Type":"ContainerDied","Data":"c4d4e642ddd616928215ca88616625ff58ce0d5d74429c6a1e9e1a08995fec54"} Oct 02 03:13:37 crc kubenswrapper[4775]: I1002 03:13:37.062037 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4d4e642ddd616928215ca88616625ff58ce0d5d74429c6a1e9e1a08995fec54" Oct 02 03:13:37 crc kubenswrapper[4775]: I1002 03:13:37.061861 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kzdtw" Oct 02 03:13:37 crc kubenswrapper[4775]: I1002 03:13:37.447611 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4zbzp" Oct 02 03:13:37 crc kubenswrapper[4775]: I1002 03:13:37.561835 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k7fs\" (UniqueName: \"kubernetes.io/projected/6960b657-431b-433a-b1b1-175d8fa098a7-kube-api-access-5k7fs\") pod \"6960b657-431b-433a-b1b1-175d8fa098a7\" (UID: \"6960b657-431b-433a-b1b1-175d8fa098a7\") " Oct 02 03:13:37 crc kubenswrapper[4775]: I1002 03:13:37.567418 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6960b657-431b-433a-b1b1-175d8fa098a7-kube-api-access-5k7fs" (OuterVolumeSpecName: "kube-api-access-5k7fs") pod "6960b657-431b-433a-b1b1-175d8fa098a7" (UID: "6960b657-431b-433a-b1b1-175d8fa098a7"). InnerVolumeSpecName "kube-api-access-5k7fs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:13:37 crc kubenswrapper[4775]: I1002 03:13:37.665051 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k7fs\" (UniqueName: \"kubernetes.io/projected/6960b657-431b-433a-b1b1-175d8fa098a7-kube-api-access-5k7fs\") on node \"crc\" DevicePath \"\"" Oct 02 03:13:38 crc kubenswrapper[4775]: I1002 03:13:38.074354 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-4zbzp" event={"ID":"6960b657-431b-433a-b1b1-175d8fa098a7","Type":"ContainerDied","Data":"0f3c9ae96b856cfa26f22225e9197e4b9f2ddb3e4727e40c0740a49e0e733077"} Oct 02 03:13:38 crc kubenswrapper[4775]: I1002 03:13:38.074415 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f3c9ae96b856cfa26f22225e9197e4b9f2ddb3e4727e40c0740a49e0e733077" Oct 02 03:13:38 crc kubenswrapper[4775]: I1002 03:13:38.074436 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-4zbzp" Oct 02 03:13:43 crc kubenswrapper[4775]: I1002 03:13:43.942176 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-f7fd-account-create-29988"] Oct 02 03:13:43 crc kubenswrapper[4775]: E1002 03:13:43.943508 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6223610a-fc09-44d4-9306-2b20384affeb" containerName="mariadb-database-create" Oct 02 03:13:43 crc kubenswrapper[4775]: I1002 03:13:43.943535 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6223610a-fc09-44d4-9306-2b20384affeb" containerName="mariadb-database-create" Oct 02 03:13:43 crc kubenswrapper[4775]: E1002 03:13:43.943568 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b563b0be-d1f0-431b-9e23-39b5a55ffc12" containerName="mariadb-database-create" Oct 02 03:13:43 crc kubenswrapper[4775]: I1002 03:13:43.943580 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b563b0be-d1f0-431b-9e23-39b5a55ffc12" containerName="mariadb-database-create" Oct 02 03:13:43 crc kubenswrapper[4775]: E1002 03:13:43.943616 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6960b657-431b-433a-b1b1-175d8fa098a7" containerName="mariadb-database-create" Oct 02 03:13:43 crc kubenswrapper[4775]: I1002 03:13:43.943629 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6960b657-431b-433a-b1b1-175d8fa098a7" containerName="mariadb-database-create" Oct 02 03:13:43 crc kubenswrapper[4775]: I1002 03:13:43.943991 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6960b657-431b-433a-b1b1-175d8fa098a7" containerName="mariadb-database-create" Oct 02 03:13:43 crc kubenswrapper[4775]: I1002 03:13:43.944022 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6223610a-fc09-44d4-9306-2b20384affeb" containerName="mariadb-database-create" Oct 02 03:13:43 crc kubenswrapper[4775]: I1002 03:13:43.944045 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b563b0be-d1f0-431b-9e23-39b5a55ffc12" containerName="mariadb-database-create" Oct 02 03:13:43 crc kubenswrapper[4775]: I1002 03:13:43.945167 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f7fd-account-create-29988" Oct 02 03:13:43 crc kubenswrapper[4775]: I1002 03:13:43.947210 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 02 03:13:43 crc kubenswrapper[4775]: I1002 03:13:43.949747 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f7fd-account-create-29988"] Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.095774 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6vpn\" (UniqueName: \"kubernetes.io/projected/7f38e9c6-c014-4516-947b-08247cdc35c7-kube-api-access-n6vpn\") pod \"nova-api-f7fd-account-create-29988\" (UID: \"7f38e9c6-c014-4516-947b-08247cdc35c7\") " pod="openstack/nova-api-f7fd-account-create-29988" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.143616 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-08e5-account-create-2rsjm"] Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.145713 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-08e5-account-create-2rsjm" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.148142 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.159314 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-08e5-account-create-2rsjm"] Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.197464 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6vpn\" (UniqueName: \"kubernetes.io/projected/7f38e9c6-c014-4516-947b-08247cdc35c7-kube-api-access-n6vpn\") pod \"nova-api-f7fd-account-create-29988\" (UID: \"7f38e9c6-c014-4516-947b-08247cdc35c7\") " pod="openstack/nova-api-f7fd-account-create-29988" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.232725 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6vpn\" (UniqueName: \"kubernetes.io/projected/7f38e9c6-c014-4516-947b-08247cdc35c7-kube-api-access-n6vpn\") pod \"nova-api-f7fd-account-create-29988\" (UID: \"7f38e9c6-c014-4516-947b-08247cdc35c7\") " pod="openstack/nova-api-f7fd-account-create-29988" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.299086 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8jjh\" (UniqueName: \"kubernetes.io/projected/63300bd0-8471-47ca-b748-ee6526e02957-kube-api-access-n8jjh\") pod \"nova-cell0-08e5-account-create-2rsjm\" (UID: \"63300bd0-8471-47ca-b748-ee6526e02957\") " pod="openstack/nova-cell0-08e5-account-create-2rsjm" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.311791 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f7fd-account-create-29988" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.342801 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-7669-account-create-lm5z5"] Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.344648 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7669-account-create-lm5z5" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.348751 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.375318 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7669-account-create-lm5z5"] Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.400588 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8jjh\" (UniqueName: \"kubernetes.io/projected/63300bd0-8471-47ca-b748-ee6526e02957-kube-api-access-n8jjh\") pod \"nova-cell0-08e5-account-create-2rsjm\" (UID: \"63300bd0-8471-47ca-b748-ee6526e02957\") " pod="openstack/nova-cell0-08e5-account-create-2rsjm" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.425754 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8jjh\" (UniqueName: \"kubernetes.io/projected/63300bd0-8471-47ca-b748-ee6526e02957-kube-api-access-n8jjh\") pod \"nova-cell0-08e5-account-create-2rsjm\" (UID: \"63300bd0-8471-47ca-b748-ee6526e02957\") " pod="openstack/nova-cell0-08e5-account-create-2rsjm" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.467902 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-08e5-account-create-2rsjm" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.502694 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs72k\" (UniqueName: \"kubernetes.io/projected/d122ae48-f8d4-46c3-bb0e-a890010f250a-kube-api-access-cs72k\") pod \"nova-cell1-7669-account-create-lm5z5\" (UID: \"d122ae48-f8d4-46c3-bb0e-a890010f250a\") " pod="openstack/nova-cell1-7669-account-create-lm5z5" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.604935 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs72k\" (UniqueName: \"kubernetes.io/projected/d122ae48-f8d4-46c3-bb0e-a890010f250a-kube-api-access-cs72k\") pod \"nova-cell1-7669-account-create-lm5z5\" (UID: \"d122ae48-f8d4-46c3-bb0e-a890010f250a\") " pod="openstack/nova-cell1-7669-account-create-lm5z5" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.629240 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs72k\" (UniqueName: \"kubernetes.io/projected/d122ae48-f8d4-46c3-bb0e-a890010f250a-kube-api-access-cs72k\") pod \"nova-cell1-7669-account-create-lm5z5\" (UID: \"d122ae48-f8d4-46c3-bb0e-a890010f250a\") " pod="openstack/nova-cell1-7669-account-create-lm5z5" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.757922 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7669-account-create-lm5z5" Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.806726 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f7fd-account-create-29988"] Oct 02 03:13:44 crc kubenswrapper[4775]: I1002 03:13:44.938818 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-08e5-account-create-2rsjm"] Oct 02 03:13:44 crc kubenswrapper[4775]: W1002 03:13:44.942676 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63300bd0_8471_47ca_b748_ee6526e02957.slice/crio-4725f5e023a3f2c4bb2f840d586f65ddf0729b43429ed2733db25a3282bd7aed WatchSource:0}: Error finding container 4725f5e023a3f2c4bb2f840d586f65ddf0729b43429ed2733db25a3282bd7aed: Status 404 returned error can't find the container with id 4725f5e023a3f2c4bb2f840d586f65ddf0729b43429ed2733db25a3282bd7aed Oct 02 03:13:45 crc kubenswrapper[4775]: I1002 03:13:45.169565 4775 generic.go:334] "Generic (PLEG): container finished" podID="7f38e9c6-c014-4516-947b-08247cdc35c7" containerID="ca8a1b9c60e3cb73943079a5feebcddd6e13a38d4652d5a7dd75f6d2f56800dc" exitCode=0 Oct 02 03:13:45 crc kubenswrapper[4775]: I1002 03:13:45.169642 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f7fd-account-create-29988" event={"ID":"7f38e9c6-c014-4516-947b-08247cdc35c7","Type":"ContainerDied","Data":"ca8a1b9c60e3cb73943079a5feebcddd6e13a38d4652d5a7dd75f6d2f56800dc"} Oct 02 03:13:45 crc kubenswrapper[4775]: I1002 03:13:45.169672 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f7fd-account-create-29988" event={"ID":"7f38e9c6-c014-4516-947b-08247cdc35c7","Type":"ContainerStarted","Data":"171c37f99a787b152ca17262bd0367257ca214f98687d462c77e80923cbf0f33"} Oct 02 03:13:45 crc kubenswrapper[4775]: I1002 03:13:45.176693 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-08e5-account-create-2rsjm" event={"ID":"63300bd0-8471-47ca-b748-ee6526e02957","Type":"ContainerStarted","Data":"4725f5e023a3f2c4bb2f840d586f65ddf0729b43429ed2733db25a3282bd7aed"} Oct 02 03:13:45 crc kubenswrapper[4775]: I1002 03:13:45.293837 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-7669-account-create-lm5z5"] Oct 02 03:13:45 crc kubenswrapper[4775]: W1002 03:13:45.355471 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd122ae48_f8d4_46c3_bb0e_a890010f250a.slice/crio-dd135200d5ce76fc779747fd112688a24006cddbc24cde04da651bdfa476a615 WatchSource:0}: Error finding container dd135200d5ce76fc779747fd112688a24006cddbc24cde04da651bdfa476a615: Status 404 returned error can't find the container with id dd135200d5ce76fc779747fd112688a24006cddbc24cde04da651bdfa476a615 Oct 02 03:13:46 crc kubenswrapper[4775]: I1002 03:13:46.193178 4775 generic.go:334] "Generic (PLEG): container finished" podID="d122ae48-f8d4-46c3-bb0e-a890010f250a" containerID="12e278188355751276f4519517fc8b44496a13bafdcb735332297fcc0afb18f3" exitCode=0 Oct 02 03:13:46 crc kubenswrapper[4775]: I1002 03:13:46.193451 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7669-account-create-lm5z5" event={"ID":"d122ae48-f8d4-46c3-bb0e-a890010f250a","Type":"ContainerDied","Data":"12e278188355751276f4519517fc8b44496a13bafdcb735332297fcc0afb18f3"} Oct 02 03:13:46 crc kubenswrapper[4775]: I1002 03:13:46.193514 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7669-account-create-lm5z5" event={"ID":"d122ae48-f8d4-46c3-bb0e-a890010f250a","Type":"ContainerStarted","Data":"dd135200d5ce76fc779747fd112688a24006cddbc24cde04da651bdfa476a615"} Oct 02 03:13:46 crc kubenswrapper[4775]: I1002 03:13:46.199829 4775 generic.go:334] "Generic (PLEG): container finished" podID="63300bd0-8471-47ca-b748-ee6526e02957" containerID="5f18572a1e2a6bf4b7db01b5cf5f77d2ded1434fe39bfaa5aca4a12897b7b110" exitCode=0 Oct 02 03:13:46 crc kubenswrapper[4775]: I1002 03:13:46.200161 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-08e5-account-create-2rsjm" event={"ID":"63300bd0-8471-47ca-b748-ee6526e02957","Type":"ContainerDied","Data":"5f18572a1e2a6bf4b7db01b5cf5f77d2ded1434fe39bfaa5aca4a12897b7b110"} Oct 02 03:13:46 crc kubenswrapper[4775]: I1002 03:13:46.669597 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f7fd-account-create-29988" Oct 02 03:13:46 crc kubenswrapper[4775]: I1002 03:13:46.746689 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6vpn\" (UniqueName: \"kubernetes.io/projected/7f38e9c6-c014-4516-947b-08247cdc35c7-kube-api-access-n6vpn\") pod \"7f38e9c6-c014-4516-947b-08247cdc35c7\" (UID: \"7f38e9c6-c014-4516-947b-08247cdc35c7\") " Oct 02 03:13:46 crc kubenswrapper[4775]: I1002 03:13:46.756630 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f38e9c6-c014-4516-947b-08247cdc35c7-kube-api-access-n6vpn" (OuterVolumeSpecName: "kube-api-access-n6vpn") pod "7f38e9c6-c014-4516-947b-08247cdc35c7" (UID: "7f38e9c6-c014-4516-947b-08247cdc35c7"). InnerVolumeSpecName "kube-api-access-n6vpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:13:46 crc kubenswrapper[4775]: I1002 03:13:46.849298 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6vpn\" (UniqueName: \"kubernetes.io/projected/7f38e9c6-c014-4516-947b-08247cdc35c7-kube-api-access-n6vpn\") on node \"crc\" DevicePath \"\"" Oct 02 03:13:47 crc kubenswrapper[4775]: I1002 03:13:47.216007 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f7fd-account-create-29988" event={"ID":"7f38e9c6-c014-4516-947b-08247cdc35c7","Type":"ContainerDied","Data":"171c37f99a787b152ca17262bd0367257ca214f98687d462c77e80923cbf0f33"} Oct 02 03:13:47 crc kubenswrapper[4775]: I1002 03:13:47.216094 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="171c37f99a787b152ca17262bd0367257ca214f98687d462c77e80923cbf0f33" Oct 02 03:13:47 crc kubenswrapper[4775]: I1002 03:13:47.216185 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f7fd-account-create-29988" Oct 02 03:13:47 crc kubenswrapper[4775]: I1002 03:13:47.675173 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-08e5-account-create-2rsjm" Oct 02 03:13:47 crc kubenswrapper[4775]: I1002 03:13:47.684582 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7669-account-create-lm5z5" Oct 02 03:13:47 crc kubenswrapper[4775]: I1002 03:13:47.765393 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n8jjh\" (UniqueName: \"kubernetes.io/projected/63300bd0-8471-47ca-b748-ee6526e02957-kube-api-access-n8jjh\") pod \"63300bd0-8471-47ca-b748-ee6526e02957\" (UID: \"63300bd0-8471-47ca-b748-ee6526e02957\") " Oct 02 03:13:47 crc kubenswrapper[4775]: I1002 03:13:47.765710 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cs72k\" (UniqueName: \"kubernetes.io/projected/d122ae48-f8d4-46c3-bb0e-a890010f250a-kube-api-access-cs72k\") pod \"d122ae48-f8d4-46c3-bb0e-a890010f250a\" (UID: \"d122ae48-f8d4-46c3-bb0e-a890010f250a\") " Oct 02 03:13:47 crc kubenswrapper[4775]: I1002 03:13:47.771041 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63300bd0-8471-47ca-b748-ee6526e02957-kube-api-access-n8jjh" (OuterVolumeSpecName: "kube-api-access-n8jjh") pod "63300bd0-8471-47ca-b748-ee6526e02957" (UID: "63300bd0-8471-47ca-b748-ee6526e02957"). InnerVolumeSpecName "kube-api-access-n8jjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:13:47 crc kubenswrapper[4775]: I1002 03:13:47.771762 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d122ae48-f8d4-46c3-bb0e-a890010f250a-kube-api-access-cs72k" (OuterVolumeSpecName: "kube-api-access-cs72k") pod "d122ae48-f8d4-46c3-bb0e-a890010f250a" (UID: "d122ae48-f8d4-46c3-bb0e-a890010f250a"). InnerVolumeSpecName "kube-api-access-cs72k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:13:47 crc kubenswrapper[4775]: I1002 03:13:47.868461 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cs72k\" (UniqueName: \"kubernetes.io/projected/d122ae48-f8d4-46c3-bb0e-a890010f250a-kube-api-access-cs72k\") on node \"crc\" DevicePath \"\"" Oct 02 03:13:47 crc kubenswrapper[4775]: I1002 03:13:47.868490 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n8jjh\" (UniqueName: \"kubernetes.io/projected/63300bd0-8471-47ca-b748-ee6526e02957-kube-api-access-n8jjh\") on node \"crc\" DevicePath \"\"" Oct 02 03:13:48 crc kubenswrapper[4775]: I1002 03:13:48.227296 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-7669-account-create-lm5z5" Oct 02 03:13:48 crc kubenswrapper[4775]: I1002 03:13:48.227337 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-7669-account-create-lm5z5" event={"ID":"d122ae48-f8d4-46c3-bb0e-a890010f250a","Type":"ContainerDied","Data":"dd135200d5ce76fc779747fd112688a24006cddbc24cde04da651bdfa476a615"} Oct 02 03:13:48 crc kubenswrapper[4775]: I1002 03:13:48.227407 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd135200d5ce76fc779747fd112688a24006cddbc24cde04da651bdfa476a615" Oct 02 03:13:48 crc kubenswrapper[4775]: I1002 03:13:48.230471 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-08e5-account-create-2rsjm" event={"ID":"63300bd0-8471-47ca-b748-ee6526e02957","Type":"ContainerDied","Data":"4725f5e023a3f2c4bb2f840d586f65ddf0729b43429ed2733db25a3282bd7aed"} Oct 02 03:13:48 crc kubenswrapper[4775]: I1002 03:13:48.230498 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4725f5e023a3f2c4bb2f840d586f65ddf0729b43429ed2733db25a3282bd7aed" Oct 02 03:13:48 crc kubenswrapper[4775]: I1002 03:13:48.230535 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-08e5-account-create-2rsjm" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.402326 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tfpzn"] Oct 02 03:13:49 crc kubenswrapper[4775]: E1002 03:13:49.402777 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d122ae48-f8d4-46c3-bb0e-a890010f250a" containerName="mariadb-account-create" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.402798 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d122ae48-f8d4-46c3-bb0e-a890010f250a" containerName="mariadb-account-create" Oct 02 03:13:49 crc kubenswrapper[4775]: E1002 03:13:49.402858 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f38e9c6-c014-4516-947b-08247cdc35c7" containerName="mariadb-account-create" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.402870 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f38e9c6-c014-4516-947b-08247cdc35c7" containerName="mariadb-account-create" Oct 02 03:13:49 crc kubenswrapper[4775]: E1002 03:13:49.402887 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63300bd0-8471-47ca-b748-ee6526e02957" containerName="mariadb-account-create" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.402898 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="63300bd0-8471-47ca-b748-ee6526e02957" containerName="mariadb-account-create" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.403207 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="63300bd0-8471-47ca-b748-ee6526e02957" containerName="mariadb-account-create" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.403242 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d122ae48-f8d4-46c3-bb0e-a890010f250a" containerName="mariadb-account-create" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.403258 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f38e9c6-c014-4516-947b-08247cdc35c7" containerName="mariadb-account-create" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.404296 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.406064 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.406585 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.411422 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bk4xq" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.425206 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tfpzn"] Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.507476 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-scripts\") pod \"nova-cell0-conductor-db-sync-tfpzn\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.507846 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l96zp\" (UniqueName: \"kubernetes.io/projected/7ecc60ed-4386-41a8-8827-421edf691ef5-kube-api-access-l96zp\") pod \"nova-cell0-conductor-db-sync-tfpzn\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.508004 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tfpzn\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.508138 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-config-data\") pod \"nova-cell0-conductor-db-sync-tfpzn\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.610403 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l96zp\" (UniqueName: \"kubernetes.io/projected/7ecc60ed-4386-41a8-8827-421edf691ef5-kube-api-access-l96zp\") pod \"nova-cell0-conductor-db-sync-tfpzn\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.610522 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tfpzn\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.610594 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-config-data\") pod \"nova-cell0-conductor-db-sync-tfpzn\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.610742 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-scripts\") pod \"nova-cell0-conductor-db-sync-tfpzn\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.635044 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-config-data\") pod \"nova-cell0-conductor-db-sync-tfpzn\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.635745 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-tfpzn\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.640376 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-scripts\") pod \"nova-cell0-conductor-db-sync-tfpzn\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.640437 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l96zp\" (UniqueName: \"kubernetes.io/projected/7ecc60ed-4386-41a8-8827-421edf691ef5-kube-api-access-l96zp\") pod \"nova-cell0-conductor-db-sync-tfpzn\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:49 crc kubenswrapper[4775]: I1002 03:13:49.726671 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:50 crc kubenswrapper[4775]: I1002 03:13:50.228026 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tfpzn"] Oct 02 03:13:50 crc kubenswrapper[4775]: I1002 03:13:50.257577 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tfpzn" event={"ID":"7ecc60ed-4386-41a8-8827-421edf691ef5","Type":"ContainerStarted","Data":"539ebf114084b4028755c403a8a02d19b347182133bcd901775391ee444396c2"} Oct 02 03:13:51 crc kubenswrapper[4775]: I1002 03:13:51.273278 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tfpzn" event={"ID":"7ecc60ed-4386-41a8-8827-421edf691ef5","Type":"ContainerStarted","Data":"4d6f5f14508cc10948f3ef5a7a4dbd1ff3ed6e25ed871567115099e112d9ceab"} Oct 02 03:13:51 crc kubenswrapper[4775]: I1002 03:13:51.309704 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-tfpzn" podStartSLOduration=2.309676513 podStartE2EDuration="2.309676513s" podCreationTimestamp="2025-10-02 03:13:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:13:51.298275018 +0000 UTC m=+5568.465019088" watchObservedRunningTime="2025-10-02 03:13:51.309676513 +0000 UTC m=+5568.476420593" Oct 02 03:13:55 crc kubenswrapper[4775]: I1002 03:13:55.320079 4775 generic.go:334] "Generic (PLEG): container finished" podID="7ecc60ed-4386-41a8-8827-421edf691ef5" containerID="4d6f5f14508cc10948f3ef5a7a4dbd1ff3ed6e25ed871567115099e112d9ceab" exitCode=0 Oct 02 03:13:55 crc kubenswrapper[4775]: I1002 03:13:55.320145 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tfpzn" event={"ID":"7ecc60ed-4386-41a8-8827-421edf691ef5","Type":"ContainerDied","Data":"4d6f5f14508cc10948f3ef5a7a4dbd1ff3ed6e25ed871567115099e112d9ceab"} Oct 02 03:13:56 crc kubenswrapper[4775]: I1002 03:13:56.797901 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:56 crc kubenswrapper[4775]: I1002 03:13:56.863539 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l96zp\" (UniqueName: \"kubernetes.io/projected/7ecc60ed-4386-41a8-8827-421edf691ef5-kube-api-access-l96zp\") pod \"7ecc60ed-4386-41a8-8827-421edf691ef5\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " Oct 02 03:13:56 crc kubenswrapper[4775]: I1002 03:13:56.864069 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-scripts\") pod \"7ecc60ed-4386-41a8-8827-421edf691ef5\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " Oct 02 03:13:56 crc kubenswrapper[4775]: I1002 03:13:56.864192 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-config-data\") pod \"7ecc60ed-4386-41a8-8827-421edf691ef5\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " Oct 02 03:13:56 crc kubenswrapper[4775]: I1002 03:13:56.864244 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-combined-ca-bundle\") pod \"7ecc60ed-4386-41a8-8827-421edf691ef5\" (UID: \"7ecc60ed-4386-41a8-8827-421edf691ef5\") " Oct 02 03:13:56 crc kubenswrapper[4775]: I1002 03:13:56.870186 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-scripts" (OuterVolumeSpecName: "scripts") pod "7ecc60ed-4386-41a8-8827-421edf691ef5" (UID: "7ecc60ed-4386-41a8-8827-421edf691ef5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:13:56 crc kubenswrapper[4775]: I1002 03:13:56.873473 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ecc60ed-4386-41a8-8827-421edf691ef5-kube-api-access-l96zp" (OuterVolumeSpecName: "kube-api-access-l96zp") pod "7ecc60ed-4386-41a8-8827-421edf691ef5" (UID: "7ecc60ed-4386-41a8-8827-421edf691ef5"). InnerVolumeSpecName "kube-api-access-l96zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:13:56 crc kubenswrapper[4775]: I1002 03:13:56.905636 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-config-data" (OuterVolumeSpecName: "config-data") pod "7ecc60ed-4386-41a8-8827-421edf691ef5" (UID: "7ecc60ed-4386-41a8-8827-421edf691ef5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:13:56 crc kubenswrapper[4775]: I1002 03:13:56.913153 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ecc60ed-4386-41a8-8827-421edf691ef5" (UID: "7ecc60ed-4386-41a8-8827-421edf691ef5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:13:56 crc kubenswrapper[4775]: I1002 03:13:56.966828 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l96zp\" (UniqueName: \"kubernetes.io/projected/7ecc60ed-4386-41a8-8827-421edf691ef5-kube-api-access-l96zp\") on node \"crc\" DevicePath \"\"" Oct 02 03:13:56 crc kubenswrapper[4775]: I1002 03:13:56.966918 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:13:56 crc kubenswrapper[4775]: I1002 03:13:56.966942 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:13:56 crc kubenswrapper[4775]: I1002 03:13:56.967009 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ecc60ed-4386-41a8-8827-421edf691ef5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.348923 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-tfpzn" event={"ID":"7ecc60ed-4386-41a8-8827-421edf691ef5","Type":"ContainerDied","Data":"539ebf114084b4028755c403a8a02d19b347182133bcd901775391ee444396c2"} Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.349032 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="539ebf114084b4028755c403a8a02d19b347182133bcd901775391ee444396c2" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.349046 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-tfpzn" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.462169 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 03:13:57 crc kubenswrapper[4775]: E1002 03:13:57.462893 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ecc60ed-4386-41a8-8827-421edf691ef5" containerName="nova-cell0-conductor-db-sync" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.463052 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ecc60ed-4386-41a8-8827-421edf691ef5" containerName="nova-cell0-conductor-db-sync" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.463506 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ecc60ed-4386-41a8-8827-421edf691ef5" containerName="nova-cell0-conductor-db-sync" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.464619 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.469168 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-bk4xq" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.469831 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.472508 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.577978 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc4855b-083c-44be-bfa6-af98a1c963fa-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5fc4855b-083c-44be-bfa6-af98a1c963fa\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.578096 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btth6\" (UniqueName: \"kubernetes.io/projected/5fc4855b-083c-44be-bfa6-af98a1c963fa-kube-api-access-btth6\") pod \"nova-cell0-conductor-0\" (UID: \"5fc4855b-083c-44be-bfa6-af98a1c963fa\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.578162 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc4855b-083c-44be-bfa6-af98a1c963fa-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5fc4855b-083c-44be-bfa6-af98a1c963fa\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.679802 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc4855b-083c-44be-bfa6-af98a1c963fa-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5fc4855b-083c-44be-bfa6-af98a1c963fa\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.679941 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc4855b-083c-44be-bfa6-af98a1c963fa-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5fc4855b-083c-44be-bfa6-af98a1c963fa\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.680027 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btth6\" (UniqueName: \"kubernetes.io/projected/5fc4855b-083c-44be-bfa6-af98a1c963fa-kube-api-access-btth6\") pod \"nova-cell0-conductor-0\" (UID: \"5fc4855b-083c-44be-bfa6-af98a1c963fa\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.684699 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc4855b-083c-44be-bfa6-af98a1c963fa-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5fc4855b-083c-44be-bfa6-af98a1c963fa\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.685644 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc4855b-083c-44be-bfa6-af98a1c963fa-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5fc4855b-083c-44be-bfa6-af98a1c963fa\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.713274 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btth6\" (UniqueName: \"kubernetes.io/projected/5fc4855b-083c-44be-bfa6-af98a1c963fa-kube-api-access-btth6\") pod \"nova-cell0-conductor-0\" (UID: \"5fc4855b-083c-44be-bfa6-af98a1c963fa\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:13:57 crc kubenswrapper[4775]: I1002 03:13:57.814131 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 03:13:58 crc kubenswrapper[4775]: I1002 03:13:58.296225 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 03:13:58 crc kubenswrapper[4775]: I1002 03:13:58.365413 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5fc4855b-083c-44be-bfa6-af98a1c963fa","Type":"ContainerStarted","Data":"08138d89b7df92c281d77c220e70ed8bd0664210a29da1a91772d46308beb8a3"} Oct 02 03:13:59 crc kubenswrapper[4775]: I1002 03:13:59.381566 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5fc4855b-083c-44be-bfa6-af98a1c963fa","Type":"ContainerStarted","Data":"72b4aa11ce93c390e5656b40ad1314a9b3d9d8fe52071d2d3ac980a62dd975d9"} Oct 02 03:13:59 crc kubenswrapper[4775]: I1002 03:13:59.381866 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 02 03:13:59 crc kubenswrapper[4775]: I1002 03:13:59.418165 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.418138352 podStartE2EDuration="2.418138352s" podCreationTimestamp="2025-10-02 03:13:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:13:59.400216488 +0000 UTC m=+5576.566960588" watchObservedRunningTime="2025-10-02 03:13:59.418138352 +0000 UTC m=+5576.584882422" Oct 02 03:14:07 crc kubenswrapper[4775]: I1002 03:14:07.862908 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.471680 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-ft765"] Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.473010 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.476597 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.477216 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.495342 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-ft765"] Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.612653 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ft765\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.612709 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-config-data\") pod \"nova-cell0-cell-mapping-ft765\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.612760 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrvsq\" (UniqueName: \"kubernetes.io/projected/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-kube-api-access-vrvsq\") pod \"nova-cell0-cell-mapping-ft765\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.612836 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-scripts\") pod \"nova-cell0-cell-mapping-ft765\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.630620 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.632034 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.637408 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.640411 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.641845 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.644583 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.650706 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.663258 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.672220 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.679347 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.687256 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.700703 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.714340 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cfd6601-23fa-48fe-8a5b-8b070233f030-logs\") pod \"nova-metadata-0\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " pod="openstack/nova-metadata-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.714383 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-logs\") pod \"nova-api-0\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " pod="openstack/nova-api-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.714416 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cfd6601-23fa-48fe-8a5b-8b070233f030-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " pod="openstack/nova-metadata-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.714436 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-config-data\") pod \"nova-api-0\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " pod="openstack/nova-api-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.714460 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdp8g\" (UniqueName: \"kubernetes.io/projected/5cfd6601-23fa-48fe-8a5b-8b070233f030-kube-api-access-gdp8g\") pod \"nova-metadata-0\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " pod="openstack/nova-metadata-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.714504 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ft765\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.714525 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-config-data\") pod \"nova-cell0-cell-mapping-ft765\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.714543 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cfd6601-23fa-48fe-8a5b-8b070233f030-config-data\") pod \"nova-metadata-0\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " pod="openstack/nova-metadata-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.714578 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrvsq\" (UniqueName: \"kubernetes.io/projected/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-kube-api-access-vrvsq\") pod \"nova-cell0-cell-mapping-ft765\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.714608 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " pod="openstack/nova-api-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.714648 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-scripts\") pod \"nova-cell0-cell-mapping-ft765\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.714667 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6qf2\" (UniqueName: \"kubernetes.io/projected/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-kube-api-access-g6qf2\") pod \"nova-api-0\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " pod="openstack/nova-api-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.727551 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-584bf664f7-7fqhb"] Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.728394 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-config-data\") pod \"nova-cell0-cell-mapping-ft765\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.729561 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ft765\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.732414 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-scripts\") pod \"nova-cell0-cell-mapping-ft765\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.748834 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-584bf664f7-7fqhb"] Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.748967 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.761182 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrvsq\" (UniqueName: \"kubernetes.io/projected/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-kube-api-access-vrvsq\") pod \"nova-cell0-cell-mapping-ft765\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.798172 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.825667 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-logs\") pod \"nova-api-0\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " pod="openstack/nova-api-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.825718 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de4c6de6-bd34-4316-9791-b360aaf48066-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"de4c6de6-bd34-4316-9791-b360aaf48066\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.825751 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cfd6601-23fa-48fe-8a5b-8b070233f030-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " pod="openstack/nova-metadata-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.825775 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-config-data\") pod \"nova-api-0\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " pod="openstack/nova-api-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.825796 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-ovsdbserver-nb\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.825819 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdp8g\" (UniqueName: \"kubernetes.io/projected/5cfd6601-23fa-48fe-8a5b-8b070233f030-kube-api-access-gdp8g\") pod \"nova-metadata-0\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " pod="openstack/nova-metadata-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.825856 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl95h\" (UniqueName: \"kubernetes.io/projected/de4c6de6-bd34-4316-9791-b360aaf48066-kube-api-access-tl95h\") pod \"nova-cell1-novncproxy-0\" (UID: \"de4c6de6-bd34-4316-9791-b360aaf48066\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.825880 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cfd6601-23fa-48fe-8a5b-8b070233f030-config-data\") pod \"nova-metadata-0\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " pod="openstack/nova-metadata-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.825910 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-config\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.825930 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4c6de6-bd34-4316-9791-b360aaf48066-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"de4c6de6-bd34-4316-9791-b360aaf48066\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.825984 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmtt4\" (UniqueName: \"kubernetes.io/projected/8473d923-d46f-40e2-9f75-58e14833dadd-kube-api-access-vmtt4\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.826014 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-dns-svc\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.826036 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " pod="openstack/nova-api-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.826069 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-ovsdbserver-sb\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.826097 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6qf2\" (UniqueName: \"kubernetes.io/projected/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-kube-api-access-g6qf2\") pod \"nova-api-0\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " pod="openstack/nova-api-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.826129 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cfd6601-23fa-48fe-8a5b-8b070233f030-logs\") pod \"nova-metadata-0\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " pod="openstack/nova-metadata-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.826487 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cfd6601-23fa-48fe-8a5b-8b070233f030-logs\") pod \"nova-metadata-0\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " pod="openstack/nova-metadata-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.826537 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-logs\") pod \"nova-api-0\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " pod="openstack/nova-api-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.829517 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.830697 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.834588 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-config-data\") pod \"nova-api-0\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " pod="openstack/nova-api-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.835082 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " pod="openstack/nova-api-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.835403 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.835920 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cfd6601-23fa-48fe-8a5b-8b070233f030-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " pod="openstack/nova-metadata-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.840362 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.840730 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cfd6601-23fa-48fe-8a5b-8b070233f030-config-data\") pod \"nova-metadata-0\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " pod="openstack/nova-metadata-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.851038 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6qf2\" (UniqueName: \"kubernetes.io/projected/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-kube-api-access-g6qf2\") pod \"nova-api-0\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " pod="openstack/nova-api-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.853614 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdp8g\" (UniqueName: \"kubernetes.io/projected/5cfd6601-23fa-48fe-8a5b-8b070233f030-kube-api-access-gdp8g\") pod \"nova-metadata-0\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " pod="openstack/nova-metadata-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.928799 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-dns-svc\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.928882 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-ovsdbserver-sb\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.928977 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de4c6de6-bd34-4316-9791-b360aaf48066-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"de4c6de6-bd34-4316-9791-b360aaf48066\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.929026 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-ovsdbserver-nb\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.929089 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02736f53-7af1-41f4-ae9b-df12578b7cfa-config-data\") pod \"nova-scheduler-0\" (UID: \"02736f53-7af1-41f4-ae9b-df12578b7cfa\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.929132 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl95h\" (UniqueName: \"kubernetes.io/projected/de4c6de6-bd34-4316-9791-b360aaf48066-kube-api-access-tl95h\") pod \"nova-cell1-novncproxy-0\" (UID: \"de4c6de6-bd34-4316-9791-b360aaf48066\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.929164 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02736f53-7af1-41f4-ae9b-df12578b7cfa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"02736f53-7af1-41f4-ae9b-df12578b7cfa\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.929189 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv8kx\" (UniqueName: \"kubernetes.io/projected/02736f53-7af1-41f4-ae9b-df12578b7cfa-kube-api-access-jv8kx\") pod \"nova-scheduler-0\" (UID: \"02736f53-7af1-41f4-ae9b-df12578b7cfa\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.929218 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-config\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.929258 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4c6de6-bd34-4316-9791-b360aaf48066-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"de4c6de6-bd34-4316-9791-b360aaf48066\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.929290 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmtt4\" (UniqueName: \"kubernetes.io/projected/8473d923-d46f-40e2-9f75-58e14833dadd-kube-api-access-vmtt4\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.930579 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-dns-svc\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.930936 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-ovsdbserver-sb\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.931979 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-ovsdbserver-nb\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.932240 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-config\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.935018 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de4c6de6-bd34-4316-9791-b360aaf48066-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"de4c6de6-bd34-4316-9791-b360aaf48066\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.946099 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.947499 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmtt4\" (UniqueName: \"kubernetes.io/projected/8473d923-d46f-40e2-9f75-58e14833dadd-kube-api-access-vmtt4\") pod \"dnsmasq-dns-584bf664f7-7fqhb\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.952403 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4c6de6-bd34-4316-9791-b360aaf48066-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"de4c6de6-bd34-4316-9791-b360aaf48066\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.954250 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl95h\" (UniqueName: \"kubernetes.io/projected/de4c6de6-bd34-4316-9791-b360aaf48066-kube-api-access-tl95h\") pod \"nova-cell1-novncproxy-0\" (UID: \"de4c6de6-bd34-4316-9791-b360aaf48066\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.960791 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 03:14:08 crc kubenswrapper[4775]: I1002 03:14:08.996950 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.031906 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02736f53-7af1-41f4-ae9b-df12578b7cfa-config-data\") pod \"nova-scheduler-0\" (UID: \"02736f53-7af1-41f4-ae9b-df12578b7cfa\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.032001 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02736f53-7af1-41f4-ae9b-df12578b7cfa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"02736f53-7af1-41f4-ae9b-df12578b7cfa\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.032047 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv8kx\" (UniqueName: \"kubernetes.io/projected/02736f53-7af1-41f4-ae9b-df12578b7cfa-kube-api-access-jv8kx\") pod \"nova-scheduler-0\" (UID: \"02736f53-7af1-41f4-ae9b-df12578b7cfa\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.036265 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02736f53-7af1-41f4-ae9b-df12578b7cfa-config-data\") pod \"nova-scheduler-0\" (UID: \"02736f53-7af1-41f4-ae9b-df12578b7cfa\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.037214 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02736f53-7af1-41f4-ae9b-df12578b7cfa-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"02736f53-7af1-41f4-ae9b-df12578b7cfa\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.046646 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv8kx\" (UniqueName: \"kubernetes.io/projected/02736f53-7af1-41f4-ae9b-df12578b7cfa-kube-api-access-jv8kx\") pod \"nova-scheduler-0\" (UID: \"02736f53-7af1-41f4-ae9b-df12578b7cfa\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.127356 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.214322 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.279566 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-ft765"] Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.423667 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:09 crc kubenswrapper[4775]: W1002 03:14:09.426921 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7df5bdbc_ea9c_4afb_b7f1_83b3f13fc3c1.slice/crio-79b72b4afbc99dbe434761c2ec3ef9fcd7997318b9b31f66db1d11424724a6a6 WatchSource:0}: Error finding container 79b72b4afbc99dbe434761c2ec3ef9fcd7997318b9b31f66db1d11424724a6a6: Status 404 returned error can't find the container with id 79b72b4afbc99dbe434761c2ec3ef9fcd7997318b9b31f66db1d11424724a6a6 Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.496066 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1","Type":"ContainerStarted","Data":"79b72b4afbc99dbe434761c2ec3ef9fcd7997318b9b31f66db1d11424724a6a6"} Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.497821 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ft765" event={"ID":"9a0025f0-7fa5-4798-b9ec-09a3cec748d7","Type":"ContainerStarted","Data":"da66786352e16c468f81770ba8f2b38ef27c81815063b0092ea1377dfdf3607c"} Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.539178 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.559111 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l7nrv"] Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.560358 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.563465 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.563635 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.566604 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.571544 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l7nrv"] Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.645143 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qtsk\" (UniqueName: \"kubernetes.io/projected/c2543f41-75cd-42d5-9f28-9b7c5f73756c-kube-api-access-5qtsk\") pod \"nova-cell1-conductor-db-sync-l7nrv\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.645197 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-l7nrv\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.645227 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-config-data\") pod \"nova-cell1-conductor-db-sync-l7nrv\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.645272 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-scripts\") pod \"nova-cell1-conductor-db-sync-l7nrv\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.736438 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.749296 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qtsk\" (UniqueName: \"kubernetes.io/projected/c2543f41-75cd-42d5-9f28-9b7c5f73756c-kube-api-access-5qtsk\") pod \"nova-cell1-conductor-db-sync-l7nrv\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.749381 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-l7nrv\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.749420 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-config-data\") pod \"nova-cell1-conductor-db-sync-l7nrv\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.749477 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-scripts\") pod \"nova-cell1-conductor-db-sync-l7nrv\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.759590 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-584bf664f7-7fqhb"] Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.766083 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qtsk\" (UniqueName: \"kubernetes.io/projected/c2543f41-75cd-42d5-9f28-9b7c5f73756c-kube-api-access-5qtsk\") pod \"nova-cell1-conductor-db-sync-l7nrv\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.768597 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-l7nrv\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.770131 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-scripts\") pod \"nova-cell1-conductor-db-sync-l7nrv\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.776438 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-config-data\") pod \"nova-cell1-conductor-db-sync-l7nrv\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:09 crc kubenswrapper[4775]: W1002 03:14:09.788784 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8473d923_d46f_40e2_9f75_58e14833dadd.slice/crio-07563fa4313842d2f16c1ea76155e4c9c1fafca77d7e19362b794863218da10c WatchSource:0}: Error finding container 07563fa4313842d2f16c1ea76155e4c9c1fafca77d7e19362b794863218da10c: Status 404 returned error can't find the container with id 07563fa4313842d2f16c1ea76155e4c9c1fafca77d7e19362b794863218da10c Oct 02 03:14:09 crc kubenswrapper[4775]: I1002 03:14:09.927766 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.363660 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l7nrv"] Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.510358 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"de4c6de6-bd34-4316-9791-b360aaf48066","Type":"ContainerStarted","Data":"253c2b542479ae7d61337ab667bbb0c152706bf9aa25f6a81c70f8682aa874d9"} Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.510399 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"de4c6de6-bd34-4316-9791-b360aaf48066","Type":"ContainerStarted","Data":"7f01083fed77a27cf203b805a21f4129fe977f8d068c64fc929032b2972e4970"} Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.514819 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1","Type":"ContainerStarted","Data":"1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9"} Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.514847 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1","Type":"ContainerStarted","Data":"ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c"} Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.518407 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cfd6601-23fa-48fe-8a5b-8b070233f030","Type":"ContainerStarted","Data":"c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18"} Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.518453 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cfd6601-23fa-48fe-8a5b-8b070233f030","Type":"ContainerStarted","Data":"3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a"} Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.518462 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cfd6601-23fa-48fe-8a5b-8b070233f030","Type":"ContainerStarted","Data":"f25f9fe413f3857a28f549206ad28536f36cfb0cf7dc631e17603ea451f3abcf"} Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.520287 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l7nrv" event={"ID":"c2543f41-75cd-42d5-9f28-9b7c5f73756c","Type":"ContainerStarted","Data":"bdf48c9317de08fc48ce3b674e10f7f953a4d9b1bb86023a778285b312be0195"} Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.521658 4775 generic.go:334] "Generic (PLEG): container finished" podID="8473d923-d46f-40e2-9f75-58e14833dadd" containerID="7d10971f75f7a0ea376bc95cdd84d53c288f132218505d3ac1eab456caabfd77" exitCode=0 Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.521695 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" event={"ID":"8473d923-d46f-40e2-9f75-58e14833dadd","Type":"ContainerDied","Data":"7d10971f75f7a0ea376bc95cdd84d53c288f132218505d3ac1eab456caabfd77"} Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.521710 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" event={"ID":"8473d923-d46f-40e2-9f75-58e14833dadd","Type":"ContainerStarted","Data":"07563fa4313842d2f16c1ea76155e4c9c1fafca77d7e19362b794863218da10c"} Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.530387 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.530368724 podStartE2EDuration="2.530368724s" podCreationTimestamp="2025-10-02 03:14:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:10.525806546 +0000 UTC m=+5587.692550586" watchObservedRunningTime="2025-10-02 03:14:10.530368724 +0000 UTC m=+5587.697112764" Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.543007 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ft765" event={"ID":"9a0025f0-7fa5-4798-b9ec-09a3cec748d7","Type":"ContainerStarted","Data":"ef652e1657b9f30f838a21acdcb8147d59deae3ab9a424a851ee7dffcd018736"} Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.559537 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.559517569 podStartE2EDuration="2.559517569s" podCreationTimestamp="2025-10-02 03:14:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:10.55067492 +0000 UTC m=+5587.717418970" watchObservedRunningTime="2025-10-02 03:14:10.559517569 +0000 UTC m=+5587.726261609" Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.562111 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02736f53-7af1-41f4-ae9b-df12578b7cfa","Type":"ContainerStarted","Data":"da66341123db38c6d3572988cbfc6a6a6ac4d088fec31963fa8b35ca27c38051"} Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.562154 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02736f53-7af1-41f4-ae9b-df12578b7cfa","Type":"ContainerStarted","Data":"70f2f05411e0c87b45711c90d290ab5899f78e7554da4612c17c2b2daf87e46a"} Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.600970 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.600938433 podStartE2EDuration="2.600938433s" podCreationTimestamp="2025-10-02 03:14:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:10.582074194 +0000 UTC m=+5587.748818244" watchObservedRunningTime="2025-10-02 03:14:10.600938433 +0000 UTC m=+5587.767682473" Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.611213 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-ft765" podStartSLOduration=2.611196028 podStartE2EDuration="2.611196028s" podCreationTimestamp="2025-10-02 03:14:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:10.597924904 +0000 UTC m=+5587.764668944" watchObservedRunningTime="2025-10-02 03:14:10.611196028 +0000 UTC m=+5587.777940058" Oct 02 03:14:10 crc kubenswrapper[4775]: I1002 03:14:10.619197 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.619181445 podStartE2EDuration="2.619181445s" podCreationTimestamp="2025-10-02 03:14:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:10.611823965 +0000 UTC m=+5587.778568005" watchObservedRunningTime="2025-10-02 03:14:10.619181445 +0000 UTC m=+5587.785925485" Oct 02 03:14:11 crc kubenswrapper[4775]: I1002 03:14:11.469622 4775 scope.go:117] "RemoveContainer" containerID="b152a10dfa6bd2a9f1aa43eaf12766e437aeb3db19855b25781a30df5c8d960e" Oct 02 03:14:11 crc kubenswrapper[4775]: I1002 03:14:11.603389 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l7nrv" event={"ID":"c2543f41-75cd-42d5-9f28-9b7c5f73756c","Type":"ContainerStarted","Data":"a3bd73887099309942bafcb6324a4078b20d9d9f62ba786f3e1506b9ac2d3ee4"} Oct 02 03:14:11 crc kubenswrapper[4775]: I1002 03:14:11.613850 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" event={"ID":"8473d923-d46f-40e2-9f75-58e14833dadd","Type":"ContainerStarted","Data":"bae7647a6403874672d4069f85fa0fc18c870bf6bea9eaa4a263f8242571c0a9"} Oct 02 03:14:11 crc kubenswrapper[4775]: I1002 03:14:11.614598 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:11 crc kubenswrapper[4775]: I1002 03:14:11.640782 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-l7nrv" podStartSLOduration=2.640720582 podStartE2EDuration="2.640720582s" podCreationTimestamp="2025-10-02 03:14:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:11.625687402 +0000 UTC m=+5588.792431442" watchObservedRunningTime="2025-10-02 03:14:11.640720582 +0000 UTC m=+5588.807464632" Oct 02 03:14:11 crc kubenswrapper[4775]: I1002 03:14:11.665829 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" podStartSLOduration=3.665807602 podStartE2EDuration="3.665807602s" podCreationTimestamp="2025-10-02 03:14:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:11.65184666 +0000 UTC m=+5588.818590700" watchObservedRunningTime="2025-10-02 03:14:11.665807602 +0000 UTC m=+5588.832551652" Oct 02 03:14:13 crc kubenswrapper[4775]: I1002 03:14:13.637934 4775 generic.go:334] "Generic (PLEG): container finished" podID="c2543f41-75cd-42d5-9f28-9b7c5f73756c" containerID="a3bd73887099309942bafcb6324a4078b20d9d9f62ba786f3e1506b9ac2d3ee4" exitCode=0 Oct 02 03:14:13 crc kubenswrapper[4775]: I1002 03:14:13.638020 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l7nrv" event={"ID":"c2543f41-75cd-42d5-9f28-9b7c5f73756c","Type":"ContainerDied","Data":"a3bd73887099309942bafcb6324a4078b20d9d9f62ba786f3e1506b9ac2d3ee4"} Oct 02 03:14:13 crc kubenswrapper[4775]: I1002 03:14:13.961486 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 03:14:13 crc kubenswrapper[4775]: I1002 03:14:13.961556 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 03:14:13 crc kubenswrapper[4775]: I1002 03:14:13.998295 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:14 crc kubenswrapper[4775]: I1002 03:14:14.215161 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 03:14:14 crc kubenswrapper[4775]: I1002 03:14:14.653520 4775 generic.go:334] "Generic (PLEG): container finished" podID="9a0025f0-7fa5-4798-b9ec-09a3cec748d7" containerID="ef652e1657b9f30f838a21acdcb8147d59deae3ab9a424a851ee7dffcd018736" exitCode=0 Oct 02 03:14:14 crc kubenswrapper[4775]: I1002 03:14:14.653585 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ft765" event={"ID":"9a0025f0-7fa5-4798-b9ec-09a3cec748d7","Type":"ContainerDied","Data":"ef652e1657b9f30f838a21acdcb8147d59deae3ab9a424a851ee7dffcd018736"} Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.136904 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.267040 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-scripts\") pod \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.267127 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-config-data\") pod \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.267159 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-combined-ca-bundle\") pod \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.267266 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qtsk\" (UniqueName: \"kubernetes.io/projected/c2543f41-75cd-42d5-9f28-9b7c5f73756c-kube-api-access-5qtsk\") pod \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\" (UID: \"c2543f41-75cd-42d5-9f28-9b7c5f73756c\") " Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.283862 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-scripts" (OuterVolumeSpecName: "scripts") pod "c2543f41-75cd-42d5-9f28-9b7c5f73756c" (UID: "c2543f41-75cd-42d5-9f28-9b7c5f73756c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.284029 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2543f41-75cd-42d5-9f28-9b7c5f73756c-kube-api-access-5qtsk" (OuterVolumeSpecName: "kube-api-access-5qtsk") pod "c2543f41-75cd-42d5-9f28-9b7c5f73756c" (UID: "c2543f41-75cd-42d5-9f28-9b7c5f73756c"). InnerVolumeSpecName "kube-api-access-5qtsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.301425 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c2543f41-75cd-42d5-9f28-9b7c5f73756c" (UID: "c2543f41-75cd-42d5-9f28-9b7c5f73756c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.303998 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-config-data" (OuterVolumeSpecName: "config-data") pod "c2543f41-75cd-42d5-9f28-9b7c5f73756c" (UID: "c2543f41-75cd-42d5-9f28-9b7c5f73756c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.369430 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.369469 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.369481 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2543f41-75cd-42d5-9f28-9b7c5f73756c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.369492 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qtsk\" (UniqueName: \"kubernetes.io/projected/c2543f41-75cd-42d5-9f28-9b7c5f73756c-kube-api-access-5qtsk\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.670151 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-l7nrv" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.670193 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-l7nrv" event={"ID":"c2543f41-75cd-42d5-9f28-9b7c5f73756c","Type":"ContainerDied","Data":"bdf48c9317de08fc48ce3b674e10f7f953a4d9b1bb86023a778285b312be0195"} Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.670254 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdf48c9317de08fc48ce3b674e10f7f953a4d9b1bb86023a778285b312be0195" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.796454 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 03:14:15 crc kubenswrapper[4775]: E1002 03:14:15.796823 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2543f41-75cd-42d5-9f28-9b7c5f73756c" containerName="nova-cell1-conductor-db-sync" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.796838 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2543f41-75cd-42d5-9f28-9b7c5f73756c" containerName="nova-cell1-conductor-db-sync" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.797106 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2543f41-75cd-42d5-9f28-9b7c5f73756c" containerName="nova-cell1-conductor-db-sync" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.797870 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.805770 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.815772 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.882946 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3f9270-7d6a-4b91-bf50-944388836500-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ca3f9270-7d6a-4b91-bf50-944388836500\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.883155 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3f9270-7d6a-4b91-bf50-944388836500-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ca3f9270-7d6a-4b91-bf50-944388836500\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.883183 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlpwq\" (UniqueName: \"kubernetes.io/projected/ca3f9270-7d6a-4b91-bf50-944388836500-kube-api-access-hlpwq\") pod \"nova-cell1-conductor-0\" (UID: \"ca3f9270-7d6a-4b91-bf50-944388836500\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.984846 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3f9270-7d6a-4b91-bf50-944388836500-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ca3f9270-7d6a-4b91-bf50-944388836500\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.985322 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3f9270-7d6a-4b91-bf50-944388836500-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ca3f9270-7d6a-4b91-bf50-944388836500\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.985349 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlpwq\" (UniqueName: \"kubernetes.io/projected/ca3f9270-7d6a-4b91-bf50-944388836500-kube-api-access-hlpwq\") pod \"nova-cell1-conductor-0\" (UID: \"ca3f9270-7d6a-4b91-bf50-944388836500\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.992736 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3f9270-7d6a-4b91-bf50-944388836500-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"ca3f9270-7d6a-4b91-bf50-944388836500\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:14:15 crc kubenswrapper[4775]: I1002 03:14:15.993199 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3f9270-7d6a-4b91-bf50-944388836500-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"ca3f9270-7d6a-4b91-bf50-944388836500\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.002540 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlpwq\" (UniqueName: \"kubernetes.io/projected/ca3f9270-7d6a-4b91-bf50-944388836500-kube-api-access-hlpwq\") pod \"nova-cell1-conductor-0\" (UID: \"ca3f9270-7d6a-4b91-bf50-944388836500\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.102886 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.122228 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.188742 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrvsq\" (UniqueName: \"kubernetes.io/projected/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-kube-api-access-vrvsq\") pod \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.188868 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-config-data\") pod \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.188916 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-combined-ca-bundle\") pod \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.188999 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-scripts\") pod \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\" (UID: \"9a0025f0-7fa5-4798-b9ec-09a3cec748d7\") " Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.194879 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-scripts" (OuterVolumeSpecName: "scripts") pod "9a0025f0-7fa5-4798-b9ec-09a3cec748d7" (UID: "9a0025f0-7fa5-4798-b9ec-09a3cec748d7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.197122 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-kube-api-access-vrvsq" (OuterVolumeSpecName: "kube-api-access-vrvsq") pod "9a0025f0-7fa5-4798-b9ec-09a3cec748d7" (UID: "9a0025f0-7fa5-4798-b9ec-09a3cec748d7"). InnerVolumeSpecName "kube-api-access-vrvsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.216047 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-config-data" (OuterVolumeSpecName: "config-data") pod "9a0025f0-7fa5-4798-b9ec-09a3cec748d7" (UID: "9a0025f0-7fa5-4798-b9ec-09a3cec748d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.218241 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a0025f0-7fa5-4798-b9ec-09a3cec748d7" (UID: "9a0025f0-7fa5-4798-b9ec-09a3cec748d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.292546 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrvsq\" (UniqueName: \"kubernetes.io/projected/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-kube-api-access-vrvsq\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.292599 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.292620 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.292636 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a0025f0-7fa5-4798-b9ec-09a3cec748d7-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.690353 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ft765" event={"ID":"9a0025f0-7fa5-4798-b9ec-09a3cec748d7","Type":"ContainerDied","Data":"da66786352e16c468f81770ba8f2b38ef27c81815063b0092ea1377dfdf3607c"} Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.690701 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da66786352e16c468f81770ba8f2b38ef27c81815063b0092ea1377dfdf3607c" Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.690467 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ft765" Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.715457 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.838328 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.838616 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" containerName="nova-api-log" containerID="cri-o://ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c" gracePeriod=30 Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.838803 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" containerName="nova-api-api" containerID="cri-o://1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9" gracePeriod=30 Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.922279 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.922530 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="02736f53-7af1-41f4-ae9b-df12578b7cfa" containerName="nova-scheduler-scheduler" containerID="cri-o://da66341123db38c6d3572988cbfc6a6a6ac4d088fec31963fa8b35ca27c38051" gracePeriod=30 Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.932764 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.933012 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5cfd6601-23fa-48fe-8a5b-8b070233f030" containerName="nova-metadata-log" containerID="cri-o://3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a" gracePeriod=30 Oct 02 03:14:16 crc kubenswrapper[4775]: I1002 03:14:16.933158 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5cfd6601-23fa-48fe-8a5b-8b070233f030" containerName="nova-metadata-metadata" containerID="cri-o://c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18" gracePeriod=30 Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.402940 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.519739 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-combined-ca-bundle\") pod \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.519908 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-config-data\") pod \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.520057 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-logs\") pod \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.520119 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6qf2\" (UniqueName: \"kubernetes.io/projected/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-kube-api-access-g6qf2\") pod \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\" (UID: \"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1\") " Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.520783 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-logs" (OuterVolumeSpecName: "logs") pod "7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" (UID: "7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.535173 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-kube-api-access-g6qf2" (OuterVolumeSpecName: "kube-api-access-g6qf2") pod "7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" (UID: "7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1"). InnerVolumeSpecName "kube-api-access-g6qf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.539749 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.543242 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" (UID: "7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.552127 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-config-data" (OuterVolumeSpecName: "config-data") pod "7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" (UID: "7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.621747 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdp8g\" (UniqueName: \"kubernetes.io/projected/5cfd6601-23fa-48fe-8a5b-8b070233f030-kube-api-access-gdp8g\") pod \"5cfd6601-23fa-48fe-8a5b-8b070233f030\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.621915 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cfd6601-23fa-48fe-8a5b-8b070233f030-logs\") pod \"5cfd6601-23fa-48fe-8a5b-8b070233f030\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.621948 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cfd6601-23fa-48fe-8a5b-8b070233f030-combined-ca-bundle\") pod \"5cfd6601-23fa-48fe-8a5b-8b070233f030\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.622005 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cfd6601-23fa-48fe-8a5b-8b070233f030-config-data\") pod \"5cfd6601-23fa-48fe-8a5b-8b070233f030\" (UID: \"5cfd6601-23fa-48fe-8a5b-8b070233f030\") " Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.622366 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.622376 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.622386 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.622394 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6qf2\" (UniqueName: \"kubernetes.io/projected/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1-kube-api-access-g6qf2\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.622865 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cfd6601-23fa-48fe-8a5b-8b070233f030-logs" (OuterVolumeSpecName: "logs") pod "5cfd6601-23fa-48fe-8a5b-8b070233f030" (UID: "5cfd6601-23fa-48fe-8a5b-8b070233f030"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.627597 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cfd6601-23fa-48fe-8a5b-8b070233f030-kube-api-access-gdp8g" (OuterVolumeSpecName: "kube-api-access-gdp8g") pod "5cfd6601-23fa-48fe-8a5b-8b070233f030" (UID: "5cfd6601-23fa-48fe-8a5b-8b070233f030"). InnerVolumeSpecName "kube-api-access-gdp8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.650108 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cfd6601-23fa-48fe-8a5b-8b070233f030-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cfd6601-23fa-48fe-8a5b-8b070233f030" (UID: "5cfd6601-23fa-48fe-8a5b-8b070233f030"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.666988 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cfd6601-23fa-48fe-8a5b-8b070233f030-config-data" (OuterVolumeSpecName: "config-data") pod "5cfd6601-23fa-48fe-8a5b-8b070233f030" (UID: "5cfd6601-23fa-48fe-8a5b-8b070233f030"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.700659 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ca3f9270-7d6a-4b91-bf50-944388836500","Type":"ContainerStarted","Data":"617d1d307ef28adc7ecba70884e00a37f469c79f8a45e6925777dabfade04be1"} Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.702295 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.702318 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ca3f9270-7d6a-4b91-bf50-944388836500","Type":"ContainerStarted","Data":"eb3d8b825139545a6b8ec216bdebabf88cda39089d87a1d35bb4b376b44d862d"} Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.702433 4775 generic.go:334] "Generic (PLEG): container finished" podID="7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" containerID="1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9" exitCode=0 Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.702446 4775 generic.go:334] "Generic (PLEG): container finished" podID="7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" containerID="ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c" exitCode=143 Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.702473 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1","Type":"ContainerDied","Data":"1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9"} Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.702490 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1","Type":"ContainerDied","Data":"ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c"} Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.702500 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1","Type":"ContainerDied","Data":"79b72b4afbc99dbe434761c2ec3ef9fcd7997318b9b31f66db1d11424724a6a6"} Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.702515 4775 scope.go:117] "RemoveContainer" containerID="1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.702553 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.704891 4775 generic.go:334] "Generic (PLEG): container finished" podID="5cfd6601-23fa-48fe-8a5b-8b070233f030" containerID="c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18" exitCode=0 Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.704907 4775 generic.go:334] "Generic (PLEG): container finished" podID="5cfd6601-23fa-48fe-8a5b-8b070233f030" containerID="3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a" exitCode=143 Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.704943 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cfd6601-23fa-48fe-8a5b-8b070233f030","Type":"ContainerDied","Data":"c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18"} Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.704978 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cfd6601-23fa-48fe-8a5b-8b070233f030","Type":"ContainerDied","Data":"3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a"} Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.704989 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5cfd6601-23fa-48fe-8a5b-8b070233f030","Type":"ContainerDied","Data":"f25f9fe413f3857a28f549206ad28536f36cfb0cf7dc631e17603ea451f3abcf"} Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.705059 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.725612 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5cfd6601-23fa-48fe-8a5b-8b070233f030-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.725644 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cfd6601-23fa-48fe-8a5b-8b070233f030-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.725654 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cfd6601-23fa-48fe-8a5b-8b070233f030-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.725667 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdp8g\" (UniqueName: \"kubernetes.io/projected/5cfd6601-23fa-48fe-8a5b-8b070233f030-kube-api-access-gdp8g\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.736714 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.736693959 podStartE2EDuration="2.736693959s" podCreationTimestamp="2025-10-02 03:14:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:17.729275687 +0000 UTC m=+5594.896019727" watchObservedRunningTime="2025-10-02 03:14:17.736693959 +0000 UTC m=+5594.903438009" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.766791 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.781483 4775 scope.go:117] "RemoveContainer" containerID="ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.798488 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.798529 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.799888 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:17 crc kubenswrapper[4775]: E1002 03:14:17.800338 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cfd6601-23fa-48fe-8a5b-8b070233f030" containerName="nova-metadata-log" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.800355 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cfd6601-23fa-48fe-8a5b-8b070233f030" containerName="nova-metadata-log" Oct 02 03:14:17 crc kubenswrapper[4775]: E1002 03:14:17.800364 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a0025f0-7fa5-4798-b9ec-09a3cec748d7" containerName="nova-manage" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.800371 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a0025f0-7fa5-4798-b9ec-09a3cec748d7" containerName="nova-manage" Oct 02 03:14:17 crc kubenswrapper[4775]: E1002 03:14:17.800399 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" containerName="nova-api-log" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.800406 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" containerName="nova-api-log" Oct 02 03:14:17 crc kubenswrapper[4775]: E1002 03:14:17.800413 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cfd6601-23fa-48fe-8a5b-8b070233f030" containerName="nova-metadata-metadata" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.800419 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cfd6601-23fa-48fe-8a5b-8b070233f030" containerName="nova-metadata-metadata" Oct 02 03:14:17 crc kubenswrapper[4775]: E1002 03:14:17.800441 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" containerName="nova-api-api" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.800448 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" containerName="nova-api-api" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.800613 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a0025f0-7fa5-4798-b9ec-09a3cec748d7" containerName="nova-manage" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.800622 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cfd6601-23fa-48fe-8a5b-8b070233f030" containerName="nova-metadata-log" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.800631 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cfd6601-23fa-48fe-8a5b-8b070233f030" containerName="nova-metadata-metadata" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.800645 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" containerName="nova-api-api" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.800656 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" containerName="nova-api-log" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.801587 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.812021 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.819104 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.823076 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.833831 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.837214 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.843166 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.843444 4775 scope.go:117] "RemoveContainer" containerID="1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9" Oct 02 03:14:17 crc kubenswrapper[4775]: E1002 03:14:17.844643 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9\": container with ID starting with 1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9 not found: ID does not exist" containerID="1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.844739 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9"} err="failed to get container status \"1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9\": rpc error: code = NotFound desc = could not find container \"1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9\": container with ID starting with 1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9 not found: ID does not exist" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.844825 4775 scope.go:117] "RemoveContainer" containerID="ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c" Oct 02 03:14:17 crc kubenswrapper[4775]: E1002 03:14:17.845779 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c\": container with ID starting with ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c not found: ID does not exist" containerID="ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.846250 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c"} err="failed to get container status \"ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c\": rpc error: code = NotFound desc = could not find container \"ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c\": container with ID starting with ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c not found: ID does not exist" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.846405 4775 scope.go:117] "RemoveContainer" containerID="1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.846722 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9"} err="failed to get container status \"1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9\": rpc error: code = NotFound desc = could not find container \"1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9\": container with ID starting with 1ae8cadc5afc007c3ea4eabcb5faf60e9bc052d24de1bc671f08143ff3783af9 not found: ID does not exist" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.846778 4775 scope.go:117] "RemoveContainer" containerID="ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.847174 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c"} err="failed to get container status \"ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c\": rpc error: code = NotFound desc = could not find container \"ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c\": container with ID starting with ba212cf95f71fe06398ea580a39ebe1b0bf16eb7907dfbfa4c68de8e7cf98c2c not found: ID does not exist" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.847277 4775 scope.go:117] "RemoveContainer" containerID="c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.847374 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.870439 4775 scope.go:117] "RemoveContainer" containerID="3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.885748 4775 scope.go:117] "RemoveContainer" containerID="c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18" Oct 02 03:14:17 crc kubenswrapper[4775]: E1002 03:14:17.887074 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18\": container with ID starting with c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18 not found: ID does not exist" containerID="c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.887118 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18"} err="failed to get container status \"c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18\": rpc error: code = NotFound desc = could not find container \"c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18\": container with ID starting with c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18 not found: ID does not exist" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.887163 4775 scope.go:117] "RemoveContainer" containerID="3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a" Oct 02 03:14:17 crc kubenswrapper[4775]: E1002 03:14:17.888105 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a\": container with ID starting with 3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a not found: ID does not exist" containerID="3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.888145 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a"} err="failed to get container status \"3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a\": rpc error: code = NotFound desc = could not find container \"3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a\": container with ID starting with 3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a not found: ID does not exist" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.888170 4775 scope.go:117] "RemoveContainer" containerID="c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.891584 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18"} err="failed to get container status \"c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18\": rpc error: code = NotFound desc = could not find container \"c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18\": container with ID starting with c2528f3ac0c943eba124f99d8bbc3054ed8e4ce2b5fec8f91cc049c8bfc74a18 not found: ID does not exist" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.891606 4775 scope.go:117] "RemoveContainer" containerID="3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.891784 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a"} err="failed to get container status \"3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a\": rpc error: code = NotFound desc = could not find container \"3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a\": container with ID starting with 3293266a683d934e8aea4a80b6a6de61e46d5a8a669657b1df81aa346a16302a not found: ID does not exist" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.929432 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-logs\") pod \"nova-api-0\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " pod="openstack/nova-api-0" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.929482 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " pod="openstack/nova-api-0" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.929509 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-config-data\") pod \"nova-api-0\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " pod="openstack/nova-api-0" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.929538 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k69d\" (UniqueName: \"kubernetes.io/projected/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-kube-api-access-4k69d\") pod \"nova-api-0\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " pod="openstack/nova-api-0" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.929566 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6eb403-8ff0-441e-99a5-4616606f1a04-config-data\") pod \"nova-metadata-0\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " pod="openstack/nova-metadata-0" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.929622 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrx7c\" (UniqueName: \"kubernetes.io/projected/2f6eb403-8ff0-441e-99a5-4616606f1a04-kube-api-access-vrx7c\") pod \"nova-metadata-0\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " pod="openstack/nova-metadata-0" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.929648 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f6eb403-8ff0-441e-99a5-4616606f1a04-logs\") pod \"nova-metadata-0\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " pod="openstack/nova-metadata-0" Oct 02 03:14:17 crc kubenswrapper[4775]: I1002 03:14:17.929676 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6eb403-8ff0-441e-99a5-4616606f1a04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " pod="openstack/nova-metadata-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.031557 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-logs\") pod \"nova-api-0\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " pod="openstack/nova-api-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.031673 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " pod="openstack/nova-api-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.031748 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-config-data\") pod \"nova-api-0\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " pod="openstack/nova-api-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.031817 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k69d\" (UniqueName: \"kubernetes.io/projected/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-kube-api-access-4k69d\") pod \"nova-api-0\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " pod="openstack/nova-api-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.031868 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6eb403-8ff0-441e-99a5-4616606f1a04-config-data\") pod \"nova-metadata-0\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " pod="openstack/nova-metadata-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.032020 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-logs\") pod \"nova-api-0\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " pod="openstack/nova-api-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.032024 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrx7c\" (UniqueName: \"kubernetes.io/projected/2f6eb403-8ff0-441e-99a5-4616606f1a04-kube-api-access-vrx7c\") pod \"nova-metadata-0\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " pod="openstack/nova-metadata-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.032136 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f6eb403-8ff0-441e-99a5-4616606f1a04-logs\") pod \"nova-metadata-0\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " pod="openstack/nova-metadata-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.032208 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6eb403-8ff0-441e-99a5-4616606f1a04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " pod="openstack/nova-metadata-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.033199 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f6eb403-8ff0-441e-99a5-4616606f1a04-logs\") pod \"nova-metadata-0\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " pod="openstack/nova-metadata-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.035183 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-config-data\") pod \"nova-api-0\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " pod="openstack/nova-api-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.036692 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6eb403-8ff0-441e-99a5-4616606f1a04-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " pod="openstack/nova-metadata-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.038585 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6eb403-8ff0-441e-99a5-4616606f1a04-config-data\") pod \"nova-metadata-0\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " pod="openstack/nova-metadata-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.038822 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " pod="openstack/nova-api-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.049067 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k69d\" (UniqueName: \"kubernetes.io/projected/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-kube-api-access-4k69d\") pod \"nova-api-0\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " pod="openstack/nova-api-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.051294 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrx7c\" (UniqueName: \"kubernetes.io/projected/2f6eb403-8ff0-441e-99a5-4616606f1a04-kube-api-access-vrx7c\") pod \"nova-metadata-0\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " pod="openstack/nova-metadata-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.124392 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.165058 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.607591 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:18 crc kubenswrapper[4775]: W1002 03:14:18.614161 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode09de27e_34c9_4a22_8fc6_aaf72ba6b583.slice/crio-a6cc987d95883344f38bf025807d1a12cc7a5377c4f90e9bb4b0b21339d4022d WatchSource:0}: Error finding container a6cc987d95883344f38bf025807d1a12cc7a5377c4f90e9bb4b0b21339d4022d: Status 404 returned error can't find the container with id a6cc987d95883344f38bf025807d1a12cc7a5377c4f90e9bb4b0b21339d4022d Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.740888 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e09de27e-34c9-4a22-8fc6-aaf72ba6b583","Type":"ContainerStarted","Data":"a6cc987d95883344f38bf025807d1a12cc7a5377c4f90e9bb4b0b21339d4022d"} Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.755325 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:18 crc kubenswrapper[4775]: W1002 03:14:18.758942 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f6eb403_8ff0_441e_99a5_4616606f1a04.slice/crio-ce6366d68737f53db17cb4fc9d61d3cdf8890d52dc010d821d285415709e114c WatchSource:0}: Error finding container ce6366d68737f53db17cb4fc9d61d3cdf8890d52dc010d821d285415709e114c: Status 404 returned error can't find the container with id ce6366d68737f53db17cb4fc9d61d3cdf8890d52dc010d821d285415709e114c Oct 02 03:14:18 crc kubenswrapper[4775]: I1002 03:14:18.998239 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.016335 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.129181 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.200113 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db697c7cc-w4mc6"] Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.200352 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" podUID="2ae17bef-94ba-4e40-ae25-fbcf8e4dab79" containerName="dnsmasq-dns" containerID="cri-o://28edfbd7b63632af67936b1c47cd2e2be628c7e79cca126a5cf1a38f1e3ce795" gracePeriod=10 Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.686101 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.762034 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-dns-svc\") pod \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.762989 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-ovsdbserver-nb\") pod \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.763116 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tnsh\" (UniqueName: \"kubernetes.io/projected/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-kube-api-access-4tnsh\") pod \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.763238 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-ovsdbserver-sb\") pod \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.763325 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-config\") pod \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\" (UID: \"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79\") " Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.775695 4775 generic.go:334] "Generic (PLEG): container finished" podID="2ae17bef-94ba-4e40-ae25-fbcf8e4dab79" containerID="28edfbd7b63632af67936b1c47cd2e2be628c7e79cca126a5cf1a38f1e3ce795" exitCode=0 Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.777163 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.777579 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cfd6601-23fa-48fe-8a5b-8b070233f030" path="/var/lib/kubelet/pods/5cfd6601-23fa-48fe-8a5b-8b070233f030/volumes" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.778242 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1" path="/var/lib/kubelet/pods/7df5bdbc-ea9c-4afb-b7f1-83b3f13fc3c1/volumes" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.783596 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-kube-api-access-4tnsh" (OuterVolumeSpecName: "kube-api-access-4tnsh") pod "2ae17bef-94ba-4e40-ae25-fbcf8e4dab79" (UID: "2ae17bef-94ba-4e40-ae25-fbcf8e4dab79"). InnerVolumeSpecName "kube-api-access-4tnsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.802517 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.802554 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f6eb403-8ff0-441e-99a5-4616606f1a04","Type":"ContainerStarted","Data":"090cf116a3f13f2459729e0fc829eab26a9d3ef8e5eb73f984f9ff4ed301bf4a"} Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.802570 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f6eb403-8ff0-441e-99a5-4616606f1a04","Type":"ContainerStarted","Data":"a37262fc5c77c394d27e02f2b5c649aaa548615bb8aaaa20b5839425834460ea"} Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.802618 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f6eb403-8ff0-441e-99a5-4616606f1a04","Type":"ContainerStarted","Data":"ce6366d68737f53db17cb4fc9d61d3cdf8890d52dc010d821d285415709e114c"} Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.802651 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e09de27e-34c9-4a22-8fc6-aaf72ba6b583","Type":"ContainerStarted","Data":"f6671679e9ace73e5ea3f2c45db281bbf4a6f45de35633a9d6a5ae4f95080a63"} Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.802671 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e09de27e-34c9-4a22-8fc6-aaf72ba6b583","Type":"ContainerStarted","Data":"47ec028efde4799589be16b23f019bbf495c1ba53eb9982646c05e6d330dc918"} Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.802683 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" event={"ID":"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79","Type":"ContainerDied","Data":"28edfbd7b63632af67936b1c47cd2e2be628c7e79cca126a5cf1a38f1e3ce795"} Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.802697 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db697c7cc-w4mc6" event={"ID":"2ae17bef-94ba-4e40-ae25-fbcf8e4dab79","Type":"ContainerDied","Data":"979c422b0cda2e52b370b2a9d7568c0035c14b0ace99efeb3a5e8967ddc29bf3"} Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.802715 4775 scope.go:117] "RemoveContainer" containerID="28edfbd7b63632af67936b1c47cd2e2be628c7e79cca126a5cf1a38f1e3ce795" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.803534 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.803186879 podStartE2EDuration="2.803186879s" podCreationTimestamp="2025-10-02 03:14:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:19.791827125 +0000 UTC m=+5596.958571155" watchObservedRunningTime="2025-10-02 03:14:19.803186879 +0000 UTC m=+5596.969930909" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.809789 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2ae17bef-94ba-4e40-ae25-fbcf8e4dab79" (UID: "2ae17bef-94ba-4e40-ae25-fbcf8e4dab79"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.821616 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-config" (OuterVolumeSpecName: "config") pod "2ae17bef-94ba-4e40-ae25-fbcf8e4dab79" (UID: "2ae17bef-94ba-4e40-ae25-fbcf8e4dab79"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.828096 4775 scope.go:117] "RemoveContainer" containerID="423c8d898cad7dc57d306509070a670cac226b66e0bcc654e0f7dea88faf5cc9" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.832556 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2ae17bef-94ba-4e40-ae25-fbcf8e4dab79" (UID: "2ae17bef-94ba-4e40-ae25-fbcf8e4dab79"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.833029 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2ae17bef-94ba-4e40-ae25-fbcf8e4dab79" (UID: "2ae17bef-94ba-4e40-ae25-fbcf8e4dab79"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.834820 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.8348024670000003 podStartE2EDuration="2.834802467s" podCreationTimestamp="2025-10-02 03:14:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:19.816257508 +0000 UTC m=+5596.983001558" watchObservedRunningTime="2025-10-02 03:14:19.834802467 +0000 UTC m=+5597.001546507" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.857883 4775 scope.go:117] "RemoveContainer" containerID="28edfbd7b63632af67936b1c47cd2e2be628c7e79cca126a5cf1a38f1e3ce795" Oct 02 03:14:19 crc kubenswrapper[4775]: E1002 03:14:19.858431 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28edfbd7b63632af67936b1c47cd2e2be628c7e79cca126a5cf1a38f1e3ce795\": container with ID starting with 28edfbd7b63632af67936b1c47cd2e2be628c7e79cca126a5cf1a38f1e3ce795 not found: ID does not exist" containerID="28edfbd7b63632af67936b1c47cd2e2be628c7e79cca126a5cf1a38f1e3ce795" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.858484 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28edfbd7b63632af67936b1c47cd2e2be628c7e79cca126a5cf1a38f1e3ce795"} err="failed to get container status \"28edfbd7b63632af67936b1c47cd2e2be628c7e79cca126a5cf1a38f1e3ce795\": rpc error: code = NotFound desc = could not find container \"28edfbd7b63632af67936b1c47cd2e2be628c7e79cca126a5cf1a38f1e3ce795\": container with ID starting with 28edfbd7b63632af67936b1c47cd2e2be628c7e79cca126a5cf1a38f1e3ce795 not found: ID does not exist" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.858510 4775 scope.go:117] "RemoveContainer" containerID="423c8d898cad7dc57d306509070a670cac226b66e0bcc654e0f7dea88faf5cc9" Oct 02 03:14:19 crc kubenswrapper[4775]: E1002 03:14:19.862195 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"423c8d898cad7dc57d306509070a670cac226b66e0bcc654e0f7dea88faf5cc9\": container with ID starting with 423c8d898cad7dc57d306509070a670cac226b66e0bcc654e0f7dea88faf5cc9 not found: ID does not exist" containerID="423c8d898cad7dc57d306509070a670cac226b66e0bcc654e0f7dea88faf5cc9" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.862242 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"423c8d898cad7dc57d306509070a670cac226b66e0bcc654e0f7dea88faf5cc9"} err="failed to get container status \"423c8d898cad7dc57d306509070a670cac226b66e0bcc654e0f7dea88faf5cc9\": rpc error: code = NotFound desc = could not find container \"423c8d898cad7dc57d306509070a670cac226b66e0bcc654e0f7dea88faf5cc9\": container with ID starting with 423c8d898cad7dc57d306509070a670cac226b66e0bcc654e0f7dea88faf5cc9 not found: ID does not exist" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.880714 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tnsh\" (UniqueName: \"kubernetes.io/projected/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-kube-api-access-4tnsh\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.881863 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.881983 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.882042 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:19 crc kubenswrapper[4775]: I1002 03:14:19.882092 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:20 crc kubenswrapper[4775]: I1002 03:14:20.128073 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db697c7cc-w4mc6"] Oct 02 03:14:20 crc kubenswrapper[4775]: I1002 03:14:20.138692 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-db697c7cc-w4mc6"] Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.176727 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.576388 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.712504 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02736f53-7af1-41f4-ae9b-df12578b7cfa-config-data\") pod \"02736f53-7af1-41f4-ae9b-df12578b7cfa\" (UID: \"02736f53-7af1-41f4-ae9b-df12578b7cfa\") " Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.712576 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jv8kx\" (UniqueName: \"kubernetes.io/projected/02736f53-7af1-41f4-ae9b-df12578b7cfa-kube-api-access-jv8kx\") pod \"02736f53-7af1-41f4-ae9b-df12578b7cfa\" (UID: \"02736f53-7af1-41f4-ae9b-df12578b7cfa\") " Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.712689 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02736f53-7af1-41f4-ae9b-df12578b7cfa-combined-ca-bundle\") pod \"02736f53-7af1-41f4-ae9b-df12578b7cfa\" (UID: \"02736f53-7af1-41f4-ae9b-df12578b7cfa\") " Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.717215 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02736f53-7af1-41f4-ae9b-df12578b7cfa-kube-api-access-jv8kx" (OuterVolumeSpecName: "kube-api-access-jv8kx") pod "02736f53-7af1-41f4-ae9b-df12578b7cfa" (UID: "02736f53-7af1-41f4-ae9b-df12578b7cfa"). InnerVolumeSpecName "kube-api-access-jv8kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.746568 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02736f53-7af1-41f4-ae9b-df12578b7cfa-config-data" (OuterVolumeSpecName: "config-data") pod "02736f53-7af1-41f4-ae9b-df12578b7cfa" (UID: "02736f53-7af1-41f4-ae9b-df12578b7cfa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.752892 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02736f53-7af1-41f4-ae9b-df12578b7cfa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02736f53-7af1-41f4-ae9b-df12578b7cfa" (UID: "02736f53-7af1-41f4-ae9b-df12578b7cfa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.776331 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ae17bef-94ba-4e40-ae25-fbcf8e4dab79" path="/var/lib/kubelet/pods/2ae17bef-94ba-4e40-ae25-fbcf8e4dab79/volumes" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.810991 4775 generic.go:334] "Generic (PLEG): container finished" podID="02736f53-7af1-41f4-ae9b-df12578b7cfa" containerID="da66341123db38c6d3572988cbfc6a6a6ac4d088fec31963fa8b35ca27c38051" exitCode=0 Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.811877 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.812412 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02736f53-7af1-41f4-ae9b-df12578b7cfa","Type":"ContainerDied","Data":"da66341123db38c6d3572988cbfc6a6a6ac4d088fec31963fa8b35ca27c38051"} Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.812448 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02736f53-7af1-41f4-ae9b-df12578b7cfa","Type":"ContainerDied","Data":"70f2f05411e0c87b45711c90d290ab5899f78e7554da4612c17c2b2daf87e46a"} Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.812464 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-99l8t"] Oct 02 03:14:21 crc kubenswrapper[4775]: E1002 03:14:21.812848 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02736f53-7af1-41f4-ae9b-df12578b7cfa" containerName="nova-scheduler-scheduler" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.813132 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="02736f53-7af1-41f4-ae9b-df12578b7cfa" containerName="nova-scheduler-scheduler" Oct 02 03:14:21 crc kubenswrapper[4775]: E1002 03:14:21.813150 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae17bef-94ba-4e40-ae25-fbcf8e4dab79" containerName="init" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.813158 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae17bef-94ba-4e40-ae25-fbcf8e4dab79" containerName="init" Oct 02 03:14:21 crc kubenswrapper[4775]: E1002 03:14:21.813196 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ae17bef-94ba-4e40-ae25-fbcf8e4dab79" containerName="dnsmasq-dns" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.813203 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ae17bef-94ba-4e40-ae25-fbcf8e4dab79" containerName="dnsmasq-dns" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.813397 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="02736f53-7af1-41f4-ae9b-df12578b7cfa" containerName="nova-scheduler-scheduler" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.813431 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ae17bef-94ba-4e40-ae25-fbcf8e4dab79" containerName="dnsmasq-dns" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.814117 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.815093 4775 scope.go:117] "RemoveContainer" containerID="da66341123db38c6d3572988cbfc6a6a6ac4d088fec31963fa8b35ca27c38051" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.816408 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02736f53-7af1-41f4-ae9b-df12578b7cfa-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.816459 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jv8kx\" (UniqueName: \"kubernetes.io/projected/02736f53-7af1-41f4-ae9b-df12578b7cfa-kube-api-access-jv8kx\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.816473 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02736f53-7af1-41f4-ae9b-df12578b7cfa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.817018 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.817265 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.834991 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-99l8t"] Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.870238 4775 scope.go:117] "RemoveContainer" containerID="da66341123db38c6d3572988cbfc6a6a6ac4d088fec31963fa8b35ca27c38051" Oct 02 03:14:21 crc kubenswrapper[4775]: E1002 03:14:21.870845 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da66341123db38c6d3572988cbfc6a6a6ac4d088fec31963fa8b35ca27c38051\": container with ID starting with da66341123db38c6d3572988cbfc6a6a6ac4d088fec31963fa8b35ca27c38051 not found: ID does not exist" containerID="da66341123db38c6d3572988cbfc6a6a6ac4d088fec31963fa8b35ca27c38051" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.870902 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da66341123db38c6d3572988cbfc6a6a6ac4d088fec31963fa8b35ca27c38051"} err="failed to get container status \"da66341123db38c6d3572988cbfc6a6a6ac4d088fec31963fa8b35ca27c38051\": rpc error: code = NotFound desc = could not find container \"da66341123db38c6d3572988cbfc6a6a6ac4d088fec31963fa8b35ca27c38051\": container with ID starting with da66341123db38c6d3572988cbfc6a6a6ac4d088fec31963fa8b35ca27c38051 not found: ID does not exist" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.897865 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.904691 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.913253 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.914447 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.917173 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.917424 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-config-data\") pod \"nova-cell1-cell-mapping-99l8t\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.917493 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhvjc\" (UniqueName: \"kubernetes.io/projected/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-kube-api-access-vhvjc\") pod \"nova-cell1-cell-mapping-99l8t\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.917548 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-99l8t\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.917626 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-scripts\") pod \"nova-cell1-cell-mapping-99l8t\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:21 crc kubenswrapper[4775]: I1002 03:14:21.922586 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.018982 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-config-data\") pod \"nova-cell1-cell-mapping-99l8t\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.019053 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhvjc\" (UniqueName: \"kubernetes.io/projected/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-kube-api-access-vhvjc\") pod \"nova-cell1-cell-mapping-99l8t\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.019087 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91aec47c-e3e1-4a2e-a344-0429ed546b57-config-data\") pod \"nova-scheduler-0\" (UID: \"91aec47c-e3e1-4a2e-a344-0429ed546b57\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.019112 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91aec47c-e3e1-4a2e-a344-0429ed546b57-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"91aec47c-e3e1-4a2e-a344-0429ed546b57\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.019131 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-99l8t\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.019158 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz78t\" (UniqueName: \"kubernetes.io/projected/91aec47c-e3e1-4a2e-a344-0429ed546b57-kube-api-access-kz78t\") pod \"nova-scheduler-0\" (UID: \"91aec47c-e3e1-4a2e-a344-0429ed546b57\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.019220 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-scripts\") pod \"nova-cell1-cell-mapping-99l8t\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.022523 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-scripts\") pod \"nova-cell1-cell-mapping-99l8t\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.023332 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-config-data\") pod \"nova-cell1-cell-mapping-99l8t\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.024622 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-99l8t\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.037051 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhvjc\" (UniqueName: \"kubernetes.io/projected/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-kube-api-access-vhvjc\") pod \"nova-cell1-cell-mapping-99l8t\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.121155 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91aec47c-e3e1-4a2e-a344-0429ed546b57-config-data\") pod \"nova-scheduler-0\" (UID: \"91aec47c-e3e1-4a2e-a344-0429ed546b57\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.121209 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91aec47c-e3e1-4a2e-a344-0429ed546b57-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"91aec47c-e3e1-4a2e-a344-0429ed546b57\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.121240 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz78t\" (UniqueName: \"kubernetes.io/projected/91aec47c-e3e1-4a2e-a344-0429ed546b57-kube-api-access-kz78t\") pod \"nova-scheduler-0\" (UID: \"91aec47c-e3e1-4a2e-a344-0429ed546b57\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.124471 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91aec47c-e3e1-4a2e-a344-0429ed546b57-config-data\") pod \"nova-scheduler-0\" (UID: \"91aec47c-e3e1-4a2e-a344-0429ed546b57\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.126823 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91aec47c-e3e1-4a2e-a344-0429ed546b57-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"91aec47c-e3e1-4a2e-a344-0429ed546b57\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.140108 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz78t\" (UniqueName: \"kubernetes.io/projected/91aec47c-e3e1-4a2e-a344-0429ed546b57-kube-api-access-kz78t\") pod \"nova-scheduler-0\" (UID: \"91aec47c-e3e1-4a2e-a344-0429ed546b57\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.144390 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.230372 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.590276 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-99l8t"] Oct 02 03:14:22 crc kubenswrapper[4775]: W1002 03:14:22.597119 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda133e8d0_41e4_46fe_b32c_81d5d6fe0888.slice/crio-c4aee8e7ff512de682d208b38fe7abff5c6993ebc41d8d2ecfe2f057833598bd WatchSource:0}: Error finding container c4aee8e7ff512de682d208b38fe7abff5c6993ebc41d8d2ecfe2f057833598bd: Status 404 returned error can't find the container with id c4aee8e7ff512de682d208b38fe7abff5c6993ebc41d8d2ecfe2f057833598bd Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.731750 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:22 crc kubenswrapper[4775]: W1002 03:14:22.733927 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91aec47c_e3e1_4a2e_a344_0429ed546b57.slice/crio-1f2708ba87f9dcef4345c1a86b48010df2bccf290c33caa107d9eefa859b3387 WatchSource:0}: Error finding container 1f2708ba87f9dcef4345c1a86b48010df2bccf290c33caa107d9eefa859b3387: Status 404 returned error can't find the container with id 1f2708ba87f9dcef4345c1a86b48010df2bccf290c33caa107d9eefa859b3387 Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.824460 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-99l8t" event={"ID":"a133e8d0-41e4-46fe-b32c-81d5d6fe0888","Type":"ContainerStarted","Data":"bee34c6b58d6da6c70e62be6a6cd4b4aa98a51b42fbc8c8a91501d8fa0c59e65"} Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.824506 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-99l8t" event={"ID":"a133e8d0-41e4-46fe-b32c-81d5d6fe0888","Type":"ContainerStarted","Data":"c4aee8e7ff512de682d208b38fe7abff5c6993ebc41d8d2ecfe2f057833598bd"} Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.827133 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"91aec47c-e3e1-4a2e-a344-0429ed546b57","Type":"ContainerStarted","Data":"1f2708ba87f9dcef4345c1a86b48010df2bccf290c33caa107d9eefa859b3387"} Oct 02 03:14:22 crc kubenswrapper[4775]: I1002 03:14:22.848634 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-99l8t" podStartSLOduration=1.848613591 podStartE2EDuration="1.848613591s" podCreationTimestamp="2025-10-02 03:14:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:22.843614112 +0000 UTC m=+5600.010358162" watchObservedRunningTime="2025-10-02 03:14:22.848613591 +0000 UTC m=+5600.015357641" Oct 02 03:14:23 crc kubenswrapper[4775]: I1002 03:14:23.166389 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 03:14:23 crc kubenswrapper[4775]: I1002 03:14:23.167384 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 03:14:23 crc kubenswrapper[4775]: I1002 03:14:23.780191 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02736f53-7af1-41f4-ae9b-df12578b7cfa" path="/var/lib/kubelet/pods/02736f53-7af1-41f4-ae9b-df12578b7cfa/volumes" Oct 02 03:14:23 crc kubenswrapper[4775]: I1002 03:14:23.840131 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"91aec47c-e3e1-4a2e-a344-0429ed546b57","Type":"ContainerStarted","Data":"290fa0a45ee8f59b2ce2d1f81c9b4e1fd366d6b2f5b60ca4919ff63230a5f859"} Oct 02 03:14:23 crc kubenswrapper[4775]: I1002 03:14:23.874723 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.874697345 podStartE2EDuration="2.874697345s" podCreationTimestamp="2025-10-02 03:14:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:23.86407401 +0000 UTC m=+5601.030818060" watchObservedRunningTime="2025-10-02 03:14:23.874697345 +0000 UTC m=+5601.041441425" Oct 02 03:14:26 crc kubenswrapper[4775]: I1002 03:14:26.873834 4775 generic.go:334] "Generic (PLEG): container finished" podID="a133e8d0-41e4-46fe-b32c-81d5d6fe0888" containerID="bee34c6b58d6da6c70e62be6a6cd4b4aa98a51b42fbc8c8a91501d8fa0c59e65" exitCode=0 Oct 02 03:14:26 crc kubenswrapper[4775]: I1002 03:14:26.873996 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-99l8t" event={"ID":"a133e8d0-41e4-46fe-b32c-81d5d6fe0888","Type":"ContainerDied","Data":"bee34c6b58d6da6c70e62be6a6cd4b4aa98a51b42fbc8c8a91501d8fa0c59e65"} Oct 02 03:14:27 crc kubenswrapper[4775]: I1002 03:14:27.231503 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.126247 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.126300 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.166136 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.166421 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.330293 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.464272 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-scripts\") pod \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.464357 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-config-data\") pod \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.464402 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-combined-ca-bundle\") pod \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.464452 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhvjc\" (UniqueName: \"kubernetes.io/projected/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-kube-api-access-vhvjc\") pod \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\" (UID: \"a133e8d0-41e4-46fe-b32c-81d5d6fe0888\") " Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.470861 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-scripts" (OuterVolumeSpecName: "scripts") pod "a133e8d0-41e4-46fe-b32c-81d5d6fe0888" (UID: "a133e8d0-41e4-46fe-b32c-81d5d6fe0888"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.471093 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-kube-api-access-vhvjc" (OuterVolumeSpecName: "kube-api-access-vhvjc") pod "a133e8d0-41e4-46fe-b32c-81d5d6fe0888" (UID: "a133e8d0-41e4-46fe-b32c-81d5d6fe0888"). InnerVolumeSpecName "kube-api-access-vhvjc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.509444 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-config-data" (OuterVolumeSpecName: "config-data") pod "a133e8d0-41e4-46fe-b32c-81d5d6fe0888" (UID: "a133e8d0-41e4-46fe-b32c-81d5d6fe0888"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.516781 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a133e8d0-41e4-46fe-b32c-81d5d6fe0888" (UID: "a133e8d0-41e4-46fe-b32c-81d5d6fe0888"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.566161 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.566441 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.566451 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.566462 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhvjc\" (UniqueName: \"kubernetes.io/projected/a133e8d0-41e4-46fe-b32c-81d5d6fe0888-kube-api-access-vhvjc\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.901854 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-99l8t" event={"ID":"a133e8d0-41e4-46fe-b32c-81d5d6fe0888","Type":"ContainerDied","Data":"c4aee8e7ff512de682d208b38fe7abff5c6993ebc41d8d2ecfe2f057833598bd"} Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.901893 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4aee8e7ff512de682d208b38fe7abff5c6993ebc41d8d2ecfe2f057833598bd" Oct 02 03:14:28 crc kubenswrapper[4775]: I1002 03:14:28.901937 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-99l8t" Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.089428 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.089754 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e09de27e-34c9-4a22-8fc6-aaf72ba6b583" containerName="nova-api-log" containerID="cri-o://47ec028efde4799589be16b23f019bbf495c1ba53eb9982646c05e6d330dc918" gracePeriod=30 Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.089807 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e09de27e-34c9-4a22-8fc6-aaf72ba6b583" containerName="nova-api-api" containerID="cri-o://f6671679e9ace73e5ea3f2c45db281bbf4a6f45de35633a9d6a5ae4f95080a63" gracePeriod=30 Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.098707 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e09de27e-34c9-4a22-8fc6-aaf72ba6b583" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.68:8774/\": EOF" Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.098763 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e09de27e-34c9-4a22-8fc6-aaf72ba6b583" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.68:8774/\": EOF" Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.101102 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.104059 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="91aec47c-e3e1-4a2e-a344-0429ed546b57" containerName="nova-scheduler-scheduler" containerID="cri-o://290fa0a45ee8f59b2ce2d1f81c9b4e1fd366d6b2f5b60ca4919ff63230a5f859" gracePeriod=30 Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.142819 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.143061 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2f6eb403-8ff0-441e-99a5-4616606f1a04" containerName="nova-metadata-log" containerID="cri-o://a37262fc5c77c394d27e02f2b5c649aaa548615bb8aaaa20b5839425834460ea" gracePeriod=30 Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.143154 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="2f6eb403-8ff0-441e-99a5-4616606f1a04" containerName="nova-metadata-metadata" containerID="cri-o://090cf116a3f13f2459729e0fc829eab26a9d3ef8e5eb73f984f9ff4ed301bf4a" gracePeriod=30 Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.149042 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2f6eb403-8ff0-441e-99a5-4616606f1a04" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.69:8775/\": EOF" Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.149086 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="2f6eb403-8ff0-441e-99a5-4616606f1a04" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.69:8775/\": EOF" Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.910861 4775 generic.go:334] "Generic (PLEG): container finished" podID="e09de27e-34c9-4a22-8fc6-aaf72ba6b583" containerID="47ec028efde4799589be16b23f019bbf495c1ba53eb9982646c05e6d330dc918" exitCode=143 Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.910937 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e09de27e-34c9-4a22-8fc6-aaf72ba6b583","Type":"ContainerDied","Data":"47ec028efde4799589be16b23f019bbf495c1ba53eb9982646c05e6d330dc918"} Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.912842 4775 generic.go:334] "Generic (PLEG): container finished" podID="2f6eb403-8ff0-441e-99a5-4616606f1a04" containerID="a37262fc5c77c394d27e02f2b5c649aaa548615bb8aaaa20b5839425834460ea" exitCode=143 Oct 02 03:14:29 crc kubenswrapper[4775]: I1002 03:14:29.912865 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f6eb403-8ff0-441e-99a5-4616606f1a04","Type":"ContainerDied","Data":"a37262fc5c77c394d27e02f2b5c649aaa548615bb8aaaa20b5839425834460ea"} Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.458976 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.616848 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91aec47c-e3e1-4a2e-a344-0429ed546b57-config-data\") pod \"91aec47c-e3e1-4a2e-a344-0429ed546b57\" (UID: \"91aec47c-e3e1-4a2e-a344-0429ed546b57\") " Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.618420 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91aec47c-e3e1-4a2e-a344-0429ed546b57-combined-ca-bundle\") pod \"91aec47c-e3e1-4a2e-a344-0429ed546b57\" (UID: \"91aec47c-e3e1-4a2e-a344-0429ed546b57\") " Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.618516 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kz78t\" (UniqueName: \"kubernetes.io/projected/91aec47c-e3e1-4a2e-a344-0429ed546b57-kube-api-access-kz78t\") pod \"91aec47c-e3e1-4a2e-a344-0429ed546b57\" (UID: \"91aec47c-e3e1-4a2e-a344-0429ed546b57\") " Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.623746 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91aec47c-e3e1-4a2e-a344-0429ed546b57-kube-api-access-kz78t" (OuterVolumeSpecName: "kube-api-access-kz78t") pod "91aec47c-e3e1-4a2e-a344-0429ed546b57" (UID: "91aec47c-e3e1-4a2e-a344-0429ed546b57"). InnerVolumeSpecName "kube-api-access-kz78t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.651106 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91aec47c-e3e1-4a2e-a344-0429ed546b57-config-data" (OuterVolumeSpecName: "config-data") pod "91aec47c-e3e1-4a2e-a344-0429ed546b57" (UID: "91aec47c-e3e1-4a2e-a344-0429ed546b57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.660271 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91aec47c-e3e1-4a2e-a344-0429ed546b57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91aec47c-e3e1-4a2e-a344-0429ed546b57" (UID: "91aec47c-e3e1-4a2e-a344-0429ed546b57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.721013 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kz78t\" (UniqueName: \"kubernetes.io/projected/91aec47c-e3e1-4a2e-a344-0429ed546b57-kube-api-access-kz78t\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.721041 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91aec47c-e3e1-4a2e-a344-0429ed546b57-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.721052 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91aec47c-e3e1-4a2e-a344-0429ed546b57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.928631 4775 generic.go:334] "Generic (PLEG): container finished" podID="91aec47c-e3e1-4a2e-a344-0429ed546b57" containerID="290fa0a45ee8f59b2ce2d1f81c9b4e1fd366d6b2f5b60ca4919ff63230a5f859" exitCode=0 Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.928686 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"91aec47c-e3e1-4a2e-a344-0429ed546b57","Type":"ContainerDied","Data":"290fa0a45ee8f59b2ce2d1f81c9b4e1fd366d6b2f5b60ca4919ff63230a5f859"} Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.928715 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"91aec47c-e3e1-4a2e-a344-0429ed546b57","Type":"ContainerDied","Data":"1f2708ba87f9dcef4345c1a86b48010df2bccf290c33caa107d9eefa859b3387"} Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.928740 4775 scope.go:117] "RemoveContainer" containerID="290fa0a45ee8f59b2ce2d1f81c9b4e1fd366d6b2f5b60ca4919ff63230a5f859" Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.928819 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.948740 4775 scope.go:117] "RemoveContainer" containerID="290fa0a45ee8f59b2ce2d1f81c9b4e1fd366d6b2f5b60ca4919ff63230a5f859" Oct 02 03:14:30 crc kubenswrapper[4775]: E1002 03:14:30.949303 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"290fa0a45ee8f59b2ce2d1f81c9b4e1fd366d6b2f5b60ca4919ff63230a5f859\": container with ID starting with 290fa0a45ee8f59b2ce2d1f81c9b4e1fd366d6b2f5b60ca4919ff63230a5f859 not found: ID does not exist" containerID="290fa0a45ee8f59b2ce2d1f81c9b4e1fd366d6b2f5b60ca4919ff63230a5f859" Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.949335 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"290fa0a45ee8f59b2ce2d1f81c9b4e1fd366d6b2f5b60ca4919ff63230a5f859"} err="failed to get container status \"290fa0a45ee8f59b2ce2d1f81c9b4e1fd366d6b2f5b60ca4919ff63230a5f859\": rpc error: code = NotFound desc = could not find container \"290fa0a45ee8f59b2ce2d1f81c9b4e1fd366d6b2f5b60ca4919ff63230a5f859\": container with ID starting with 290fa0a45ee8f59b2ce2d1f81c9b4e1fd366d6b2f5b60ca4919ff63230a5f859 not found: ID does not exist" Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.969151 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:30 crc kubenswrapper[4775]: I1002 03:14:30.976468 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.007192 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:31 crc kubenswrapper[4775]: E1002 03:14:31.007631 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91aec47c-e3e1-4a2e-a344-0429ed546b57" containerName="nova-scheduler-scheduler" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.007652 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="91aec47c-e3e1-4a2e-a344-0429ed546b57" containerName="nova-scheduler-scheduler" Oct 02 03:14:31 crc kubenswrapper[4775]: E1002 03:14:31.007674 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a133e8d0-41e4-46fe-b32c-81d5d6fe0888" containerName="nova-manage" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.007685 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a133e8d0-41e4-46fe-b32c-81d5d6fe0888" containerName="nova-manage" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.007922 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a133e8d0-41e4-46fe-b32c-81d5d6fe0888" containerName="nova-manage" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.007947 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="91aec47c-e3e1-4a2e-a344-0429ed546b57" containerName="nova-scheduler-scheduler" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.008714 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.010191 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.015224 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.132124 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-config-data\") pod \"nova-scheduler-0\" (UID: \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.132202 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.132243 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsv4d\" (UniqueName: \"kubernetes.io/projected/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-kube-api-access-qsv4d\") pod \"nova-scheduler-0\" (UID: \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.234828 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-config-data\") pod \"nova-scheduler-0\" (UID: \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.234919 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.234979 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsv4d\" (UniqueName: \"kubernetes.io/projected/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-kube-api-access-qsv4d\") pod \"nova-scheduler-0\" (UID: \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.240298 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.240406 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-config-data\") pod \"nova-scheduler-0\" (UID: \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.259673 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsv4d\" (UniqueName: \"kubernetes.io/projected/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-kube-api-access-qsv4d\") pod \"nova-scheduler-0\" (UID: \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\") " pod="openstack/nova-scheduler-0" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.347004 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.785542 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91aec47c-e3e1-4a2e-a344-0429ed546b57" path="/var/lib/kubelet/pods/91aec47c-e3e1-4a2e-a344-0429ed546b57/volumes" Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.873596 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:14:31 crc kubenswrapper[4775]: I1002 03:14:31.943314 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6","Type":"ContainerStarted","Data":"eaccc751aaef3b57d17001d4b967c99661ac7a6f44e0605e23fbfef6f34d419f"} Oct 02 03:14:32 crc kubenswrapper[4775]: I1002 03:14:32.962005 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6","Type":"ContainerStarted","Data":"1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935"} Oct 02 03:14:32 crc kubenswrapper[4775]: I1002 03:14:32.995174 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.995146062 podStartE2EDuration="2.995146062s" podCreationTimestamp="2025-10-02 03:14:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:32.989141977 +0000 UTC m=+5610.155886057" watchObservedRunningTime="2025-10-02 03:14:32.995146062 +0000 UTC m=+5610.161890132" Oct 02 03:14:33 crc kubenswrapper[4775]: I1002 03:14:33.974985 4775 generic.go:334] "Generic (PLEG): container finished" podID="2f6eb403-8ff0-441e-99a5-4616606f1a04" containerID="090cf116a3f13f2459729e0fc829eab26a9d3ef8e5eb73f984f9ff4ed301bf4a" exitCode=0 Oct 02 03:14:33 crc kubenswrapper[4775]: I1002 03:14:33.975212 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f6eb403-8ff0-441e-99a5-4616606f1a04","Type":"ContainerDied","Data":"090cf116a3f13f2459729e0fc829eab26a9d3ef8e5eb73f984f9ff4ed301bf4a"} Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.055977 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.206706 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vrx7c\" (UniqueName: \"kubernetes.io/projected/2f6eb403-8ff0-441e-99a5-4616606f1a04-kube-api-access-vrx7c\") pod \"2f6eb403-8ff0-441e-99a5-4616606f1a04\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.206755 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f6eb403-8ff0-441e-99a5-4616606f1a04-logs\") pod \"2f6eb403-8ff0-441e-99a5-4616606f1a04\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.206898 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6eb403-8ff0-441e-99a5-4616606f1a04-config-data\") pod \"2f6eb403-8ff0-441e-99a5-4616606f1a04\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.206939 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6eb403-8ff0-441e-99a5-4616606f1a04-combined-ca-bundle\") pod \"2f6eb403-8ff0-441e-99a5-4616606f1a04\" (UID: \"2f6eb403-8ff0-441e-99a5-4616606f1a04\") " Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.208628 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f6eb403-8ff0-441e-99a5-4616606f1a04-logs" (OuterVolumeSpecName: "logs") pod "2f6eb403-8ff0-441e-99a5-4616606f1a04" (UID: "2f6eb403-8ff0-441e-99a5-4616606f1a04"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.237209 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f6eb403-8ff0-441e-99a5-4616606f1a04-kube-api-access-vrx7c" (OuterVolumeSpecName: "kube-api-access-vrx7c") pod "2f6eb403-8ff0-441e-99a5-4616606f1a04" (UID: "2f6eb403-8ff0-441e-99a5-4616606f1a04"). InnerVolumeSpecName "kube-api-access-vrx7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.275937 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f6eb403-8ff0-441e-99a5-4616606f1a04-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f6eb403-8ff0-441e-99a5-4616606f1a04" (UID: "2f6eb403-8ff0-441e-99a5-4616606f1a04"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.276812 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f6eb403-8ff0-441e-99a5-4616606f1a04-config-data" (OuterVolumeSpecName: "config-data") pod "2f6eb403-8ff0-441e-99a5-4616606f1a04" (UID: "2f6eb403-8ff0-441e-99a5-4616606f1a04"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.309343 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f6eb403-8ff0-441e-99a5-4616606f1a04-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.309372 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f6eb403-8ff0-441e-99a5-4616606f1a04-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.309381 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f6eb403-8ff0-441e-99a5-4616606f1a04-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.309392 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vrx7c\" (UniqueName: \"kubernetes.io/projected/2f6eb403-8ff0-441e-99a5-4616606f1a04-kube-api-access-vrx7c\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.982589 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.984326 4775 generic.go:334] "Generic (PLEG): container finished" podID="e09de27e-34c9-4a22-8fc6-aaf72ba6b583" containerID="f6671679e9ace73e5ea3f2c45db281bbf4a6f45de35633a9d6a5ae4f95080a63" exitCode=0 Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.985060 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e09de27e-34c9-4a22-8fc6-aaf72ba6b583","Type":"ContainerDied","Data":"f6671679e9ace73e5ea3f2c45db281bbf4a6f45de35633a9d6a5ae4f95080a63"} Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.985096 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e09de27e-34c9-4a22-8fc6-aaf72ba6b583","Type":"ContainerDied","Data":"a6cc987d95883344f38bf025807d1a12cc7a5377c4f90e9bb4b0b21339d4022d"} Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.985118 4775 scope.go:117] "RemoveContainer" containerID="f6671679e9ace73e5ea3f2c45db281bbf4a6f45de35633a9d6a5ae4f95080a63" Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.988115 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"2f6eb403-8ff0-441e-99a5-4616606f1a04","Type":"ContainerDied","Data":"ce6366d68737f53db17cb4fc9d61d3cdf8890d52dc010d821d285415709e114c"} Oct 02 03:14:34 crc kubenswrapper[4775]: I1002 03:14:34.988195 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.007759 4775 scope.go:117] "RemoveContainer" containerID="47ec028efde4799589be16b23f019bbf495c1ba53eb9982646c05e6d330dc918" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.028414 4775 scope.go:117] "RemoveContainer" containerID="f6671679e9ace73e5ea3f2c45db281bbf4a6f45de35633a9d6a5ae4f95080a63" Oct 02 03:14:35 crc kubenswrapper[4775]: E1002 03:14:35.028908 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6671679e9ace73e5ea3f2c45db281bbf4a6f45de35633a9d6a5ae4f95080a63\": container with ID starting with f6671679e9ace73e5ea3f2c45db281bbf4a6f45de35633a9d6a5ae4f95080a63 not found: ID does not exist" containerID="f6671679e9ace73e5ea3f2c45db281bbf4a6f45de35633a9d6a5ae4f95080a63" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.029047 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6671679e9ace73e5ea3f2c45db281bbf4a6f45de35633a9d6a5ae4f95080a63"} err="failed to get container status \"f6671679e9ace73e5ea3f2c45db281bbf4a6f45de35633a9d6a5ae4f95080a63\": rpc error: code = NotFound desc = could not find container \"f6671679e9ace73e5ea3f2c45db281bbf4a6f45de35633a9d6a5ae4f95080a63\": container with ID starting with f6671679e9ace73e5ea3f2c45db281bbf4a6f45de35633a9d6a5ae4f95080a63 not found: ID does not exist" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.029149 4775 scope.go:117] "RemoveContainer" containerID="47ec028efde4799589be16b23f019bbf495c1ba53eb9982646c05e6d330dc918" Oct 02 03:14:35 crc kubenswrapper[4775]: E1002 03:14:35.030422 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47ec028efde4799589be16b23f019bbf495c1ba53eb9982646c05e6d330dc918\": container with ID starting with 47ec028efde4799589be16b23f019bbf495c1ba53eb9982646c05e6d330dc918 not found: ID does not exist" containerID="47ec028efde4799589be16b23f019bbf495c1ba53eb9982646c05e6d330dc918" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.030467 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47ec028efde4799589be16b23f019bbf495c1ba53eb9982646c05e6d330dc918"} err="failed to get container status \"47ec028efde4799589be16b23f019bbf495c1ba53eb9982646c05e6d330dc918\": rpc error: code = NotFound desc = could not find container \"47ec028efde4799589be16b23f019bbf495c1ba53eb9982646c05e6d330dc918\": container with ID starting with 47ec028efde4799589be16b23f019bbf495c1ba53eb9982646c05e6d330dc918 not found: ID does not exist" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.030494 4775 scope.go:117] "RemoveContainer" containerID="090cf116a3f13f2459729e0fc829eab26a9d3ef8e5eb73f984f9ff4ed301bf4a" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.038602 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.049289 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.057863 4775 scope.go:117] "RemoveContainer" containerID="a37262fc5c77c394d27e02f2b5c649aaa548615bb8aaaa20b5839425834460ea" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.066298 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:35 crc kubenswrapper[4775]: E1002 03:14:35.066819 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f6eb403-8ff0-441e-99a5-4616606f1a04" containerName="nova-metadata-metadata" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.066839 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f6eb403-8ff0-441e-99a5-4616606f1a04" containerName="nova-metadata-metadata" Oct 02 03:14:35 crc kubenswrapper[4775]: E1002 03:14:35.066852 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e09de27e-34c9-4a22-8fc6-aaf72ba6b583" containerName="nova-api-log" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.066859 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e09de27e-34c9-4a22-8fc6-aaf72ba6b583" containerName="nova-api-log" Oct 02 03:14:35 crc kubenswrapper[4775]: E1002 03:14:35.066873 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e09de27e-34c9-4a22-8fc6-aaf72ba6b583" containerName="nova-api-api" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.066879 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e09de27e-34c9-4a22-8fc6-aaf72ba6b583" containerName="nova-api-api" Oct 02 03:14:35 crc kubenswrapper[4775]: E1002 03:14:35.066898 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f6eb403-8ff0-441e-99a5-4616606f1a04" containerName="nova-metadata-log" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.066904 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f6eb403-8ff0-441e-99a5-4616606f1a04" containerName="nova-metadata-log" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.067091 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f6eb403-8ff0-441e-99a5-4616606f1a04" containerName="nova-metadata-log" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.067111 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e09de27e-34c9-4a22-8fc6-aaf72ba6b583" containerName="nova-api-api" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.067125 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e09de27e-34c9-4a22-8fc6-aaf72ba6b583" containerName="nova-api-log" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.067141 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f6eb403-8ff0-441e-99a5-4616606f1a04" containerName="nova-metadata-metadata" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.070739 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.074230 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.076253 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.123478 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-config-data\") pod \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.123520 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-logs\") pod \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.123560 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4k69d\" (UniqueName: \"kubernetes.io/projected/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-kube-api-access-4k69d\") pod \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.123616 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-combined-ca-bundle\") pod \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\" (UID: \"e09de27e-34c9-4a22-8fc6-aaf72ba6b583\") " Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.124119 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-logs" (OuterVolumeSpecName: "logs") pod "e09de27e-34c9-4a22-8fc6-aaf72ba6b583" (UID: "e09de27e-34c9-4a22-8fc6-aaf72ba6b583"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.124331 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.141219 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-kube-api-access-4k69d" (OuterVolumeSpecName: "kube-api-access-4k69d") pod "e09de27e-34c9-4a22-8fc6-aaf72ba6b583" (UID: "e09de27e-34c9-4a22-8fc6-aaf72ba6b583"). InnerVolumeSpecName "kube-api-access-4k69d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.145432 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-config-data" (OuterVolumeSpecName: "config-data") pod "e09de27e-34c9-4a22-8fc6-aaf72ba6b583" (UID: "e09de27e-34c9-4a22-8fc6-aaf72ba6b583"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.161044 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e09de27e-34c9-4a22-8fc6-aaf72ba6b583" (UID: "e09de27e-34c9-4a22-8fc6-aaf72ba6b583"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.225396 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4a80c5f-4c44-46f7-a694-a6904a158bfc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.225496 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2t8dg\" (UniqueName: \"kubernetes.io/projected/a4a80c5f-4c44-46f7-a694-a6904a158bfc-kube-api-access-2t8dg\") pod \"nova-metadata-0\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.225536 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4a80c5f-4c44-46f7-a694-a6904a158bfc-config-data\") pod \"nova-metadata-0\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.225567 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4a80c5f-4c44-46f7-a694-a6904a158bfc-logs\") pod \"nova-metadata-0\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.225629 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.225641 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4k69d\" (UniqueName: \"kubernetes.io/projected/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-kube-api-access-4k69d\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.225691 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e09de27e-34c9-4a22-8fc6-aaf72ba6b583-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.327272 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2t8dg\" (UniqueName: \"kubernetes.io/projected/a4a80c5f-4c44-46f7-a694-a6904a158bfc-kube-api-access-2t8dg\") pod \"nova-metadata-0\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.327363 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4a80c5f-4c44-46f7-a694-a6904a158bfc-config-data\") pod \"nova-metadata-0\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.327413 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4a80c5f-4c44-46f7-a694-a6904a158bfc-logs\") pod \"nova-metadata-0\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.327461 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4a80c5f-4c44-46f7-a694-a6904a158bfc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.328402 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4a80c5f-4c44-46f7-a694-a6904a158bfc-logs\") pod \"nova-metadata-0\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.332441 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4a80c5f-4c44-46f7-a694-a6904a158bfc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.332572 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4a80c5f-4c44-46f7-a694-a6904a158bfc-config-data\") pod \"nova-metadata-0\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.356164 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2t8dg\" (UniqueName: \"kubernetes.io/projected/a4a80c5f-4c44-46f7-a694-a6904a158bfc-kube-api-access-2t8dg\") pod \"nova-metadata-0\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.392978 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.732712 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:14:35 crc kubenswrapper[4775]: W1002 03:14:35.743378 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4a80c5f_4c44_46f7_a694_a6904a158bfc.slice/crio-28ebddd32a3f9f450c32bb5f9789b43d6a18ffea587107b0ea167db9dd649793 WatchSource:0}: Error finding container 28ebddd32a3f9f450c32bb5f9789b43d6a18ffea587107b0ea167db9dd649793: Status 404 returned error can't find the container with id 28ebddd32a3f9f450c32bb5f9789b43d6a18ffea587107b0ea167db9dd649793 Oct 02 03:14:35 crc kubenswrapper[4775]: I1002 03:14:35.785558 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f6eb403-8ff0-441e-99a5-4616606f1a04" path="/var/lib/kubelet/pods/2f6eb403-8ff0-441e-99a5-4616606f1a04/volumes" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.001182 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.004486 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a4a80c5f-4c44-46f7-a694-a6904a158bfc","Type":"ContainerStarted","Data":"64ceeabff7ae4e873af57d8fd56c5a730409ce674f71ff304b80f43af51430d4"} Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.004535 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a4a80c5f-4c44-46f7-a694-a6904a158bfc","Type":"ContainerStarted","Data":"28ebddd32a3f9f450c32bb5f9789b43d6a18ffea587107b0ea167db9dd649793"} Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.073287 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.124861 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.124928 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.131463 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.134286 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.146297 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.248490 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r4kb\" (UniqueName: \"kubernetes.io/projected/690c75e7-b17b-4aba-b6a6-2d969ad8137d-kube-api-access-7r4kb\") pod \"nova-api-0\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.248630 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/690c75e7-b17b-4aba-b6a6-2d969ad8137d-logs\") pod \"nova-api-0\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.248660 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690c75e7-b17b-4aba-b6a6-2d969ad8137d-config-data\") pod \"nova-api-0\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.248721 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690c75e7-b17b-4aba-b6a6-2d969ad8137d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.347989 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.350250 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690c75e7-b17b-4aba-b6a6-2d969ad8137d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.350368 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r4kb\" (UniqueName: \"kubernetes.io/projected/690c75e7-b17b-4aba-b6a6-2d969ad8137d-kube-api-access-7r4kb\") pod \"nova-api-0\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.350535 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/690c75e7-b17b-4aba-b6a6-2d969ad8137d-logs\") pod \"nova-api-0\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.350578 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690c75e7-b17b-4aba-b6a6-2d969ad8137d-config-data\") pod \"nova-api-0\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.351389 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/690c75e7-b17b-4aba-b6a6-2d969ad8137d-logs\") pod \"nova-api-0\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.356446 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690c75e7-b17b-4aba-b6a6-2d969ad8137d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.356508 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690c75e7-b17b-4aba-b6a6-2d969ad8137d-config-data\") pod \"nova-api-0\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.372029 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r4kb\" (UniqueName: \"kubernetes.io/projected/690c75e7-b17b-4aba-b6a6-2d969ad8137d-kube-api-access-7r4kb\") pod \"nova-api-0\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.457263 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 03:14:36 crc kubenswrapper[4775]: I1002 03:14:36.711032 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:14:36 crc kubenswrapper[4775]: W1002 03:14:36.722293 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod690c75e7_b17b_4aba_b6a6_2d969ad8137d.slice/crio-8fec26988832127ac7a3349ec5ae25c6760c24b9af4088b4735862c9ae586374 WatchSource:0}: Error finding container 8fec26988832127ac7a3349ec5ae25c6760c24b9af4088b4735862c9ae586374: Status 404 returned error can't find the container with id 8fec26988832127ac7a3349ec5ae25c6760c24b9af4088b4735862c9ae586374 Oct 02 03:14:37 crc kubenswrapper[4775]: I1002 03:14:37.014620 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"690c75e7-b17b-4aba-b6a6-2d969ad8137d","Type":"ContainerStarted","Data":"81ce762c940c90ea0299fd779112ade0eab09c4e18b02b0bd4b3270faeb3da57"} Oct 02 03:14:37 crc kubenswrapper[4775]: I1002 03:14:37.014658 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"690c75e7-b17b-4aba-b6a6-2d969ad8137d","Type":"ContainerStarted","Data":"8fec26988832127ac7a3349ec5ae25c6760c24b9af4088b4735862c9ae586374"} Oct 02 03:14:37 crc kubenswrapper[4775]: I1002 03:14:37.016689 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a4a80c5f-4c44-46f7-a694-a6904a158bfc","Type":"ContainerStarted","Data":"fe7fd197864de5e576d1dd56a195c436013c7f5ce89446da93b812dc73917bb4"} Oct 02 03:14:37 crc kubenswrapper[4775]: I1002 03:14:37.049697 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.04967994 podStartE2EDuration="2.04967994s" podCreationTimestamp="2025-10-02 03:14:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:37.04192716 +0000 UTC m=+5614.208671200" watchObservedRunningTime="2025-10-02 03:14:37.04967994 +0000 UTC m=+5614.216423980" Oct 02 03:14:37 crc kubenswrapper[4775]: I1002 03:14:37.240669 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:14:37 crc kubenswrapper[4775]: I1002 03:14:37.241094 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:14:37 crc kubenswrapper[4775]: I1002 03:14:37.783492 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e09de27e-34c9-4a22-8fc6-aaf72ba6b583" path="/var/lib/kubelet/pods/e09de27e-34c9-4a22-8fc6-aaf72ba6b583/volumes" Oct 02 03:14:38 crc kubenswrapper[4775]: I1002 03:14:38.031605 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"690c75e7-b17b-4aba-b6a6-2d969ad8137d","Type":"ContainerStarted","Data":"2aa7bec967ea4a34d7201995bd474c11c9a3aa1f616f8053076acda37a854993"} Oct 02 03:14:38 crc kubenswrapper[4775]: I1002 03:14:38.057481 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.05745654 podStartE2EDuration="2.05745654s" podCreationTimestamp="2025-10-02 03:14:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:14:38.050395767 +0000 UTC m=+5615.217139827" watchObservedRunningTime="2025-10-02 03:14:38.05745654 +0000 UTC m=+5615.224200590" Oct 02 03:14:40 crc kubenswrapper[4775]: I1002 03:14:40.394093 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 03:14:40 crc kubenswrapper[4775]: I1002 03:14:40.394593 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 03:14:41 crc kubenswrapper[4775]: I1002 03:14:41.348240 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 03:14:41 crc kubenswrapper[4775]: I1002 03:14:41.396910 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 03:14:42 crc kubenswrapper[4775]: I1002 03:14:42.127833 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 03:14:45 crc kubenswrapper[4775]: I1002 03:14:45.393547 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 03:14:45 crc kubenswrapper[4775]: I1002 03:14:45.394099 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 03:14:46 crc kubenswrapper[4775]: I1002 03:14:46.457581 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 03:14:46 crc kubenswrapper[4775]: I1002 03:14:46.457922 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 03:14:46 crc kubenswrapper[4775]: I1002 03:14:46.476226 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.73:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 03:14:46 crc kubenswrapper[4775]: I1002 03:14:46.477987 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.73:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 03:14:47 crc kubenswrapper[4775]: I1002 03:14:47.540173 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="690c75e7-b17b-4aba-b6a6-2d969ad8137d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.74:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 03:14:47 crc kubenswrapper[4775]: I1002 03:14:47.540248 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="690c75e7-b17b-4aba-b6a6-2d969ad8137d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.74:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 03:14:55 crc kubenswrapper[4775]: I1002 03:14:55.396277 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 03:14:55 crc kubenswrapper[4775]: I1002 03:14:55.398428 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 03:14:55 crc kubenswrapper[4775]: I1002 03:14:55.400303 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 03:14:56 crc kubenswrapper[4775]: I1002 03:14:56.244985 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 03:14:56 crc kubenswrapper[4775]: I1002 03:14:56.462385 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 03:14:56 crc kubenswrapper[4775]: I1002 03:14:56.462781 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 03:14:56 crc kubenswrapper[4775]: I1002 03:14:56.463388 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 03:14:56 crc kubenswrapper[4775]: I1002 03:14:56.465811 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.253289 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.257894 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.472139 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-854565669c-g7wps"] Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.477767 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.486778 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-854565669c-g7wps"] Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.602497 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-dns-svc\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.602585 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-config\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.602628 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjzv8\" (UniqueName: \"kubernetes.io/projected/8eb457c8-28e9-4955-b9db-775c9e59634a-kube-api-access-rjzv8\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.602753 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-ovsdbserver-nb\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.602800 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-ovsdbserver-sb\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.704076 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-dns-svc\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.704128 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-config\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.704153 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjzv8\" (UniqueName: \"kubernetes.io/projected/8eb457c8-28e9-4955-b9db-775c9e59634a-kube-api-access-rjzv8\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.704220 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-ovsdbserver-nb\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.704251 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-ovsdbserver-sb\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.705047 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-ovsdbserver-sb\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.705563 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-config\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.705727 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-dns-svc\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.706917 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-ovsdbserver-nb\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.752911 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjzv8\" (UniqueName: \"kubernetes.io/projected/8eb457c8-28e9-4955-b9db-775c9e59634a-kube-api-access-rjzv8\") pod \"dnsmasq-dns-854565669c-g7wps\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:57 crc kubenswrapper[4775]: I1002 03:14:57.801369 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:14:58 crc kubenswrapper[4775]: I1002 03:14:58.388850 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-854565669c-g7wps"] Oct 02 03:14:59 crc kubenswrapper[4775]: I1002 03:14:59.288200 4775 generic.go:334] "Generic (PLEG): container finished" podID="8eb457c8-28e9-4955-b9db-775c9e59634a" containerID="f81a3646cdba9c0f9249e5d57d329a8ab236821f2bae4c01cbe49f41e0437f0b" exitCode=0 Oct 02 03:14:59 crc kubenswrapper[4775]: I1002 03:14:59.288319 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-854565669c-g7wps" event={"ID":"8eb457c8-28e9-4955-b9db-775c9e59634a","Type":"ContainerDied","Data":"f81a3646cdba9c0f9249e5d57d329a8ab236821f2bae4c01cbe49f41e0437f0b"} Oct 02 03:14:59 crc kubenswrapper[4775]: I1002 03:14:59.289514 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-854565669c-g7wps" event={"ID":"8eb457c8-28e9-4955-b9db-775c9e59634a","Type":"ContainerStarted","Data":"8a3deda0866e2841f2f14ec002e874dca613d104773324b5bc53c064b7e010f0"} Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.133413 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc"] Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.135434 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.137691 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.137694 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.144205 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc"] Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.264480 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-config-volume\") pod \"collect-profiles-29322915-sp6bc\" (UID: \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.264585 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chs75\" (UniqueName: \"kubernetes.io/projected/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-kube-api-access-chs75\") pod \"collect-profiles-29322915-sp6bc\" (UID: \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.264616 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-secret-volume\") pod \"collect-profiles-29322915-sp6bc\" (UID: \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.300993 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-854565669c-g7wps" event={"ID":"8eb457c8-28e9-4955-b9db-775c9e59634a","Type":"ContainerStarted","Data":"a8e4faa64780b7e2ce07f300e8ea4f33e62a0b8ea95320e75a9acc09a85c0ae1"} Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.301236 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.328666 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-854565669c-g7wps" podStartSLOduration=3.328642353 podStartE2EDuration="3.328642353s" podCreationTimestamp="2025-10-02 03:14:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:15:00.320782569 +0000 UTC m=+5637.487526609" watchObservedRunningTime="2025-10-02 03:15:00.328642353 +0000 UTC m=+5637.495386393" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.366890 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-config-volume\") pod \"collect-profiles-29322915-sp6bc\" (UID: \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.367304 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chs75\" (UniqueName: \"kubernetes.io/projected/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-kube-api-access-chs75\") pod \"collect-profiles-29322915-sp6bc\" (UID: \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.367435 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-secret-volume\") pod \"collect-profiles-29322915-sp6bc\" (UID: \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.368006 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-config-volume\") pod \"collect-profiles-29322915-sp6bc\" (UID: \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.377009 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-secret-volume\") pod \"collect-profiles-29322915-sp6bc\" (UID: \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.385199 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chs75\" (UniqueName: \"kubernetes.io/projected/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-kube-api-access-chs75\") pod \"collect-profiles-29322915-sp6bc\" (UID: \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.463409 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" Oct 02 03:15:00 crc kubenswrapper[4775]: I1002 03:15:00.955534 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc"] Oct 02 03:15:00 crc kubenswrapper[4775]: W1002 03:15:00.974109 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf9df4f5_dbb7_41f3_aff8_1b86fed890a0.slice/crio-b10a2bf2bb53a07fc3bb9c81e5acf51c841b488d126aeeb62db27fb210d744bd WatchSource:0}: Error finding container b10a2bf2bb53a07fc3bb9c81e5acf51c841b488d126aeeb62db27fb210d744bd: Status 404 returned error can't find the container with id b10a2bf2bb53a07fc3bb9c81e5acf51c841b488d126aeeb62db27fb210d744bd Oct 02 03:15:01 crc kubenswrapper[4775]: I1002 03:15:01.312291 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" event={"ID":"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0","Type":"ContainerStarted","Data":"82b6bf51ffc146bca85b24301f4c020584752fd5b210ba7c3f32194dc21f434a"} Oct 02 03:15:01 crc kubenswrapper[4775]: I1002 03:15:01.313506 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" event={"ID":"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0","Type":"ContainerStarted","Data":"b10a2bf2bb53a07fc3bb9c81e5acf51c841b488d126aeeb62db27fb210d744bd"} Oct 02 03:15:01 crc kubenswrapper[4775]: I1002 03:15:01.344986 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" podStartSLOduration=1.344934674 podStartE2EDuration="1.344934674s" podCreationTimestamp="2025-10-02 03:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:15:01.332109312 +0000 UTC m=+5638.498853352" watchObservedRunningTime="2025-10-02 03:15:01.344934674 +0000 UTC m=+5638.511678744" Oct 02 03:15:02 crc kubenswrapper[4775]: I1002 03:15:02.343051 4775 generic.go:334] "Generic (PLEG): container finished" podID="bf9df4f5-dbb7-41f3-aff8-1b86fed890a0" containerID="82b6bf51ffc146bca85b24301f4c020584752fd5b210ba7c3f32194dc21f434a" exitCode=0 Oct 02 03:15:02 crc kubenswrapper[4775]: I1002 03:15:02.343396 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" event={"ID":"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0","Type":"ContainerDied","Data":"82b6bf51ffc146bca85b24301f4c020584752fd5b210ba7c3f32194dc21f434a"} Oct 02 03:15:03 crc kubenswrapper[4775]: I1002 03:15:03.816949 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" Oct 02 03:15:03 crc kubenswrapper[4775]: I1002 03:15:03.949192 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chs75\" (UniqueName: \"kubernetes.io/projected/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-kube-api-access-chs75\") pod \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\" (UID: \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\") " Oct 02 03:15:03 crc kubenswrapper[4775]: I1002 03:15:03.950247 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-config-volume\") pod \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\" (UID: \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\") " Oct 02 03:15:03 crc kubenswrapper[4775]: I1002 03:15:03.950331 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-secret-volume\") pod \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\" (UID: \"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0\") " Oct 02 03:15:03 crc kubenswrapper[4775]: I1002 03:15:03.951489 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-config-volume" (OuterVolumeSpecName: "config-volume") pod "bf9df4f5-dbb7-41f3-aff8-1b86fed890a0" (UID: "bf9df4f5-dbb7-41f3-aff8-1b86fed890a0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:15:03 crc kubenswrapper[4775]: I1002 03:15:03.958708 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "bf9df4f5-dbb7-41f3-aff8-1b86fed890a0" (UID: "bf9df4f5-dbb7-41f3-aff8-1b86fed890a0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:03 crc kubenswrapper[4775]: I1002 03:15:03.961368 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-kube-api-access-chs75" (OuterVolumeSpecName: "kube-api-access-chs75") pod "bf9df4f5-dbb7-41f3-aff8-1b86fed890a0" (UID: "bf9df4f5-dbb7-41f3-aff8-1b86fed890a0"). InnerVolumeSpecName "kube-api-access-chs75". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:15:04 crc kubenswrapper[4775]: I1002 03:15:04.052829 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chs75\" (UniqueName: \"kubernetes.io/projected/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-kube-api-access-chs75\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:04 crc kubenswrapper[4775]: I1002 03:15:04.052868 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:04 crc kubenswrapper[4775]: I1002 03:15:04.052885 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:04 crc kubenswrapper[4775]: I1002 03:15:04.369854 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" event={"ID":"bf9df4f5-dbb7-41f3-aff8-1b86fed890a0","Type":"ContainerDied","Data":"b10a2bf2bb53a07fc3bb9c81e5acf51c841b488d126aeeb62db27fb210d744bd"} Oct 02 03:15:04 crc kubenswrapper[4775]: I1002 03:15:04.369909 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b10a2bf2bb53a07fc3bb9c81e5acf51c841b488d126aeeb62db27fb210d744bd" Oct 02 03:15:04 crc kubenswrapper[4775]: I1002 03:15:04.369979 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc" Oct 02 03:15:04 crc kubenswrapper[4775]: I1002 03:15:04.431734 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95"] Oct 02 03:15:04 crc kubenswrapper[4775]: I1002 03:15:04.437822 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322870-6lc95"] Oct 02 03:15:05 crc kubenswrapper[4775]: I1002 03:15:05.787513 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5" path="/var/lib/kubelet/pods/aa175dd1-85e7-4d2d-9ce2-7ece5b688cc5/volumes" Oct 02 03:15:07 crc kubenswrapper[4775]: I1002 03:15:07.234067 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:15:07 crc kubenswrapper[4775]: I1002 03:15:07.234142 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:15:07 crc kubenswrapper[4775]: I1002 03:15:07.804201 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:15:07 crc kubenswrapper[4775]: I1002 03:15:07.878653 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-584bf664f7-7fqhb"] Oct 02 03:15:07 crc kubenswrapper[4775]: I1002 03:15:07.878907 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" podUID="8473d923-d46f-40e2-9f75-58e14833dadd" containerName="dnsmasq-dns" containerID="cri-o://bae7647a6403874672d4069f85fa0fc18c870bf6bea9eaa4a263f8242571c0a9" gracePeriod=10 Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.413231 4775 generic.go:334] "Generic (PLEG): container finished" podID="8473d923-d46f-40e2-9f75-58e14833dadd" containerID="bae7647a6403874672d4069f85fa0fc18c870bf6bea9eaa4a263f8242571c0a9" exitCode=0 Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.413320 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" event={"ID":"8473d923-d46f-40e2-9f75-58e14833dadd","Type":"ContainerDied","Data":"bae7647a6403874672d4069f85fa0fc18c870bf6bea9eaa4a263f8242571c0a9"} Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.413553 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" event={"ID":"8473d923-d46f-40e2-9f75-58e14833dadd","Type":"ContainerDied","Data":"07563fa4313842d2f16c1ea76155e4c9c1fafca77d7e19362b794863218da10c"} Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.413568 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="07563fa4313842d2f16c1ea76155e4c9c1fafca77d7e19362b794863218da10c" Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.433364 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.542587 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-dns-svc\") pod \"8473d923-d46f-40e2-9f75-58e14833dadd\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.542631 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-config\") pod \"8473d923-d46f-40e2-9f75-58e14833dadd\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.542716 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-ovsdbserver-nb\") pod \"8473d923-d46f-40e2-9f75-58e14833dadd\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.542736 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-ovsdbserver-sb\") pod \"8473d923-d46f-40e2-9f75-58e14833dadd\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.542765 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmtt4\" (UniqueName: \"kubernetes.io/projected/8473d923-d46f-40e2-9f75-58e14833dadd-kube-api-access-vmtt4\") pod \"8473d923-d46f-40e2-9f75-58e14833dadd\" (UID: \"8473d923-d46f-40e2-9f75-58e14833dadd\") " Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.547725 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8473d923-d46f-40e2-9f75-58e14833dadd-kube-api-access-vmtt4" (OuterVolumeSpecName: "kube-api-access-vmtt4") pod "8473d923-d46f-40e2-9f75-58e14833dadd" (UID: "8473d923-d46f-40e2-9f75-58e14833dadd"). InnerVolumeSpecName "kube-api-access-vmtt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.586909 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-config" (OuterVolumeSpecName: "config") pod "8473d923-d46f-40e2-9f75-58e14833dadd" (UID: "8473d923-d46f-40e2-9f75-58e14833dadd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.597908 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8473d923-d46f-40e2-9f75-58e14833dadd" (UID: "8473d923-d46f-40e2-9f75-58e14833dadd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.605575 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8473d923-d46f-40e2-9f75-58e14833dadd" (UID: "8473d923-d46f-40e2-9f75-58e14833dadd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.612472 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8473d923-d46f-40e2-9f75-58e14833dadd" (UID: "8473d923-d46f-40e2-9f75-58e14833dadd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.644443 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.644482 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.644497 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.644510 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8473d923-d46f-40e2-9f75-58e14833dadd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:08 crc kubenswrapper[4775]: I1002 03:15:08.644523 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmtt4\" (UniqueName: \"kubernetes.io/projected/8473d923-d46f-40e2-9f75-58e14833dadd-kube-api-access-vmtt4\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:09 crc kubenswrapper[4775]: I1002 03:15:09.425517 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-584bf664f7-7fqhb" Oct 02 03:15:09 crc kubenswrapper[4775]: I1002 03:15:09.500066 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-584bf664f7-7fqhb"] Oct 02 03:15:09 crc kubenswrapper[4775]: I1002 03:15:09.511967 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-584bf664f7-7fqhb"] Oct 02 03:15:09 crc kubenswrapper[4775]: E1002 03:15:09.663741 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8473d923_d46f_40e2_9f75_58e14833dadd.slice/crio-07563fa4313842d2f16c1ea76155e4c9c1fafca77d7e19362b794863218da10c\": RecentStats: unable to find data in memory cache]" Oct 02 03:15:09 crc kubenswrapper[4775]: I1002 03:15:09.776892 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8473d923-d46f-40e2-9f75-58e14833dadd" path="/var/lib/kubelet/pods/8473d923-d46f-40e2-9f75-58e14833dadd/volumes" Oct 02 03:15:11 crc kubenswrapper[4775]: I1002 03:15:11.523215 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-cpx6g"] Oct 02 03:15:11 crc kubenswrapper[4775]: E1002 03:15:11.523648 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf9df4f5-dbb7-41f3-aff8-1b86fed890a0" containerName="collect-profiles" Oct 02 03:15:11 crc kubenswrapper[4775]: I1002 03:15:11.523662 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf9df4f5-dbb7-41f3-aff8-1b86fed890a0" containerName="collect-profiles" Oct 02 03:15:11 crc kubenswrapper[4775]: E1002 03:15:11.523682 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8473d923-d46f-40e2-9f75-58e14833dadd" containerName="dnsmasq-dns" Oct 02 03:15:11 crc kubenswrapper[4775]: I1002 03:15:11.523687 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8473d923-d46f-40e2-9f75-58e14833dadd" containerName="dnsmasq-dns" Oct 02 03:15:11 crc kubenswrapper[4775]: E1002 03:15:11.523713 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8473d923-d46f-40e2-9f75-58e14833dadd" containerName="init" Oct 02 03:15:11 crc kubenswrapper[4775]: I1002 03:15:11.523720 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8473d923-d46f-40e2-9f75-58e14833dadd" containerName="init" Oct 02 03:15:11 crc kubenswrapper[4775]: I1002 03:15:11.523876 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf9df4f5-dbb7-41f3-aff8-1b86fed890a0" containerName="collect-profiles" Oct 02 03:15:11 crc kubenswrapper[4775]: I1002 03:15:11.523903 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8473d923-d46f-40e2-9f75-58e14833dadd" containerName="dnsmasq-dns" Oct 02 03:15:11 crc kubenswrapper[4775]: I1002 03:15:11.524601 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cpx6g" Oct 02 03:15:11 crc kubenswrapper[4775]: I1002 03:15:11.538520 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-cpx6g"] Oct 02 03:15:11 crc kubenswrapper[4775]: I1002 03:15:11.539889 4775 scope.go:117] "RemoveContainer" containerID="eae24cc858e46a80e35ec9e64496af5a44f475e2b250320c7378f243778f2f5e" Oct 02 03:15:11 crc kubenswrapper[4775]: I1002 03:15:11.640182 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ltv5\" (UniqueName: \"kubernetes.io/projected/84ad728b-c2c6-44a3-8fb9-23a29a82ef65-kube-api-access-6ltv5\") pod \"cinder-db-create-cpx6g\" (UID: \"84ad728b-c2c6-44a3-8fb9-23a29a82ef65\") " pod="openstack/cinder-db-create-cpx6g" Oct 02 03:15:11 crc kubenswrapper[4775]: I1002 03:15:11.742710 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ltv5\" (UniqueName: \"kubernetes.io/projected/84ad728b-c2c6-44a3-8fb9-23a29a82ef65-kube-api-access-6ltv5\") pod \"cinder-db-create-cpx6g\" (UID: \"84ad728b-c2c6-44a3-8fb9-23a29a82ef65\") " pod="openstack/cinder-db-create-cpx6g" Oct 02 03:15:11 crc kubenswrapper[4775]: I1002 03:15:11.765427 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ltv5\" (UniqueName: \"kubernetes.io/projected/84ad728b-c2c6-44a3-8fb9-23a29a82ef65-kube-api-access-6ltv5\") pod \"cinder-db-create-cpx6g\" (UID: \"84ad728b-c2c6-44a3-8fb9-23a29a82ef65\") " pod="openstack/cinder-db-create-cpx6g" Oct 02 03:15:11 crc kubenswrapper[4775]: I1002 03:15:11.856296 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cpx6g" Oct 02 03:15:12 crc kubenswrapper[4775]: I1002 03:15:12.336492 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-cpx6g"] Oct 02 03:15:12 crc kubenswrapper[4775]: I1002 03:15:12.470494 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cpx6g" event={"ID":"84ad728b-c2c6-44a3-8fb9-23a29a82ef65","Type":"ContainerStarted","Data":"e725a568df6b7eccc0ddd8ff1fd587def00f988bc3b3fc00d2cb2038716383ef"} Oct 02 03:15:13 crc kubenswrapper[4775]: I1002 03:15:13.491525 4775 generic.go:334] "Generic (PLEG): container finished" podID="84ad728b-c2c6-44a3-8fb9-23a29a82ef65" containerID="4b52b5567f3c3250afa5f90b477c6ddb3b3c6791df1fc3a566eb950d56971c4e" exitCode=0 Oct 02 03:15:13 crc kubenswrapper[4775]: I1002 03:15:13.491600 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cpx6g" event={"ID":"84ad728b-c2c6-44a3-8fb9-23a29a82ef65","Type":"ContainerDied","Data":"4b52b5567f3c3250afa5f90b477c6ddb3b3c6791df1fc3a566eb950d56971c4e"} Oct 02 03:15:14 crc kubenswrapper[4775]: I1002 03:15:14.948657 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cpx6g" Oct 02 03:15:15 crc kubenswrapper[4775]: I1002 03:15:15.023410 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ltv5\" (UniqueName: \"kubernetes.io/projected/84ad728b-c2c6-44a3-8fb9-23a29a82ef65-kube-api-access-6ltv5\") pod \"84ad728b-c2c6-44a3-8fb9-23a29a82ef65\" (UID: \"84ad728b-c2c6-44a3-8fb9-23a29a82ef65\") " Oct 02 03:15:15 crc kubenswrapper[4775]: I1002 03:15:15.032065 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84ad728b-c2c6-44a3-8fb9-23a29a82ef65-kube-api-access-6ltv5" (OuterVolumeSpecName: "kube-api-access-6ltv5") pod "84ad728b-c2c6-44a3-8fb9-23a29a82ef65" (UID: "84ad728b-c2c6-44a3-8fb9-23a29a82ef65"). InnerVolumeSpecName "kube-api-access-6ltv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:15:15 crc kubenswrapper[4775]: I1002 03:15:15.125765 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ltv5\" (UniqueName: \"kubernetes.io/projected/84ad728b-c2c6-44a3-8fb9-23a29a82ef65-kube-api-access-6ltv5\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:15 crc kubenswrapper[4775]: I1002 03:15:15.520547 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-cpx6g" event={"ID":"84ad728b-c2c6-44a3-8fb9-23a29a82ef65","Type":"ContainerDied","Data":"e725a568df6b7eccc0ddd8ff1fd587def00f988bc3b3fc00d2cb2038716383ef"} Oct 02 03:15:15 crc kubenswrapper[4775]: I1002 03:15:15.520604 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-cpx6g" Oct 02 03:15:15 crc kubenswrapper[4775]: I1002 03:15:15.520604 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e725a568df6b7eccc0ddd8ff1fd587def00f988bc3b3fc00d2cb2038716383ef" Oct 02 03:15:21 crc kubenswrapper[4775]: I1002 03:15:21.660368 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-f5a2-account-create-bwcrk"] Oct 02 03:15:21 crc kubenswrapper[4775]: E1002 03:15:21.661609 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84ad728b-c2c6-44a3-8fb9-23a29a82ef65" containerName="mariadb-database-create" Oct 02 03:15:21 crc kubenswrapper[4775]: I1002 03:15:21.661631 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="84ad728b-c2c6-44a3-8fb9-23a29a82ef65" containerName="mariadb-database-create" Oct 02 03:15:21 crc kubenswrapper[4775]: I1002 03:15:21.661997 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="84ad728b-c2c6-44a3-8fb9-23a29a82ef65" containerName="mariadb-database-create" Oct 02 03:15:21 crc kubenswrapper[4775]: I1002 03:15:21.662950 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f5a2-account-create-bwcrk" Oct 02 03:15:21 crc kubenswrapper[4775]: I1002 03:15:21.665906 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 02 03:15:21 crc kubenswrapper[4775]: I1002 03:15:21.677564 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f5a2-account-create-bwcrk"] Oct 02 03:15:21 crc kubenswrapper[4775]: I1002 03:15:21.834881 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhdb9\" (UniqueName: \"kubernetes.io/projected/013cf425-7401-4e5d-979f-d63a79d9db31-kube-api-access-jhdb9\") pod \"cinder-f5a2-account-create-bwcrk\" (UID: \"013cf425-7401-4e5d-979f-d63a79d9db31\") " pod="openstack/cinder-f5a2-account-create-bwcrk" Oct 02 03:15:21 crc kubenswrapper[4775]: I1002 03:15:21.936517 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhdb9\" (UniqueName: \"kubernetes.io/projected/013cf425-7401-4e5d-979f-d63a79d9db31-kube-api-access-jhdb9\") pod \"cinder-f5a2-account-create-bwcrk\" (UID: \"013cf425-7401-4e5d-979f-d63a79d9db31\") " pod="openstack/cinder-f5a2-account-create-bwcrk" Oct 02 03:15:21 crc kubenswrapper[4775]: I1002 03:15:21.963318 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhdb9\" (UniqueName: \"kubernetes.io/projected/013cf425-7401-4e5d-979f-d63a79d9db31-kube-api-access-jhdb9\") pod \"cinder-f5a2-account-create-bwcrk\" (UID: \"013cf425-7401-4e5d-979f-d63a79d9db31\") " pod="openstack/cinder-f5a2-account-create-bwcrk" Oct 02 03:15:21 crc kubenswrapper[4775]: I1002 03:15:21.992305 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f5a2-account-create-bwcrk" Oct 02 03:15:22 crc kubenswrapper[4775]: I1002 03:15:22.493477 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-f5a2-account-create-bwcrk"] Oct 02 03:15:22 crc kubenswrapper[4775]: I1002 03:15:22.606738 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f5a2-account-create-bwcrk" event={"ID":"013cf425-7401-4e5d-979f-d63a79d9db31","Type":"ContainerStarted","Data":"468a5b31e9e82c8b269d1d1d7f5dba62aa6bb7094b6843992aa7f5191721ff25"} Oct 02 03:15:23 crc kubenswrapper[4775]: I1002 03:15:23.623056 4775 generic.go:334] "Generic (PLEG): container finished" podID="013cf425-7401-4e5d-979f-d63a79d9db31" containerID="13f0c894d70e585132b7ce1ad0445019d51afe10558d20a630b589d1b45956be" exitCode=0 Oct 02 03:15:23 crc kubenswrapper[4775]: I1002 03:15:23.623143 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f5a2-account-create-bwcrk" event={"ID":"013cf425-7401-4e5d-979f-d63a79d9db31","Type":"ContainerDied","Data":"13f0c894d70e585132b7ce1ad0445019d51afe10558d20a630b589d1b45956be"} Oct 02 03:15:25 crc kubenswrapper[4775]: I1002 03:15:25.045035 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f5a2-account-create-bwcrk" Oct 02 03:15:25 crc kubenswrapper[4775]: I1002 03:15:25.109132 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhdb9\" (UniqueName: \"kubernetes.io/projected/013cf425-7401-4e5d-979f-d63a79d9db31-kube-api-access-jhdb9\") pod \"013cf425-7401-4e5d-979f-d63a79d9db31\" (UID: \"013cf425-7401-4e5d-979f-d63a79d9db31\") " Oct 02 03:15:25 crc kubenswrapper[4775]: I1002 03:15:25.117684 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/013cf425-7401-4e5d-979f-d63a79d9db31-kube-api-access-jhdb9" (OuterVolumeSpecName: "kube-api-access-jhdb9") pod "013cf425-7401-4e5d-979f-d63a79d9db31" (UID: "013cf425-7401-4e5d-979f-d63a79d9db31"). InnerVolumeSpecName "kube-api-access-jhdb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:15:25 crc kubenswrapper[4775]: I1002 03:15:25.210594 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhdb9\" (UniqueName: \"kubernetes.io/projected/013cf425-7401-4e5d-979f-d63a79d9db31-kube-api-access-jhdb9\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:25 crc kubenswrapper[4775]: I1002 03:15:25.655320 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-f5a2-account-create-bwcrk" event={"ID":"013cf425-7401-4e5d-979f-d63a79d9db31","Type":"ContainerDied","Data":"468a5b31e9e82c8b269d1d1d7f5dba62aa6bb7094b6843992aa7f5191721ff25"} Oct 02 03:15:25 crc kubenswrapper[4775]: I1002 03:15:25.656008 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="468a5b31e9e82c8b269d1d1d7f5dba62aa6bb7094b6843992aa7f5191721ff25" Oct 02 03:15:25 crc kubenswrapper[4775]: I1002 03:15:25.655404 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-f5a2-account-create-bwcrk" Oct 02 03:15:26 crc kubenswrapper[4775]: I1002 03:15:26.900942 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-fklbq"] Oct 02 03:15:26 crc kubenswrapper[4775]: E1002 03:15:26.901449 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="013cf425-7401-4e5d-979f-d63a79d9db31" containerName="mariadb-account-create" Oct 02 03:15:26 crc kubenswrapper[4775]: I1002 03:15:26.901466 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="013cf425-7401-4e5d-979f-d63a79d9db31" containerName="mariadb-account-create" Oct 02 03:15:26 crc kubenswrapper[4775]: I1002 03:15:26.901714 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="013cf425-7401-4e5d-979f-d63a79d9db31" containerName="mariadb-account-create" Oct 02 03:15:26 crc kubenswrapper[4775]: I1002 03:15:26.902487 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:26 crc kubenswrapper[4775]: I1002 03:15:26.905376 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 03:15:26 crc kubenswrapper[4775]: I1002 03:15:26.905632 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 03:15:26 crc kubenswrapper[4775]: I1002 03:15:26.905447 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9zxst" Oct 02 03:15:26 crc kubenswrapper[4775]: I1002 03:15:26.911905 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-fklbq"] Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.048199 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af1edb92-72cb-46db-9ad8-6e22d9206bc5-etc-machine-id\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.048293 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-config-data\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.048352 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-scripts\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.048531 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qb5b6\" (UniqueName: \"kubernetes.io/projected/af1edb92-72cb-46db-9ad8-6e22d9206bc5-kube-api-access-qb5b6\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.048584 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-combined-ca-bundle\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.048971 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-db-sync-config-data\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.150542 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qb5b6\" (UniqueName: \"kubernetes.io/projected/af1edb92-72cb-46db-9ad8-6e22d9206bc5-kube-api-access-qb5b6\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.150622 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-combined-ca-bundle\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.150741 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-db-sync-config-data\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.150825 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af1edb92-72cb-46db-9ad8-6e22d9206bc5-etc-machine-id\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.150857 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-config-data\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.150900 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-scripts\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.151041 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af1edb92-72cb-46db-9ad8-6e22d9206bc5-etc-machine-id\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.159295 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-combined-ca-bundle\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.160830 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-config-data\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.167094 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-db-sync-config-data\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.169610 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-scripts\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.175935 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qb5b6\" (UniqueName: \"kubernetes.io/projected/af1edb92-72cb-46db-9ad8-6e22d9206bc5-kube-api-access-qb5b6\") pod \"cinder-db-sync-fklbq\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.255764 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:27 crc kubenswrapper[4775]: W1002 03:15:27.526161 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf1edb92_72cb_46db_9ad8_6e22d9206bc5.slice/crio-90fbceb4c5c04827ab46e94513d1239af97c666e65ebba081145b5759eedd128 WatchSource:0}: Error finding container 90fbceb4c5c04827ab46e94513d1239af97c666e65ebba081145b5759eedd128: Status 404 returned error can't find the container with id 90fbceb4c5c04827ab46e94513d1239af97c666e65ebba081145b5759eedd128 Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.531250 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-fklbq"] Oct 02 03:15:27 crc kubenswrapper[4775]: I1002 03:15:27.692179 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fklbq" event={"ID":"af1edb92-72cb-46db-9ad8-6e22d9206bc5","Type":"ContainerStarted","Data":"90fbceb4c5c04827ab46e94513d1239af97c666e65ebba081145b5759eedd128"} Oct 02 03:15:28 crc kubenswrapper[4775]: I1002 03:15:28.720149 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fklbq" event={"ID":"af1edb92-72cb-46db-9ad8-6e22d9206bc5","Type":"ContainerStarted","Data":"b9c22cb65f85af7868611432cb1a21b0ca9aeaef6bf00ce5287f0651041d5055"} Oct 02 03:15:28 crc kubenswrapper[4775]: I1002 03:15:28.752826 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-fklbq" podStartSLOduration=2.75280007 podStartE2EDuration="2.75280007s" podCreationTimestamp="2025-10-02 03:15:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:15:28.744043373 +0000 UTC m=+5665.910787413" watchObservedRunningTime="2025-10-02 03:15:28.75280007 +0000 UTC m=+5665.919544140" Oct 02 03:15:30 crc kubenswrapper[4775]: I1002 03:15:30.745727 4775 generic.go:334] "Generic (PLEG): container finished" podID="af1edb92-72cb-46db-9ad8-6e22d9206bc5" containerID="b9c22cb65f85af7868611432cb1a21b0ca9aeaef6bf00ce5287f0651041d5055" exitCode=0 Oct 02 03:15:30 crc kubenswrapper[4775]: I1002 03:15:30.745922 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fklbq" event={"ID":"af1edb92-72cb-46db-9ad8-6e22d9206bc5","Type":"ContainerDied","Data":"b9c22cb65f85af7868611432cb1a21b0ca9aeaef6bf00ce5287f0651041d5055"} Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.185283 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.200554 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qb5b6\" (UniqueName: \"kubernetes.io/projected/af1edb92-72cb-46db-9ad8-6e22d9206bc5-kube-api-access-qb5b6\") pod \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.200796 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af1edb92-72cb-46db-9ad8-6e22d9206bc5-etc-machine-id\") pod \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.201007 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-scripts\") pod \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.201071 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-config-data\") pod \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.201119 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af1edb92-72cb-46db-9ad8-6e22d9206bc5-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "af1edb92-72cb-46db-9ad8-6e22d9206bc5" (UID: "af1edb92-72cb-46db-9ad8-6e22d9206bc5"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.201151 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-db-sync-config-data\") pod \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.201287 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-combined-ca-bundle\") pod \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\" (UID: \"af1edb92-72cb-46db-9ad8-6e22d9206bc5\") " Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.210605 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af1edb92-72cb-46db-9ad8-6e22d9206bc5-kube-api-access-qb5b6" (OuterVolumeSpecName: "kube-api-access-qb5b6") pod "af1edb92-72cb-46db-9ad8-6e22d9206bc5" (UID: "af1edb92-72cb-46db-9ad8-6e22d9206bc5"). InnerVolumeSpecName "kube-api-access-qb5b6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.215433 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "af1edb92-72cb-46db-9ad8-6e22d9206bc5" (UID: "af1edb92-72cb-46db-9ad8-6e22d9206bc5"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.219308 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qb5b6\" (UniqueName: \"kubernetes.io/projected/af1edb92-72cb-46db-9ad8-6e22d9206bc5-kube-api-access-qb5b6\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.221767 4775 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/af1edb92-72cb-46db-9ad8-6e22d9206bc5-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.222161 4775 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.221238 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-scripts" (OuterVolumeSpecName: "scripts") pod "af1edb92-72cb-46db-9ad8-6e22d9206bc5" (UID: "af1edb92-72cb-46db-9ad8-6e22d9206bc5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.241134 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af1edb92-72cb-46db-9ad8-6e22d9206bc5" (UID: "af1edb92-72cb-46db-9ad8-6e22d9206bc5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.297715 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-config-data" (OuterVolumeSpecName: "config-data") pod "af1edb92-72cb-46db-9ad8-6e22d9206bc5" (UID: "af1edb92-72cb-46db-9ad8-6e22d9206bc5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.323231 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.323259 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.323270 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af1edb92-72cb-46db-9ad8-6e22d9206bc5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.777749 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-fklbq" event={"ID":"af1edb92-72cb-46db-9ad8-6e22d9206bc5","Type":"ContainerDied","Data":"90fbceb4c5c04827ab46e94513d1239af97c666e65ebba081145b5759eedd128"} Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.777854 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90fbceb4c5c04827ab46e94513d1239af97c666e65ebba081145b5759eedd128" Oct 02 03:15:32 crc kubenswrapper[4775]: I1002 03:15:32.778035 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-fklbq" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.157681 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bd57d768c-b47z2"] Oct 02 03:15:33 crc kubenswrapper[4775]: E1002 03:15:33.158069 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af1edb92-72cb-46db-9ad8-6e22d9206bc5" containerName="cinder-db-sync" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.158080 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="af1edb92-72cb-46db-9ad8-6e22d9206bc5" containerName="cinder-db-sync" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.158269 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="af1edb92-72cb-46db-9ad8-6e22d9206bc5" containerName="cinder-db-sync" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.159184 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.182893 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bd57d768c-b47z2"] Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.239080 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-ovsdbserver-nb\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.239160 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-ovsdbserver-sb\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.239203 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-dns-svc\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.239229 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-config\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.239244 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b482g\" (UniqueName: \"kubernetes.io/projected/5cbdb932-01a7-409c-bc31-4dceb903ccad-kube-api-access-b482g\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.336260 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.337581 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.339549 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.339761 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.340308 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-ovsdbserver-nb\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.340399 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-ovsdbserver-sb\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.340420 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-dns-svc\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.340444 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-config\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.340459 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b482g\" (UniqueName: \"kubernetes.io/projected/5cbdb932-01a7-409c-bc31-4dceb903ccad-kube-api-access-b482g\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.341171 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-ovsdbserver-nb\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.341881 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-ovsdbserver-sb\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.344021 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.345628 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-dns-svc\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.345665 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-config\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.349853 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9zxst" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.357359 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.361040 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b482g\" (UniqueName: \"kubernetes.io/projected/5cbdb932-01a7-409c-bc31-4dceb903ccad-kube-api-access-b482g\") pod \"dnsmasq-dns-6bd57d768c-b47z2\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.442396 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-config-data\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.442468 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-etc-machine-id\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.442491 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.442710 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm26h\" (UniqueName: \"kubernetes.io/projected/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-kube-api-access-pm26h\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.442808 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-scripts\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.442858 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-logs\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.442921 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-config-data-custom\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.505655 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.549635 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-config-data\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.549916 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-etc-machine-id\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.549941 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.550039 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm26h\" (UniqueName: \"kubernetes.io/projected/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-kube-api-access-pm26h\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.550069 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-etc-machine-id\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.550902 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-scripts\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.550985 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-logs\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.551086 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-config-data-custom\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.552405 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-logs\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.556433 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-scripts\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.556730 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-config-data-custom\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.567668 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.568157 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-config-data\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.568199 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm26h\" (UniqueName: \"kubernetes.io/projected/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-kube-api-access-pm26h\") pod \"cinder-api-0\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.655708 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 03:15:33 crc kubenswrapper[4775]: I1002 03:15:33.973830 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bd57d768c-b47z2"] Oct 02 03:15:34 crc kubenswrapper[4775]: I1002 03:15:34.131175 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 03:15:34 crc kubenswrapper[4775]: W1002 03:15:34.145653 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53d272f6_0ff6_4f2f_b386_903a04d3f3ef.slice/crio-79e30db6022f42ebc4cb7f73179c6864b1ced8053bf25f3b7b5b63dd1e339ad3 WatchSource:0}: Error finding container 79e30db6022f42ebc4cb7f73179c6864b1ced8053bf25f3b7b5b63dd1e339ad3: Status 404 returned error can't find the container with id 79e30db6022f42ebc4cb7f73179c6864b1ced8053bf25f3b7b5b63dd1e339ad3 Oct 02 03:15:34 crc kubenswrapper[4775]: I1002 03:15:34.797791 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"53d272f6-0ff6-4f2f-b386-903a04d3f3ef","Type":"ContainerStarted","Data":"5f96cf4b3c64f46ee14b304a352fcdfcc326262df09b1c84cb7b5ead68dba9e8"} Oct 02 03:15:34 crc kubenswrapper[4775]: I1002 03:15:34.798047 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"53d272f6-0ff6-4f2f-b386-903a04d3f3ef","Type":"ContainerStarted","Data":"79e30db6022f42ebc4cb7f73179c6864b1ced8053bf25f3b7b5b63dd1e339ad3"} Oct 02 03:15:34 crc kubenswrapper[4775]: I1002 03:15:34.801680 4775 generic.go:334] "Generic (PLEG): container finished" podID="5cbdb932-01a7-409c-bc31-4dceb903ccad" containerID="7a6dd5f8679e03c842934fac8237531a771fe43ac712c7b34032696f3841d01a" exitCode=0 Oct 02 03:15:34 crc kubenswrapper[4775]: I1002 03:15:34.801725 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" event={"ID":"5cbdb932-01a7-409c-bc31-4dceb903ccad","Type":"ContainerDied","Data":"7a6dd5f8679e03c842934fac8237531a771fe43ac712c7b34032696f3841d01a"} Oct 02 03:15:34 crc kubenswrapper[4775]: I1002 03:15:34.801752 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" event={"ID":"5cbdb932-01a7-409c-bc31-4dceb903ccad","Type":"ContainerStarted","Data":"51729983878fe4d5fb3aeb93797a9a0e42a49ff8fce583d8be4667f55125ba91"} Oct 02 03:15:35 crc kubenswrapper[4775]: I1002 03:15:35.816381 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"53d272f6-0ff6-4f2f-b386-903a04d3f3ef","Type":"ContainerStarted","Data":"697f8eaf8a54462a09e6c76a1581e618c690316ac6e544ba1ab4552aa24cd90e"} Oct 02 03:15:35 crc kubenswrapper[4775]: I1002 03:15:35.817089 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 03:15:35 crc kubenswrapper[4775]: I1002 03:15:35.821367 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" event={"ID":"5cbdb932-01a7-409c-bc31-4dceb903ccad","Type":"ContainerStarted","Data":"0ca58f904e02e5c88e9e6959d4f51da3eb893f3e04ecbd6afcc108c76aa71c10"} Oct 02 03:15:35 crc kubenswrapper[4775]: I1002 03:15:35.821524 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:35 crc kubenswrapper[4775]: I1002 03:15:35.836370 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=2.836351694 podStartE2EDuration="2.836351694s" podCreationTimestamp="2025-10-02 03:15:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:15:35.834469015 +0000 UTC m=+5673.001213095" watchObservedRunningTime="2025-10-02 03:15:35.836351694 +0000 UTC m=+5673.003095744" Oct 02 03:15:35 crc kubenswrapper[4775]: I1002 03:15:35.866864 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" podStartSLOduration=2.866838034 podStartE2EDuration="2.866838034s" podCreationTimestamp="2025-10-02 03:15:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:15:35.859601276 +0000 UTC m=+5673.026345336" watchObservedRunningTime="2025-10-02 03:15:35.866838034 +0000 UTC m=+5673.033582094" Oct 02 03:15:37 crc kubenswrapper[4775]: I1002 03:15:37.233547 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:15:37 crc kubenswrapper[4775]: I1002 03:15:37.234031 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:15:37 crc kubenswrapper[4775]: I1002 03:15:37.234120 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 03:15:37 crc kubenswrapper[4775]: I1002 03:15:37.234927 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 03:15:37 crc kubenswrapper[4775]: I1002 03:15:37.235113 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" gracePeriod=600 Oct 02 03:15:37 crc kubenswrapper[4775]: E1002 03:15:37.352690 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:15:37 crc kubenswrapper[4775]: I1002 03:15:37.859932 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" exitCode=0 Oct 02 03:15:37 crc kubenswrapper[4775]: I1002 03:15:37.859993 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6"} Oct 02 03:15:37 crc kubenswrapper[4775]: I1002 03:15:37.860309 4775 scope.go:117] "RemoveContainer" containerID="5c0c35a295e1473ae177f82875e76ff0ae056f4db6af4a094e5ee23221bb6584" Oct 02 03:15:37 crc kubenswrapper[4775]: I1002 03:15:37.861660 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:15:37 crc kubenswrapper[4775]: E1002 03:15:37.862102 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:15:43 crc kubenswrapper[4775]: I1002 03:15:43.508301 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:15:43 crc kubenswrapper[4775]: I1002 03:15:43.596985 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-854565669c-g7wps"] Oct 02 03:15:43 crc kubenswrapper[4775]: I1002 03:15:43.597631 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-854565669c-g7wps" podUID="8eb457c8-28e9-4955-b9db-775c9e59634a" containerName="dnsmasq-dns" containerID="cri-o://a8e4faa64780b7e2ce07f300e8ea4f33e62a0b8ea95320e75a9acc09a85c0ae1" gracePeriod=10 Oct 02 03:15:43 crc kubenswrapper[4775]: I1002 03:15:43.970973 4775 generic.go:334] "Generic (PLEG): container finished" podID="8eb457c8-28e9-4955-b9db-775c9e59634a" containerID="a8e4faa64780b7e2ce07f300e8ea4f33e62a0b8ea95320e75a9acc09a85c0ae1" exitCode=0 Oct 02 03:15:43 crc kubenswrapper[4775]: I1002 03:15:43.971022 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-854565669c-g7wps" event={"ID":"8eb457c8-28e9-4955-b9db-775c9e59634a","Type":"ContainerDied","Data":"a8e4faa64780b7e2ce07f300e8ea4f33e62a0b8ea95320e75a9acc09a85c0ae1"} Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.054442 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.079298 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-dns-svc\") pod \"8eb457c8-28e9-4955-b9db-775c9e59634a\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.079394 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjzv8\" (UniqueName: \"kubernetes.io/projected/8eb457c8-28e9-4955-b9db-775c9e59634a-kube-api-access-rjzv8\") pod \"8eb457c8-28e9-4955-b9db-775c9e59634a\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.079480 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-ovsdbserver-sb\") pod \"8eb457c8-28e9-4955-b9db-775c9e59634a\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.079514 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-ovsdbserver-nb\") pod \"8eb457c8-28e9-4955-b9db-775c9e59634a\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.079547 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-config\") pod \"8eb457c8-28e9-4955-b9db-775c9e59634a\" (UID: \"8eb457c8-28e9-4955-b9db-775c9e59634a\") " Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.111846 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8eb457c8-28e9-4955-b9db-775c9e59634a-kube-api-access-rjzv8" (OuterVolumeSpecName: "kube-api-access-rjzv8") pod "8eb457c8-28e9-4955-b9db-775c9e59634a" (UID: "8eb457c8-28e9-4955-b9db-775c9e59634a"). InnerVolumeSpecName "kube-api-access-rjzv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.134360 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8eb457c8-28e9-4955-b9db-775c9e59634a" (UID: "8eb457c8-28e9-4955-b9db-775c9e59634a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.139436 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8eb457c8-28e9-4955-b9db-775c9e59634a" (UID: "8eb457c8-28e9-4955-b9db-775c9e59634a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.142586 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8eb457c8-28e9-4955-b9db-775c9e59634a" (UID: "8eb457c8-28e9-4955-b9db-775c9e59634a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.152798 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-config" (OuterVolumeSpecName: "config") pod "8eb457c8-28e9-4955-b9db-775c9e59634a" (UID: "8eb457c8-28e9-4955-b9db-775c9e59634a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.180732 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.180764 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.180776 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.180785 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8eb457c8-28e9-4955-b9db-775c9e59634a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.180795 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjzv8\" (UniqueName: \"kubernetes.io/projected/8eb457c8-28e9-4955-b9db-775c9e59634a-kube-api-access-rjzv8\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.981477 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-854565669c-g7wps" event={"ID":"8eb457c8-28e9-4955-b9db-775c9e59634a","Type":"ContainerDied","Data":"8a3deda0866e2841f2f14ec002e874dca613d104773324b5bc53c064b7e010f0"} Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.981803 4775 scope.go:117] "RemoveContainer" containerID="a8e4faa64780b7e2ce07f300e8ea4f33e62a0b8ea95320e75a9acc09a85c0ae1" Oct 02 03:15:44 crc kubenswrapper[4775]: I1002 03:15:44.981740 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-854565669c-g7wps" Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.016354 4775 scope.go:117] "RemoveContainer" containerID="f81a3646cdba9c0f9249e5d57d329a8ab236821f2bae4c01cbe49f41e0437f0b" Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.034801 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-854565669c-g7wps"] Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.042633 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-854565669c-g7wps"] Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.174559 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.174815 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6" containerName="nova-scheduler-scheduler" containerID="cri-o://1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935" gracePeriod=30 Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.186716 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.186935 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" containerName="nova-metadata-log" containerID="cri-o://64ceeabff7ae4e873af57d8fd56c5a730409ce674f71ff304b80f43af51430d4" gracePeriod=30 Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.187058 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" containerName="nova-metadata-metadata" containerID="cri-o://fe7fd197864de5e576d1dd56a195c436013c7f5ce89446da93b812dc73917bb4" gracePeriod=30 Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.200863 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.201120 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="5fc4855b-083c-44be-bfa6-af98a1c963fa" containerName="nova-cell0-conductor-conductor" containerID="cri-o://72b4aa11ce93c390e5656b40ad1314a9b3d9d8fe52071d2d3ac980a62dd975d9" gracePeriod=30 Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.219832 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.220107 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="690c75e7-b17b-4aba-b6a6-2d969ad8137d" containerName="nova-api-log" containerID="cri-o://81ce762c940c90ea0299fd779112ade0eab09c4e18b02b0bd4b3270faeb3da57" gracePeriod=30 Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.220162 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="690c75e7-b17b-4aba-b6a6-2d969ad8137d" containerName="nova-api-api" containerID="cri-o://2aa7bec967ea4a34d7201995bd474c11c9a3aa1f616f8053076acda37a854993" gracePeriod=30 Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.231097 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.231327 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="de4c6de6-bd34-4316-9791-b360aaf48066" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://253c2b542479ae7d61337ab667bbb0c152706bf9aa25f6a81c70f8682aa874d9" gracePeriod=30 Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.332165 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.332363 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="ca3f9270-7d6a-4b91-bf50-944388836500" containerName="nova-cell1-conductor-conductor" containerID="cri-o://617d1d307ef28adc7ecba70884e00a37f469c79f8a45e6925777dabfade04be1" gracePeriod=30 Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.600081 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.775774 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8eb457c8-28e9-4955-b9db-775c9e59634a" path="/var/lib/kubelet/pods/8eb457c8-28e9-4955-b9db-775c9e59634a/volumes" Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.995186 4775 generic.go:334] "Generic (PLEG): container finished" podID="690c75e7-b17b-4aba-b6a6-2d969ad8137d" containerID="81ce762c940c90ea0299fd779112ade0eab09c4e18b02b0bd4b3270faeb3da57" exitCode=143 Oct 02 03:15:45 crc kubenswrapper[4775]: I1002 03:15:45.995249 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"690c75e7-b17b-4aba-b6a6-2d969ad8137d","Type":"ContainerDied","Data":"81ce762c940c90ea0299fd779112ade0eab09c4e18b02b0bd4b3270faeb3da57"} Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.002467 4775 generic.go:334] "Generic (PLEG): container finished" podID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" containerID="64ceeabff7ae4e873af57d8fd56c5a730409ce674f71ff304b80f43af51430d4" exitCode=143 Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.002551 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a4a80c5f-4c44-46f7-a694-a6904a158bfc","Type":"ContainerDied","Data":"64ceeabff7ae4e873af57d8fd56c5a730409ce674f71ff304b80f43af51430d4"} Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.005632 4775 generic.go:334] "Generic (PLEG): container finished" podID="de4c6de6-bd34-4316-9791-b360aaf48066" containerID="253c2b542479ae7d61337ab667bbb0c152706bf9aa25f6a81c70f8682aa874d9" exitCode=0 Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.005721 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"de4c6de6-bd34-4316-9791-b360aaf48066","Type":"ContainerDied","Data":"253c2b542479ae7d61337ab667bbb0c152706bf9aa25f6a81c70f8682aa874d9"} Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.005746 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"de4c6de6-bd34-4316-9791-b360aaf48066","Type":"ContainerDied","Data":"7f01083fed77a27cf203b805a21f4129fe977f8d068c64fc929032b2972e4970"} Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.005759 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f01083fed77a27cf203b805a21f4129fe977f8d068c64fc929032b2972e4970" Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.037867 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:46 crc kubenswrapper[4775]: E1002 03:15:46.125989 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="617d1d307ef28adc7ecba70884e00a37f469c79f8a45e6925777dabfade04be1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 03:15:46 crc kubenswrapper[4775]: E1002 03:15:46.127086 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="617d1d307ef28adc7ecba70884e00a37f469c79f8a45e6925777dabfade04be1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 03:15:46 crc kubenswrapper[4775]: E1002 03:15:46.128862 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="617d1d307ef28adc7ecba70884e00a37f469c79f8a45e6925777dabfade04be1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 03:15:46 crc kubenswrapper[4775]: E1002 03:15:46.128894 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="ca3f9270-7d6a-4b91-bf50-944388836500" containerName="nova-cell1-conductor-conductor" Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.213400 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4c6de6-bd34-4316-9791-b360aaf48066-combined-ca-bundle\") pod \"de4c6de6-bd34-4316-9791-b360aaf48066\" (UID: \"de4c6de6-bd34-4316-9791-b360aaf48066\") " Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.213500 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de4c6de6-bd34-4316-9791-b360aaf48066-config-data\") pod \"de4c6de6-bd34-4316-9791-b360aaf48066\" (UID: \"de4c6de6-bd34-4316-9791-b360aaf48066\") " Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.213543 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl95h\" (UniqueName: \"kubernetes.io/projected/de4c6de6-bd34-4316-9791-b360aaf48066-kube-api-access-tl95h\") pod \"de4c6de6-bd34-4316-9791-b360aaf48066\" (UID: \"de4c6de6-bd34-4316-9791-b360aaf48066\") " Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.219598 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de4c6de6-bd34-4316-9791-b360aaf48066-kube-api-access-tl95h" (OuterVolumeSpecName: "kube-api-access-tl95h") pod "de4c6de6-bd34-4316-9791-b360aaf48066" (UID: "de4c6de6-bd34-4316-9791-b360aaf48066"). InnerVolumeSpecName "kube-api-access-tl95h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.240831 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de4c6de6-bd34-4316-9791-b360aaf48066-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de4c6de6-bd34-4316-9791-b360aaf48066" (UID: "de4c6de6-bd34-4316-9791-b360aaf48066"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.243131 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de4c6de6-bd34-4316-9791-b360aaf48066-config-data" (OuterVolumeSpecName: "config-data") pod "de4c6de6-bd34-4316-9791-b360aaf48066" (UID: "de4c6de6-bd34-4316-9791-b360aaf48066"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.314918 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de4c6de6-bd34-4316-9791-b360aaf48066-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.314947 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl95h\" (UniqueName: \"kubernetes.io/projected/de4c6de6-bd34-4316-9791-b360aaf48066-kube-api-access-tl95h\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.314971 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4c6de6-bd34-4316-9791-b360aaf48066-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:46 crc kubenswrapper[4775]: E1002 03:15:46.350153 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 03:15:46 crc kubenswrapper[4775]: E1002 03:15:46.351627 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 03:15:46 crc kubenswrapper[4775]: E1002 03:15:46.352708 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 03:15:46 crc kubenswrapper[4775]: E1002 03:15:46.352770 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6" containerName="nova-scheduler-scheduler" Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.790456 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.930321 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc4855b-083c-44be-bfa6-af98a1c963fa-combined-ca-bundle\") pod \"5fc4855b-083c-44be-bfa6-af98a1c963fa\" (UID: \"5fc4855b-083c-44be-bfa6-af98a1c963fa\") " Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.930378 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc4855b-083c-44be-bfa6-af98a1c963fa-config-data\") pod \"5fc4855b-083c-44be-bfa6-af98a1c963fa\" (UID: \"5fc4855b-083c-44be-bfa6-af98a1c963fa\") " Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.930419 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-btth6\" (UniqueName: \"kubernetes.io/projected/5fc4855b-083c-44be-bfa6-af98a1c963fa-kube-api-access-btth6\") pod \"5fc4855b-083c-44be-bfa6-af98a1c963fa\" (UID: \"5fc4855b-083c-44be-bfa6-af98a1c963fa\") " Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.934778 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fc4855b-083c-44be-bfa6-af98a1c963fa-kube-api-access-btth6" (OuterVolumeSpecName: "kube-api-access-btth6") pod "5fc4855b-083c-44be-bfa6-af98a1c963fa" (UID: "5fc4855b-083c-44be-bfa6-af98a1c963fa"). InnerVolumeSpecName "kube-api-access-btth6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.963868 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc4855b-083c-44be-bfa6-af98a1c963fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fc4855b-083c-44be-bfa6-af98a1c963fa" (UID: "5fc4855b-083c-44be-bfa6-af98a1c963fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:46 crc kubenswrapper[4775]: I1002 03:15:46.976180 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fc4855b-083c-44be-bfa6-af98a1c963fa-config-data" (OuterVolumeSpecName: "config-data") pod "5fc4855b-083c-44be-bfa6-af98a1c963fa" (UID: "5fc4855b-083c-44be-bfa6-af98a1c963fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.022980 4775 generic.go:334] "Generic (PLEG): container finished" podID="5fc4855b-083c-44be-bfa6-af98a1c963fa" containerID="72b4aa11ce93c390e5656b40ad1314a9b3d9d8fe52071d2d3ac980a62dd975d9" exitCode=0 Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.023071 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.023077 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5fc4855b-083c-44be-bfa6-af98a1c963fa","Type":"ContainerDied","Data":"72b4aa11ce93c390e5656b40ad1314a9b3d9d8fe52071d2d3ac980a62dd975d9"} Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.023151 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5fc4855b-083c-44be-bfa6-af98a1c963fa","Type":"ContainerDied","Data":"08138d89b7df92c281d77c220e70ed8bd0664210a29da1a91772d46308beb8a3"} Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.023173 4775 scope.go:117] "RemoveContainer" containerID="72b4aa11ce93c390e5656b40ad1314a9b3d9d8fe52071d2d3ac980a62dd975d9" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.024138 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.032072 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fc4855b-083c-44be-bfa6-af98a1c963fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.032101 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fc4855b-083c-44be-bfa6-af98a1c963fa-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.032111 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-btth6\" (UniqueName: \"kubernetes.io/projected/5fc4855b-083c-44be-bfa6-af98a1c963fa-kube-api-access-btth6\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.050451 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.061683 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.067246 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.088094 4775 scope.go:117] "RemoveContainer" containerID="72b4aa11ce93c390e5656b40ad1314a9b3d9d8fe52071d2d3ac980a62dd975d9" Oct 02 03:15:47 crc kubenswrapper[4775]: E1002 03:15:47.088658 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72b4aa11ce93c390e5656b40ad1314a9b3d9d8fe52071d2d3ac980a62dd975d9\": container with ID starting with 72b4aa11ce93c390e5656b40ad1314a9b3d9d8fe52071d2d3ac980a62dd975d9 not found: ID does not exist" containerID="72b4aa11ce93c390e5656b40ad1314a9b3d9d8fe52071d2d3ac980a62dd975d9" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.088695 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72b4aa11ce93c390e5656b40ad1314a9b3d9d8fe52071d2d3ac980a62dd975d9"} err="failed to get container status \"72b4aa11ce93c390e5656b40ad1314a9b3d9d8fe52071d2d3ac980a62dd975d9\": rpc error: code = NotFound desc = could not find container \"72b4aa11ce93c390e5656b40ad1314a9b3d9d8fe52071d2d3ac980a62dd975d9\": container with ID starting with 72b4aa11ce93c390e5656b40ad1314a9b3d9d8fe52071d2d3ac980a62dd975d9 not found: ID does not exist" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.114975 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 03:15:47 crc kubenswrapper[4775]: E1002 03:15:47.115693 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fc4855b-083c-44be-bfa6-af98a1c963fa" containerName="nova-cell0-conductor-conductor" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.115729 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fc4855b-083c-44be-bfa6-af98a1c963fa" containerName="nova-cell0-conductor-conductor" Oct 02 03:15:47 crc kubenswrapper[4775]: E1002 03:15:47.115748 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eb457c8-28e9-4955-b9db-775c9e59634a" containerName="dnsmasq-dns" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.115754 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eb457c8-28e9-4955-b9db-775c9e59634a" containerName="dnsmasq-dns" Oct 02 03:15:47 crc kubenswrapper[4775]: E1002 03:15:47.115777 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de4c6de6-bd34-4316-9791-b360aaf48066" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.115784 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="de4c6de6-bd34-4316-9791-b360aaf48066" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 03:15:47 crc kubenswrapper[4775]: E1002 03:15:47.115794 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8eb457c8-28e9-4955-b9db-775c9e59634a" containerName="init" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.115800 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8eb457c8-28e9-4955-b9db-775c9e59634a" containerName="init" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.116001 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fc4855b-083c-44be-bfa6-af98a1c963fa" containerName="nova-cell0-conductor-conductor" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.116023 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="de4c6de6-bd34-4316-9791-b360aaf48066" containerName="nova-cell1-novncproxy-novncproxy" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.116042 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8eb457c8-28e9-4955-b9db-775c9e59634a" containerName="dnsmasq-dns" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.116975 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.119513 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.125096 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.134752 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.140760 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.142413 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.146286 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.146986 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.236225 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/207f84a9-5284-46d4-856c-ab11901ae6e7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"207f84a9-5284-46d4-856c-ab11901ae6e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.236361 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5hcc\" (UniqueName: \"kubernetes.io/projected/207f84a9-5284-46d4-856c-ab11901ae6e7-kube-api-access-m5hcc\") pod \"nova-cell1-novncproxy-0\" (UID: \"207f84a9-5284-46d4-856c-ab11901ae6e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.236503 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207f84a9-5284-46d4-856c-ab11901ae6e7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"207f84a9-5284-46d4-856c-ab11901ae6e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.339372 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e865292-9dda-49fc-acdf-3afc04304aac-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4e865292-9dda-49fc-acdf-3afc04304aac\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.339481 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e865292-9dda-49fc-acdf-3afc04304aac-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4e865292-9dda-49fc-acdf-3afc04304aac\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.339555 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5hcc\" (UniqueName: \"kubernetes.io/projected/207f84a9-5284-46d4-856c-ab11901ae6e7-kube-api-access-m5hcc\") pod \"nova-cell1-novncproxy-0\" (UID: \"207f84a9-5284-46d4-856c-ab11901ae6e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.339656 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207f84a9-5284-46d4-856c-ab11901ae6e7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"207f84a9-5284-46d4-856c-ab11901ae6e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.339796 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/207f84a9-5284-46d4-856c-ab11901ae6e7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"207f84a9-5284-46d4-856c-ab11901ae6e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.339830 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn6wm\" (UniqueName: \"kubernetes.io/projected/4e865292-9dda-49fc-acdf-3afc04304aac-kube-api-access-mn6wm\") pod \"nova-cell0-conductor-0\" (UID: \"4e865292-9dda-49fc-acdf-3afc04304aac\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.346701 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/207f84a9-5284-46d4-856c-ab11901ae6e7-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"207f84a9-5284-46d4-856c-ab11901ae6e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.348698 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/207f84a9-5284-46d4-856c-ab11901ae6e7-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"207f84a9-5284-46d4-856c-ab11901ae6e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.356515 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5hcc\" (UniqueName: \"kubernetes.io/projected/207f84a9-5284-46d4-856c-ab11901ae6e7-kube-api-access-m5hcc\") pod \"nova-cell1-novncproxy-0\" (UID: \"207f84a9-5284-46d4-856c-ab11901ae6e7\") " pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.432625 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.440865 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn6wm\" (UniqueName: \"kubernetes.io/projected/4e865292-9dda-49fc-acdf-3afc04304aac-kube-api-access-mn6wm\") pod \"nova-cell0-conductor-0\" (UID: \"4e865292-9dda-49fc-acdf-3afc04304aac\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.440924 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e865292-9dda-49fc-acdf-3afc04304aac-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4e865292-9dda-49fc-acdf-3afc04304aac\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.440974 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e865292-9dda-49fc-acdf-3afc04304aac-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4e865292-9dda-49fc-acdf-3afc04304aac\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.445129 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e865292-9dda-49fc-acdf-3afc04304aac-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"4e865292-9dda-49fc-acdf-3afc04304aac\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.446468 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e865292-9dda-49fc-acdf-3afc04304aac-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"4e865292-9dda-49fc-acdf-3afc04304aac\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.458597 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn6wm\" (UniqueName: \"kubernetes.io/projected/4e865292-9dda-49fc-acdf-3afc04304aac-kube-api-access-mn6wm\") pod \"nova-cell0-conductor-0\" (UID: \"4e865292-9dda-49fc-acdf-3afc04304aac\") " pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.756348 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.780679 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fc4855b-083c-44be-bfa6-af98a1c963fa" path="/var/lib/kubelet/pods/5fc4855b-083c-44be-bfa6-af98a1c963fa/volumes" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.781730 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de4c6de6-bd34-4316-9791-b360aaf48066" path="/var/lib/kubelet/pods/de4c6de6-bd34-4316-9791-b360aaf48066/volumes" Oct 02 03:15:47 crc kubenswrapper[4775]: I1002 03:15:47.960720 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.040342 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"207f84a9-5284-46d4-856c-ab11901ae6e7","Type":"ContainerStarted","Data":"596756daa9df075dd49ea6525fe70ca97b091bf6a07a298cc388dea965cddd2a"} Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.336400 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 03:15:48 crc kubenswrapper[4775]: W1002 03:15:48.339253 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e865292_9dda_49fc_acdf_3afc04304aac.slice/crio-c89a4734eb7e14a5d749a8ae821c124e6672db8a27a6f24bd494d1b09799be33 WatchSource:0}: Error finding container c89a4734eb7e14a5d749a8ae821c124e6672db8a27a6f24bd494d1b09799be33: Status 404 returned error can't find the container with id c89a4734eb7e14a5d749a8ae821c124e6672db8a27a6f24bd494d1b09799be33 Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.339771 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.73:8775/\": read tcp 10.217.0.2:60680->10.217.1.73:8775: read: connection reset by peer" Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.339798 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.73:8775/\": read tcp 10.217.0.2:60682->10.217.1.73:8775: read: connection reset by peer" Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.772131 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.840200 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.876689 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/690c75e7-b17b-4aba-b6a6-2d969ad8137d-logs\") pod \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.876896 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690c75e7-b17b-4aba-b6a6-2d969ad8137d-config-data\") pod \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.876921 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690c75e7-b17b-4aba-b6a6-2d969ad8137d-combined-ca-bundle\") pod \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.876976 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7r4kb\" (UniqueName: \"kubernetes.io/projected/690c75e7-b17b-4aba-b6a6-2d969ad8137d-kube-api-access-7r4kb\") pod \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\" (UID: \"690c75e7-b17b-4aba-b6a6-2d969ad8137d\") " Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.878390 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/690c75e7-b17b-4aba-b6a6-2d969ad8137d-logs" (OuterVolumeSpecName: "logs") pod "690c75e7-b17b-4aba-b6a6-2d969ad8137d" (UID: "690c75e7-b17b-4aba-b6a6-2d969ad8137d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.888590 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/690c75e7-b17b-4aba-b6a6-2d969ad8137d-kube-api-access-7r4kb" (OuterVolumeSpecName: "kube-api-access-7r4kb") pod "690c75e7-b17b-4aba-b6a6-2d969ad8137d" (UID: "690c75e7-b17b-4aba-b6a6-2d969ad8137d"). InnerVolumeSpecName "kube-api-access-7r4kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.918071 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/690c75e7-b17b-4aba-b6a6-2d969ad8137d-config-data" (OuterVolumeSpecName: "config-data") pod "690c75e7-b17b-4aba-b6a6-2d969ad8137d" (UID: "690c75e7-b17b-4aba-b6a6-2d969ad8137d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.934084 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/690c75e7-b17b-4aba-b6a6-2d969ad8137d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "690c75e7-b17b-4aba-b6a6-2d969ad8137d" (UID: "690c75e7-b17b-4aba-b6a6-2d969ad8137d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.977995 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4a80c5f-4c44-46f7-a694-a6904a158bfc-config-data\") pod \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.978103 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4a80c5f-4c44-46f7-a694-a6904a158bfc-logs\") pod \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.978174 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4a80c5f-4c44-46f7-a694-a6904a158bfc-combined-ca-bundle\") pod \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.978273 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2t8dg\" (UniqueName: \"kubernetes.io/projected/a4a80c5f-4c44-46f7-a694-a6904a158bfc-kube-api-access-2t8dg\") pod \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\" (UID: \"a4a80c5f-4c44-46f7-a694-a6904a158bfc\") " Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.978550 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4a80c5f-4c44-46f7-a694-a6904a158bfc-logs" (OuterVolumeSpecName: "logs") pod "a4a80c5f-4c44-46f7-a694-a6904a158bfc" (UID: "a4a80c5f-4c44-46f7-a694-a6904a158bfc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.978632 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/690c75e7-b17b-4aba-b6a6-2d969ad8137d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.978649 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/690c75e7-b17b-4aba-b6a6-2d969ad8137d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.978668 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7r4kb\" (UniqueName: \"kubernetes.io/projected/690c75e7-b17b-4aba-b6a6-2d969ad8137d-kube-api-access-7r4kb\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.978678 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/690c75e7-b17b-4aba-b6a6-2d969ad8137d-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.978685 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a4a80c5f-4c44-46f7-a694-a6904a158bfc-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:48 crc kubenswrapper[4775]: I1002 03:15:48.988792 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4a80c5f-4c44-46f7-a694-a6904a158bfc-kube-api-access-2t8dg" (OuterVolumeSpecName: "kube-api-access-2t8dg") pod "a4a80c5f-4c44-46f7-a694-a6904a158bfc" (UID: "a4a80c5f-4c44-46f7-a694-a6904a158bfc"). InnerVolumeSpecName "kube-api-access-2t8dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.008452 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4a80c5f-4c44-46f7-a694-a6904a158bfc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4a80c5f-4c44-46f7-a694-a6904a158bfc" (UID: "a4a80c5f-4c44-46f7-a694-a6904a158bfc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.059038 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4a80c5f-4c44-46f7-a694-a6904a158bfc-config-data" (OuterVolumeSpecName: "config-data") pod "a4a80c5f-4c44-46f7-a694-a6904a158bfc" (UID: "a4a80c5f-4c44-46f7-a694-a6904a158bfc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.061796 4775 generic.go:334] "Generic (PLEG): container finished" podID="690c75e7-b17b-4aba-b6a6-2d969ad8137d" containerID="2aa7bec967ea4a34d7201995bd474c11c9a3aa1f616f8053076acda37a854993" exitCode=0 Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.061842 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.061864 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"690c75e7-b17b-4aba-b6a6-2d969ad8137d","Type":"ContainerDied","Data":"2aa7bec967ea4a34d7201995bd474c11c9a3aa1f616f8053076acda37a854993"} Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.062372 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"690c75e7-b17b-4aba-b6a6-2d969ad8137d","Type":"ContainerDied","Data":"8fec26988832127ac7a3349ec5ae25c6760c24b9af4088b4735862c9ae586374"} Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.062391 4775 scope.go:117] "RemoveContainer" containerID="2aa7bec967ea4a34d7201995bd474c11c9a3aa1f616f8053076acda37a854993" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.070133 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4e865292-9dda-49fc-acdf-3afc04304aac","Type":"ContainerStarted","Data":"6e4b2842c6945dc728e9ab20f6eae59f1424306f45f162bb0205a42c38e05645"} Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.070179 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4e865292-9dda-49fc-acdf-3afc04304aac","Type":"ContainerStarted","Data":"c89a4734eb7e14a5d749a8ae821c124e6672db8a27a6f24bd494d1b09799be33"} Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.070665 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.080449 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2t8dg\" (UniqueName: \"kubernetes.io/projected/a4a80c5f-4c44-46f7-a694-a6904a158bfc-kube-api-access-2t8dg\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.080479 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4a80c5f-4c44-46f7-a694-a6904a158bfc-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.080490 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4a80c5f-4c44-46f7-a694-a6904a158bfc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.081978 4775 generic.go:334] "Generic (PLEG): container finished" podID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" containerID="fe7fd197864de5e576d1dd56a195c436013c7f5ce89446da93b812dc73917bb4" exitCode=0 Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.082092 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a4a80c5f-4c44-46f7-a694-a6904a158bfc","Type":"ContainerDied","Data":"fe7fd197864de5e576d1dd56a195c436013c7f5ce89446da93b812dc73917bb4"} Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.082112 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.082124 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a4a80c5f-4c44-46f7-a694-a6904a158bfc","Type":"ContainerDied","Data":"28ebddd32a3f9f450c32bb5f9789b43d6a18ffea587107b0ea167db9dd649793"} Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.092445 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"207f84a9-5284-46d4-856c-ab11901ae6e7","Type":"ContainerStarted","Data":"c3c351e69b7dfd3b1fac06f6a5732846898cc2633e0ba8bf6cb6989a8b5ee22a"} Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.103060 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.103042084 podStartE2EDuration="2.103042084s" podCreationTimestamp="2025-10-02 03:15:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:15:49.089529804 +0000 UTC m=+5686.256273844" watchObservedRunningTime="2025-10-02 03:15:49.103042084 +0000 UTC m=+5686.269786124" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.117834 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.117815937 podStartE2EDuration="2.117815937s" podCreationTimestamp="2025-10-02 03:15:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:15:49.107466549 +0000 UTC m=+5686.274210589" watchObservedRunningTime="2025-10-02 03:15:49.117815937 +0000 UTC m=+5686.284559977" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.126137 4775 scope.go:117] "RemoveContainer" containerID="81ce762c940c90ea0299fd779112ade0eab09c4e18b02b0bd4b3270faeb3da57" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.151017 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.169569 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.175519 4775 scope.go:117] "RemoveContainer" containerID="2aa7bec967ea4a34d7201995bd474c11c9a3aa1f616f8053076acda37a854993" Oct 02 03:15:49 crc kubenswrapper[4775]: E1002 03:15:49.176844 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2aa7bec967ea4a34d7201995bd474c11c9a3aa1f616f8053076acda37a854993\": container with ID starting with 2aa7bec967ea4a34d7201995bd474c11c9a3aa1f616f8053076acda37a854993 not found: ID does not exist" containerID="2aa7bec967ea4a34d7201995bd474c11c9a3aa1f616f8053076acda37a854993" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.176891 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2aa7bec967ea4a34d7201995bd474c11c9a3aa1f616f8053076acda37a854993"} err="failed to get container status \"2aa7bec967ea4a34d7201995bd474c11c9a3aa1f616f8053076acda37a854993\": rpc error: code = NotFound desc = could not find container \"2aa7bec967ea4a34d7201995bd474c11c9a3aa1f616f8053076acda37a854993\": container with ID starting with 2aa7bec967ea4a34d7201995bd474c11c9a3aa1f616f8053076acda37a854993 not found: ID does not exist" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.176918 4775 scope.go:117] "RemoveContainer" containerID="81ce762c940c90ea0299fd779112ade0eab09c4e18b02b0bd4b3270faeb3da57" Oct 02 03:15:49 crc kubenswrapper[4775]: E1002 03:15:49.178102 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81ce762c940c90ea0299fd779112ade0eab09c4e18b02b0bd4b3270faeb3da57\": container with ID starting with 81ce762c940c90ea0299fd779112ade0eab09c4e18b02b0bd4b3270faeb3da57 not found: ID does not exist" containerID="81ce762c940c90ea0299fd779112ade0eab09c4e18b02b0bd4b3270faeb3da57" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.178131 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81ce762c940c90ea0299fd779112ade0eab09c4e18b02b0bd4b3270faeb3da57"} err="failed to get container status \"81ce762c940c90ea0299fd779112ade0eab09c4e18b02b0bd4b3270faeb3da57\": rpc error: code = NotFound desc = could not find container \"81ce762c940c90ea0299fd779112ade0eab09c4e18b02b0bd4b3270faeb3da57\": container with ID starting with 81ce762c940c90ea0299fd779112ade0eab09c4e18b02b0bd4b3270faeb3da57 not found: ID does not exist" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.178151 4775 scope.go:117] "RemoveContainer" containerID="fe7fd197864de5e576d1dd56a195c436013c7f5ce89446da93b812dc73917bb4" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.189867 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 03:15:49 crc kubenswrapper[4775]: E1002 03:15:49.190311 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690c75e7-b17b-4aba-b6a6-2d969ad8137d" containerName="nova-api-api" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.190328 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="690c75e7-b17b-4aba-b6a6-2d969ad8137d" containerName="nova-api-api" Oct 02 03:15:49 crc kubenswrapper[4775]: E1002 03:15:49.190343 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" containerName="nova-metadata-metadata" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.190350 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" containerName="nova-metadata-metadata" Oct 02 03:15:49 crc kubenswrapper[4775]: E1002 03:15:49.190363 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="690c75e7-b17b-4aba-b6a6-2d969ad8137d" containerName="nova-api-log" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.190369 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="690c75e7-b17b-4aba-b6a6-2d969ad8137d" containerName="nova-api-log" Oct 02 03:15:49 crc kubenswrapper[4775]: E1002 03:15:49.190378 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" containerName="nova-metadata-log" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.190385 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" containerName="nova-metadata-log" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.190542 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" containerName="nova-metadata-metadata" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.190554 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" containerName="nova-metadata-log" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.190570 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="690c75e7-b17b-4aba-b6a6-2d969ad8137d" containerName="nova-api-api" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.190582 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="690c75e7-b17b-4aba-b6a6-2d969ad8137d" containerName="nova-api-log" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.191509 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.195416 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.206261 4775 scope.go:117] "RemoveContainer" containerID="64ceeabff7ae4e873af57d8fd56c5a730409ce674f71ff304b80f43af51430d4" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.212314 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.231002 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.233182 4775 scope.go:117] "RemoveContainer" containerID="fe7fd197864de5e576d1dd56a195c436013c7f5ce89446da93b812dc73917bb4" Oct 02 03:15:49 crc kubenswrapper[4775]: E1002 03:15:49.234205 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe7fd197864de5e576d1dd56a195c436013c7f5ce89446da93b812dc73917bb4\": container with ID starting with fe7fd197864de5e576d1dd56a195c436013c7f5ce89446da93b812dc73917bb4 not found: ID does not exist" containerID="fe7fd197864de5e576d1dd56a195c436013c7f5ce89446da93b812dc73917bb4" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.234248 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe7fd197864de5e576d1dd56a195c436013c7f5ce89446da93b812dc73917bb4"} err="failed to get container status \"fe7fd197864de5e576d1dd56a195c436013c7f5ce89446da93b812dc73917bb4\": rpc error: code = NotFound desc = could not find container \"fe7fd197864de5e576d1dd56a195c436013c7f5ce89446da93b812dc73917bb4\": container with ID starting with fe7fd197864de5e576d1dd56a195c436013c7f5ce89446da93b812dc73917bb4 not found: ID does not exist" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.234306 4775 scope.go:117] "RemoveContainer" containerID="64ceeabff7ae4e873af57d8fd56c5a730409ce674f71ff304b80f43af51430d4" Oct 02 03:15:49 crc kubenswrapper[4775]: E1002 03:15:49.235028 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64ceeabff7ae4e873af57d8fd56c5a730409ce674f71ff304b80f43af51430d4\": container with ID starting with 64ceeabff7ae4e873af57d8fd56c5a730409ce674f71ff304b80f43af51430d4 not found: ID does not exist" containerID="64ceeabff7ae4e873af57d8fd56c5a730409ce674f71ff304b80f43af51430d4" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.235066 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64ceeabff7ae4e873af57d8fd56c5a730409ce674f71ff304b80f43af51430d4"} err="failed to get container status \"64ceeabff7ae4e873af57d8fd56c5a730409ce674f71ff304b80f43af51430d4\": rpc error: code = NotFound desc = could not find container \"64ceeabff7ae4e873af57d8fd56c5a730409ce674f71ff304b80f43af51430d4\": container with ID starting with 64ceeabff7ae4e873af57d8fd56c5a730409ce674f71ff304b80f43af51430d4 not found: ID does not exist" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.243770 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.251366 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.252847 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.254208 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.257880 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.287127 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jthvb\" (UniqueName: \"kubernetes.io/projected/8986cac6-b070-4ea8-a961-24481fbb7bdd-kube-api-access-jthvb\") pod \"nova-api-0\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.287187 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8986cac6-b070-4ea8-a961-24481fbb7bdd-config-data\") pod \"nova-api-0\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.287340 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8986cac6-b070-4ea8-a961-24481fbb7bdd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.287583 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8986cac6-b070-4ea8-a961-24481fbb7bdd-logs\") pod \"nova-api-0\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.389798 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jthvb\" (UniqueName: \"kubernetes.io/projected/8986cac6-b070-4ea8-a961-24481fbb7bdd-kube-api-access-jthvb\") pod \"nova-api-0\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.389854 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8986cac6-b070-4ea8-a961-24481fbb7bdd-config-data\") pod \"nova-api-0\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.389879 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/def5a435-381a-4636-932e-8e4881989de3-logs\") pod \"nova-metadata-0\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.389897 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8986cac6-b070-4ea8-a961-24481fbb7bdd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.389925 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/def5a435-381a-4636-932e-8e4881989de3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.389968 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/def5a435-381a-4636-932e-8e4881989de3-config-data\") pod \"nova-metadata-0\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.390001 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljlqn\" (UniqueName: \"kubernetes.io/projected/def5a435-381a-4636-932e-8e4881989de3-kube-api-access-ljlqn\") pod \"nova-metadata-0\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.390049 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8986cac6-b070-4ea8-a961-24481fbb7bdd-logs\") pod \"nova-api-0\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.390384 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8986cac6-b070-4ea8-a961-24481fbb7bdd-logs\") pod \"nova-api-0\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.397027 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8986cac6-b070-4ea8-a961-24481fbb7bdd-config-data\") pod \"nova-api-0\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.398600 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8986cac6-b070-4ea8-a961-24481fbb7bdd-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.411306 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jthvb\" (UniqueName: \"kubernetes.io/projected/8986cac6-b070-4ea8-a961-24481fbb7bdd-kube-api-access-jthvb\") pod \"nova-api-0\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.491326 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/def5a435-381a-4636-932e-8e4881989de3-logs\") pod \"nova-metadata-0\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.491380 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/def5a435-381a-4636-932e-8e4881989de3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.491408 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/def5a435-381a-4636-932e-8e4881989de3-config-data\") pod \"nova-metadata-0\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.491444 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljlqn\" (UniqueName: \"kubernetes.io/projected/def5a435-381a-4636-932e-8e4881989de3-kube-api-access-ljlqn\") pod \"nova-metadata-0\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.493319 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/def5a435-381a-4636-932e-8e4881989de3-logs\") pod \"nova-metadata-0\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.494721 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/def5a435-381a-4636-932e-8e4881989de3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.496154 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/def5a435-381a-4636-932e-8e4881989de3-config-data\") pod \"nova-metadata-0\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.513932 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.526507 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljlqn\" (UniqueName: \"kubernetes.io/projected/def5a435-381a-4636-932e-8e4881989de3-kube-api-access-ljlqn\") pod \"nova-metadata-0\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.569204 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.781432 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="690c75e7-b17b-4aba-b6a6-2d969ad8137d" path="/var/lib/kubelet/pods/690c75e7-b17b-4aba-b6a6-2d969ad8137d/volumes" Oct 02 03:15:49 crc kubenswrapper[4775]: I1002 03:15:49.782331 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4a80c5f-4c44-46f7-a694-a6904a158bfc" path="/var/lib/kubelet/pods/a4a80c5f-4c44-46f7-a694-a6904a158bfc/volumes" Oct 02 03:15:50 crc kubenswrapper[4775]: I1002 03:15:50.054764 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 03:15:50 crc kubenswrapper[4775]: I1002 03:15:50.099661 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 03:15:50 crc kubenswrapper[4775]: I1002 03:15:50.116271 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8986cac6-b070-4ea8-a961-24481fbb7bdd","Type":"ContainerStarted","Data":"49734b270b32bcae1851d56616d6f946eaf31914db2772c5124d03d9e0cd5089"} Oct 02 03:15:50 crc kubenswrapper[4775]: I1002 03:15:50.684017 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 03:15:50 crc kubenswrapper[4775]: I1002 03:15:50.765872 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:15:50 crc kubenswrapper[4775]: E1002 03:15:50.766202 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:15:50 crc kubenswrapper[4775]: I1002 03:15:50.819496 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsv4d\" (UniqueName: \"kubernetes.io/projected/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-kube-api-access-qsv4d\") pod \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\" (UID: \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\") " Oct 02 03:15:50 crc kubenswrapper[4775]: I1002 03:15:50.819561 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-config-data\") pod \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\" (UID: \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\") " Oct 02 03:15:50 crc kubenswrapper[4775]: I1002 03:15:50.819751 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-combined-ca-bundle\") pod \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\" (UID: \"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6\") " Oct 02 03:15:50 crc kubenswrapper[4775]: I1002 03:15:50.823972 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-kube-api-access-qsv4d" (OuterVolumeSpecName: "kube-api-access-qsv4d") pod "ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6" (UID: "ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6"). InnerVolumeSpecName "kube-api-access-qsv4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:15:50 crc kubenswrapper[4775]: I1002 03:15:50.843633 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6" (UID: "ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:50 crc kubenswrapper[4775]: I1002 03:15:50.851999 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-config-data" (OuterVolumeSpecName: "config-data") pod "ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6" (UID: "ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:50 crc kubenswrapper[4775]: I1002 03:15:50.922487 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:50 crc kubenswrapper[4775]: I1002 03:15:50.922542 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsv4d\" (UniqueName: \"kubernetes.io/projected/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-kube-api-access-qsv4d\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:50 crc kubenswrapper[4775]: I1002 03:15:50.922563 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:51 crc kubenswrapper[4775]: E1002 03:15:51.129344 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="617d1d307ef28adc7ecba70884e00a37f469c79f8a45e6925777dabfade04be1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 03:15:51 crc kubenswrapper[4775]: E1002 03:15:51.132281 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="617d1d307ef28adc7ecba70884e00a37f469c79f8a45e6925777dabfade04be1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 03:15:51 crc kubenswrapper[4775]: E1002 03:15:51.134399 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="617d1d307ef28adc7ecba70884e00a37f469c79f8a45e6925777dabfade04be1" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 03:15:51 crc kubenswrapper[4775]: E1002 03:15:51.134501 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="ca3f9270-7d6a-4b91-bf50-944388836500" containerName="nova-cell1-conductor-conductor" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.145137 4775 generic.go:334] "Generic (PLEG): container finished" podID="ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6" containerID="1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935" exitCode=0 Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.145259 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.145348 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6","Type":"ContainerDied","Data":"1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935"} Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.145417 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6","Type":"ContainerDied","Data":"eaccc751aaef3b57d17001d4b967c99661ac7a6f44e0605e23fbfef6f34d419f"} Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.145461 4775 scope.go:117] "RemoveContainer" containerID="1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.170577 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8986cac6-b070-4ea8-a961-24481fbb7bdd","Type":"ContainerStarted","Data":"276887ff0f0d687bacfb0513089001e62180ad08e2d3bcfc486bc4e5bd888e9f"} Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.170687 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8986cac6-b070-4ea8-a961-24481fbb7bdd","Type":"ContainerStarted","Data":"126f1fe0b63d9c60ba34b2fc8a728348a966ea5fb359156765cd0bce9ddb9ee4"} Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.176486 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"def5a435-381a-4636-932e-8e4881989de3","Type":"ContainerStarted","Data":"4c136b0ad6b8a589bc80d338cd4a23c9b9f0075532593ecff610f11104b1f4a9"} Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.176537 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"def5a435-381a-4636-932e-8e4881989de3","Type":"ContainerStarted","Data":"36093f2c01943f0cca62904703c8989a7da89a2ab6946a9b8cec36ca74fb80b4"} Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.176557 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"def5a435-381a-4636-932e-8e4881989de3","Type":"ContainerStarted","Data":"298fd9920f4ecbc49ae3edaaccc776395e881b56c77c9178b9ebab80fcee16c2"} Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.218734 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.225081 4775 scope.go:117] "RemoveContainer" containerID="1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935" Oct 02 03:15:51 crc kubenswrapper[4775]: E1002 03:15:51.225539 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935\": container with ID starting with 1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935 not found: ID does not exist" containerID="1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.225587 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935"} err="failed to get container status \"1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935\": rpc error: code = NotFound desc = could not find container \"1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935\": container with ID starting with 1ae1366eca6805bae087e728a84e869552fb02ddb39409aa3d9df9542a8d6935 not found: ID does not exist" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.232433 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.246079 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:15:51 crc kubenswrapper[4775]: E1002 03:15:51.246486 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6" containerName="nova-scheduler-scheduler" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.246504 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6" containerName="nova-scheduler-scheduler" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.246709 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6" containerName="nova-scheduler-scheduler" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.247325 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.248400 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.247933394 podStartE2EDuration="2.247933394s" podCreationTimestamp="2025-10-02 03:15:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:15:51.214265092 +0000 UTC m=+5688.381009142" watchObservedRunningTime="2025-10-02 03:15:51.247933394 +0000 UTC m=+5688.414677434" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.249181 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.259471 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.262209 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.262200334 podStartE2EDuration="2.262200334s" podCreationTimestamp="2025-10-02 03:15:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:15:51.243324935 +0000 UTC m=+5688.410068965" watchObservedRunningTime="2025-10-02 03:15:51.262200334 +0000 UTC m=+5688.428944374" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.329425 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02de0e68-b4be-42b1-80eb-b4cb50dbd456-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\") " pod="openstack/nova-scheduler-0" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.329581 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02de0e68-b4be-42b1-80eb-b4cb50dbd456-config-data\") pod \"nova-scheduler-0\" (UID: \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\") " pod="openstack/nova-scheduler-0" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.329793 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnwcp\" (UniqueName: \"kubernetes.io/projected/02de0e68-b4be-42b1-80eb-b4cb50dbd456-kube-api-access-bnwcp\") pod \"nova-scheduler-0\" (UID: \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\") " pod="openstack/nova-scheduler-0" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.432753 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02de0e68-b4be-42b1-80eb-b4cb50dbd456-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\") " pod="openstack/nova-scheduler-0" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.432880 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02de0e68-b4be-42b1-80eb-b4cb50dbd456-config-data\") pod \"nova-scheduler-0\" (UID: \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\") " pod="openstack/nova-scheduler-0" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.433063 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnwcp\" (UniqueName: \"kubernetes.io/projected/02de0e68-b4be-42b1-80eb-b4cb50dbd456-kube-api-access-bnwcp\") pod \"nova-scheduler-0\" (UID: \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\") " pod="openstack/nova-scheduler-0" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.439574 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02de0e68-b4be-42b1-80eb-b4cb50dbd456-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\") " pod="openstack/nova-scheduler-0" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.456771 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02de0e68-b4be-42b1-80eb-b4cb50dbd456-config-data\") pod \"nova-scheduler-0\" (UID: \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\") " pod="openstack/nova-scheduler-0" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.464921 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnwcp\" (UniqueName: \"kubernetes.io/projected/02de0e68-b4be-42b1-80eb-b4cb50dbd456-kube-api-access-bnwcp\") pod \"nova-scheduler-0\" (UID: \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\") " pod="openstack/nova-scheduler-0" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.581916 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 03:15:51 crc kubenswrapper[4775]: I1002 03:15:51.780340 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6" path="/var/lib/kubelet/pods/ca438004-f9c3-4aa4-b6cb-abd07dbf8cf6/volumes" Oct 02 03:15:52 crc kubenswrapper[4775]: I1002 03:15:52.072368 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 03:15:52 crc kubenswrapper[4775]: I1002 03:15:52.191539 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02de0e68-b4be-42b1-80eb-b4cb50dbd456","Type":"ContainerStarted","Data":"96ad8cd1898f5a30da72eaf36eeb585010c87fc803f6f93c740b42da44ecd9e1"} Oct 02 03:15:52 crc kubenswrapper[4775]: I1002 03:15:52.433777 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:53 crc kubenswrapper[4775]: I1002 03:15:53.205830 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02de0e68-b4be-42b1-80eb-b4cb50dbd456","Type":"ContainerStarted","Data":"2d5246027af2edbc8f1f0bcfb9f5c32756879a301b041860ced369a8eae8dc35"} Oct 02 03:15:53 crc kubenswrapper[4775]: I1002 03:15:53.235748 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.235719255 podStartE2EDuration="2.235719255s" podCreationTimestamp="2025-10-02 03:15:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:15:53.230355116 +0000 UTC m=+5690.397099196" watchObservedRunningTime="2025-10-02 03:15:53.235719255 +0000 UTC m=+5690.402463335" Oct 02 03:15:54 crc kubenswrapper[4775]: I1002 03:15:54.570519 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 03:15:54 crc kubenswrapper[4775]: I1002 03:15:54.570808 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 03:15:55 crc kubenswrapper[4775]: I1002 03:15:55.236947 4775 generic.go:334] "Generic (PLEG): container finished" podID="ca3f9270-7d6a-4b91-bf50-944388836500" containerID="617d1d307ef28adc7ecba70884e00a37f469c79f8a45e6925777dabfade04be1" exitCode=0 Oct 02 03:15:55 crc kubenswrapper[4775]: I1002 03:15:55.237021 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ca3f9270-7d6a-4b91-bf50-944388836500","Type":"ContainerDied","Data":"617d1d307ef28adc7ecba70884e00a37f469c79f8a45e6925777dabfade04be1"} Oct 02 03:15:55 crc kubenswrapper[4775]: I1002 03:15:55.396775 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 03:15:55 crc kubenswrapper[4775]: I1002 03:15:55.524849 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3f9270-7d6a-4b91-bf50-944388836500-combined-ca-bundle\") pod \"ca3f9270-7d6a-4b91-bf50-944388836500\" (UID: \"ca3f9270-7d6a-4b91-bf50-944388836500\") " Oct 02 03:15:55 crc kubenswrapper[4775]: I1002 03:15:55.524983 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3f9270-7d6a-4b91-bf50-944388836500-config-data\") pod \"ca3f9270-7d6a-4b91-bf50-944388836500\" (UID: \"ca3f9270-7d6a-4b91-bf50-944388836500\") " Oct 02 03:15:55 crc kubenswrapper[4775]: I1002 03:15:55.525097 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlpwq\" (UniqueName: \"kubernetes.io/projected/ca3f9270-7d6a-4b91-bf50-944388836500-kube-api-access-hlpwq\") pod \"ca3f9270-7d6a-4b91-bf50-944388836500\" (UID: \"ca3f9270-7d6a-4b91-bf50-944388836500\") " Oct 02 03:15:55 crc kubenswrapper[4775]: I1002 03:15:55.532193 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca3f9270-7d6a-4b91-bf50-944388836500-kube-api-access-hlpwq" (OuterVolumeSpecName: "kube-api-access-hlpwq") pod "ca3f9270-7d6a-4b91-bf50-944388836500" (UID: "ca3f9270-7d6a-4b91-bf50-944388836500"). InnerVolumeSpecName "kube-api-access-hlpwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:15:55 crc kubenswrapper[4775]: I1002 03:15:55.548439 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca3f9270-7d6a-4b91-bf50-944388836500-config-data" (OuterVolumeSpecName: "config-data") pod "ca3f9270-7d6a-4b91-bf50-944388836500" (UID: "ca3f9270-7d6a-4b91-bf50-944388836500"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:55 crc kubenswrapper[4775]: I1002 03:15:55.568102 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca3f9270-7d6a-4b91-bf50-944388836500-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca3f9270-7d6a-4b91-bf50-944388836500" (UID: "ca3f9270-7d6a-4b91-bf50-944388836500"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:15:55 crc kubenswrapper[4775]: I1002 03:15:55.627907 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlpwq\" (UniqueName: \"kubernetes.io/projected/ca3f9270-7d6a-4b91-bf50-944388836500-kube-api-access-hlpwq\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:55 crc kubenswrapper[4775]: I1002 03:15:55.627982 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca3f9270-7d6a-4b91-bf50-944388836500-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:55 crc kubenswrapper[4775]: I1002 03:15:55.628002 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca3f9270-7d6a-4b91-bf50-944388836500-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.250154 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"ca3f9270-7d6a-4b91-bf50-944388836500","Type":"ContainerDied","Data":"eb3d8b825139545a6b8ec216bdebabf88cda39089d87a1d35bb4b376b44d862d"} Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.250590 4775 scope.go:117] "RemoveContainer" containerID="617d1d307ef28adc7ecba70884e00a37f469c79f8a45e6925777dabfade04be1" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.250287 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.290774 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.315066 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.339239 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 03:15:56 crc kubenswrapper[4775]: E1002 03:15:56.339691 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca3f9270-7d6a-4b91-bf50-944388836500" containerName="nova-cell1-conductor-conductor" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.339715 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca3f9270-7d6a-4b91-bf50-944388836500" containerName="nova-cell1-conductor-conductor" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.339914 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca3f9270-7d6a-4b91-bf50-944388836500" containerName="nova-cell1-conductor-conductor" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.340563 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.340661 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.344025 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.443504 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fababa8-2f3f-47dc-90cd-38105370efbe-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6fababa8-2f3f-47dc-90cd-38105370efbe\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.443610 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fababa8-2f3f-47dc-90cd-38105370efbe-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6fababa8-2f3f-47dc-90cd-38105370efbe\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.443846 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z94ql\" (UniqueName: \"kubernetes.io/projected/6fababa8-2f3f-47dc-90cd-38105370efbe-kube-api-access-z94ql\") pod \"nova-cell1-conductor-0\" (UID: \"6fababa8-2f3f-47dc-90cd-38105370efbe\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.552994 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z94ql\" (UniqueName: \"kubernetes.io/projected/6fababa8-2f3f-47dc-90cd-38105370efbe-kube-api-access-z94ql\") pod \"nova-cell1-conductor-0\" (UID: \"6fababa8-2f3f-47dc-90cd-38105370efbe\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.553158 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fababa8-2f3f-47dc-90cd-38105370efbe-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6fababa8-2f3f-47dc-90cd-38105370efbe\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.553193 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fababa8-2f3f-47dc-90cd-38105370efbe-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6fababa8-2f3f-47dc-90cd-38105370efbe\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.565148 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fababa8-2f3f-47dc-90cd-38105370efbe-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6fababa8-2f3f-47dc-90cd-38105370efbe\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.567263 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fababa8-2f3f-47dc-90cd-38105370efbe-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6fababa8-2f3f-47dc-90cd-38105370efbe\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.579227 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z94ql\" (UniqueName: \"kubernetes.io/projected/6fababa8-2f3f-47dc-90cd-38105370efbe-kube-api-access-z94ql\") pod \"nova-cell1-conductor-0\" (UID: \"6fababa8-2f3f-47dc-90cd-38105370efbe\") " pod="openstack/nova-cell1-conductor-0" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.582759 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 03:15:56 crc kubenswrapper[4775]: I1002 03:15:56.693662 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 03:15:57 crc kubenswrapper[4775]: W1002 03:15:57.217813 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fababa8_2f3f_47dc_90cd_38105370efbe.slice/crio-cb2d5bd18f0c9a1f44fffb7031c59ce8d7066d4c9bc6109c23dd69e176b7a43a WatchSource:0}: Error finding container cb2d5bd18f0c9a1f44fffb7031c59ce8d7066d4c9bc6109c23dd69e176b7a43a: Status 404 returned error can't find the container with id cb2d5bd18f0c9a1f44fffb7031c59ce8d7066d4c9bc6109c23dd69e176b7a43a Oct 02 03:15:57 crc kubenswrapper[4775]: I1002 03:15:57.227510 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 03:15:57 crc kubenswrapper[4775]: I1002 03:15:57.268394 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6fababa8-2f3f-47dc-90cd-38105370efbe","Type":"ContainerStarted","Data":"cb2d5bd18f0c9a1f44fffb7031c59ce8d7066d4c9bc6109c23dd69e176b7a43a"} Oct 02 03:15:57 crc kubenswrapper[4775]: I1002 03:15:57.433489 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:57 crc kubenswrapper[4775]: I1002 03:15:57.448125 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:57 crc kubenswrapper[4775]: I1002 03:15:57.777585 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca3f9270-7d6a-4b91-bf50-944388836500" path="/var/lib/kubelet/pods/ca3f9270-7d6a-4b91-bf50-944388836500/volumes" Oct 02 03:15:57 crc kubenswrapper[4775]: I1002 03:15:57.788172 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 02 03:15:58 crc kubenswrapper[4775]: I1002 03:15:58.284238 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6fababa8-2f3f-47dc-90cd-38105370efbe","Type":"ContainerStarted","Data":"3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f"} Oct 02 03:15:58 crc kubenswrapper[4775]: I1002 03:15:58.300649 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 02 03:15:58 crc kubenswrapper[4775]: I1002 03:15:58.329697 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.329666262 podStartE2EDuration="2.329666262s" podCreationTimestamp="2025-10-02 03:15:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:15:58.306338037 +0000 UTC m=+5695.473082117" watchObservedRunningTime="2025-10-02 03:15:58.329666262 +0000 UTC m=+5695.496410332" Oct 02 03:15:59 crc kubenswrapper[4775]: I1002 03:15:59.293755 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 02 03:15:59 crc kubenswrapper[4775]: I1002 03:15:59.514679 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 03:15:59 crc kubenswrapper[4775]: I1002 03:15:59.514732 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 03:15:59 crc kubenswrapper[4775]: I1002 03:15:59.570617 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 03:15:59 crc kubenswrapper[4775]: I1002 03:15:59.571116 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 03:16:00 crc kubenswrapper[4775]: I1002 03:16:00.598335 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8986cac6-b070-4ea8-a961-24481fbb7bdd" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.84:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 03:16:00 crc kubenswrapper[4775]: I1002 03:16:00.598303 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="8986cac6-b070-4ea8-a961-24481fbb7bdd" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.84:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 03:16:00 crc kubenswrapper[4775]: I1002 03:16:00.681354 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="def5a435-381a-4636-932e-8e4881989de3" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.85:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 03:16:00 crc kubenswrapper[4775]: I1002 03:16:00.681360 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="def5a435-381a-4636-932e-8e4881989de3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.85:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 03:16:01 crc kubenswrapper[4775]: I1002 03:16:01.583526 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 03:16:01 crc kubenswrapper[4775]: I1002 03:16:01.620127 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 03:16:02 crc kubenswrapper[4775]: I1002 03:16:02.354854 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 03:16:02 crc kubenswrapper[4775]: I1002 03:16:02.766852 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:16:02 crc kubenswrapper[4775]: E1002 03:16:02.767144 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.170394 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.177383 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.182006 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.191450 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.249131 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.249193 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-scripts\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.249232 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-config-data\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.249432 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brkq5\" (UniqueName: \"kubernetes.io/projected/2b443853-28ea-4f48-bb18-4f1adf10961a-kube-api-access-brkq5\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.249523 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.249628 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b443853-28ea-4f48-bb18-4f1adf10961a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.351177 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brkq5\" (UniqueName: \"kubernetes.io/projected/2b443853-28ea-4f48-bb18-4f1adf10961a-kube-api-access-brkq5\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.351296 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.351412 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b443853-28ea-4f48-bb18-4f1adf10961a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.351491 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.351530 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-scripts\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.351580 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-config-data\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.351599 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b443853-28ea-4f48-bb18-4f1adf10961a-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.358231 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-scripts\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.360368 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.364462 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.364566 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-config-data\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.381323 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brkq5\" (UniqueName: \"kubernetes.io/projected/2b443853-28ea-4f48-bb18-4f1adf10961a-kube-api-access-brkq5\") pod \"cinder-scheduler-0\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.506860 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 03:16:05 crc kubenswrapper[4775]: I1002 03:16:05.809920 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 03:16:06 crc kubenswrapper[4775]: I1002 03:16:06.374868 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2b443853-28ea-4f48-bb18-4f1adf10961a","Type":"ContainerStarted","Data":"9bba9ad0051e7306980ff7e54177132f7a1cdd01a776b8d83c9f56ca24e8c972"} Oct 02 03:16:06 crc kubenswrapper[4775]: I1002 03:16:06.748961 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 02 03:16:06 crc kubenswrapper[4775]: I1002 03:16:06.839270 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 03:16:06 crc kubenswrapper[4775]: I1002 03:16:06.839565 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="53d272f6-0ff6-4f2f-b386-903a04d3f3ef" containerName="cinder-api-log" containerID="cri-o://5f96cf4b3c64f46ee14b304a352fcdfcc326262df09b1c84cb7b5ead68dba9e8" gracePeriod=30 Oct 02 03:16:06 crc kubenswrapper[4775]: I1002 03:16:06.839831 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="53d272f6-0ff6-4f2f-b386-903a04d3f3ef" containerName="cinder-api" containerID="cri-o://697f8eaf8a54462a09e6c76a1581e618c690316ac6e544ba1ab4552aa24cd90e" gracePeriod=30 Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.383240 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d272f6-0ff6-4f2f-b386-903a04d3f3ef" containerID="5f96cf4b3c64f46ee14b304a352fcdfcc326262df09b1c84cb7b5ead68dba9e8" exitCode=143 Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.383372 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"53d272f6-0ff6-4f2f-b386-903a04d3f3ef","Type":"ContainerDied","Data":"5f96cf4b3c64f46ee14b304a352fcdfcc326262df09b1c84cb7b5ead68dba9e8"} Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.385483 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2b443853-28ea-4f48-bb18-4f1adf10961a","Type":"ContainerStarted","Data":"9a77ab16b75e89acc319f1545b286cbeb894e8d0f4f377c5bd36b4a679a8dfe2"} Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.385569 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2b443853-28ea-4f48-bb18-4f1adf10961a","Type":"ContainerStarted","Data":"01888fcf624bb0d784c04b9a18d9ae756e2805a8eafc6ab7ee3f1f6d4ac35205"} Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.407350 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.40733033 podStartE2EDuration="2.40733033s" podCreationTimestamp="2025-10-02 03:16:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:16:07.404747463 +0000 UTC m=+5704.571491523" watchObservedRunningTime="2025-10-02 03:16:07.40733033 +0000 UTC m=+5704.574074380" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.607490 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.611711 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.613674 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.624946 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.705882 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.705936 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e763e57d-3848-40c3-ac6f-29987d3070cf-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.705990 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e763e57d-3848-40c3-ac6f-29987d3070cf-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.706026 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.706051 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-dev\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.706076 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e763e57d-3848-40c3-ac6f-29987d3070cf-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.706119 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.706140 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6vsq\" (UniqueName: \"kubernetes.io/projected/e763e57d-3848-40c3-ac6f-29987d3070cf-kube-api-access-z6vsq\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.706204 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-run\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.706248 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-sys\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.706287 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.706310 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e763e57d-3848-40c3-ac6f-29987d3070cf-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.706341 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.706358 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.706398 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.706419 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e763e57d-3848-40c3-ac6f-29987d3070cf-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.807728 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e763e57d-3848-40c3-ac6f-29987d3070cf-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.808001 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.808099 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.808190 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.808254 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e763e57d-3848-40c3-ac6f-29987d3070cf-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.808336 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.808400 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e763e57d-3848-40c3-ac6f-29987d3070cf-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.808514 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.808256 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.808717 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.809079 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.808471 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e763e57d-3848-40c3-ac6f-29987d3070cf-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.809553 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.809637 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-dev\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.809751 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e763e57d-3848-40c3-ac6f-29987d3070cf-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.809858 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.809924 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6vsq\" (UniqueName: \"kubernetes.io/projected/e763e57d-3848-40c3-ac6f-29987d3070cf-kube-api-access-z6vsq\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.810039 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-run\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.810119 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-sys\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.809631 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.809671 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-dev\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.810313 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-sys\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.810389 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-run\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.810538 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.810793 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.810749 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e763e57d-3848-40c3-ac6f-29987d3070cf-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.813370 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e763e57d-3848-40c3-ac6f-29987d3070cf-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.814244 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e763e57d-3848-40c3-ac6f-29987d3070cf-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.814438 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e763e57d-3848-40c3-ac6f-29987d3070cf-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.815089 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e763e57d-3848-40c3-ac6f-29987d3070cf-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.831475 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e763e57d-3848-40c3-ac6f-29987d3070cf-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.834096 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6vsq\" (UniqueName: \"kubernetes.io/projected/e763e57d-3848-40c3-ac6f-29987d3070cf-kube-api-access-z6vsq\") pod \"cinder-volume-volume1-0\" (UID: \"e763e57d-3848-40c3-ac6f-29987d3070cf\") " pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:07 crc kubenswrapper[4775]: I1002 03:16:07.939688 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.125345 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.127377 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.137969 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.166894 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.228439 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.228877 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-lib-modules\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.228923 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.231577 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-sys\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.231661 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-dev\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.231695 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.231744 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.231770 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j25w\" (UniqueName: \"kubernetes.io/projected/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-kube-api-access-5j25w\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.231789 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-config-data-custom\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.231807 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-scripts\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.231836 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.231884 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-etc-nvme\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.231903 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-ceph\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.231933 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-config-data\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.231967 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.232101 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-run\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333216 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333362 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j25w\" (UniqueName: \"kubernetes.io/projected/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-kube-api-access-5j25w\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333384 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-config-data-custom\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333398 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-scripts\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333419 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333445 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-etc-nvme\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333460 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-ceph\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333477 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-config-data\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333494 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333513 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-run\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333556 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333591 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-lib-modules\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333620 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333635 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-sys\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333663 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-dev\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333682 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333757 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.333302 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.335051 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-etc-nvme\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.335084 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-lib-modules\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.335156 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-run\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.335221 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.335253 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.335278 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-dev\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.335816 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-sys\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.335945 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.341785 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-scripts\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.342365 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-config-data-custom\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.346748 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.351022 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-config-data\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.358041 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-ceph\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.359593 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j25w\" (UniqueName: \"kubernetes.io/projected/dc936ddc-e6af-4e0a-85b0-732eddb8ca82-kube-api-access-5j25w\") pod \"cinder-backup-0\" (UID: \"dc936ddc-e6af-4e0a-85b0-732eddb8ca82\") " pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.477459 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.665439 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 02 03:16:08 crc kubenswrapper[4775]: I1002 03:16:08.669813 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 03:16:09 crc kubenswrapper[4775]: I1002 03:16:09.010320 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 02 03:16:09 crc kubenswrapper[4775]: W1002 03:16:09.015860 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc936ddc_e6af_4e0a_85b0_732eddb8ca82.slice/crio-a1b17bff4db75719219dc75b6eb20af66486785b7490fbe0a45f5bc640bcf012 WatchSource:0}: Error finding container a1b17bff4db75719219dc75b6eb20af66486785b7490fbe0a45f5bc640bcf012: Status 404 returned error can't find the container with id a1b17bff4db75719219dc75b6eb20af66486785b7490fbe0a45f5bc640bcf012 Oct 02 03:16:09 crc kubenswrapper[4775]: I1002 03:16:09.418175 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"dc936ddc-e6af-4e0a-85b0-732eddb8ca82","Type":"ContainerStarted","Data":"a1b17bff4db75719219dc75b6eb20af66486785b7490fbe0a45f5bc640bcf012"} Oct 02 03:16:09 crc kubenswrapper[4775]: I1002 03:16:09.423254 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"e763e57d-3848-40c3-ac6f-29987d3070cf","Type":"ContainerStarted","Data":"672a071707222c2fabafc0c1a86588de8e6794406b1222848b0a8e3779c8044f"} Oct 02 03:16:09 crc kubenswrapper[4775]: I1002 03:16:09.518985 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 03:16:09 crc kubenswrapper[4775]: I1002 03:16:09.519587 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 03:16:09 crc kubenswrapper[4775]: I1002 03:16:09.520590 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 03:16:09 crc kubenswrapper[4775]: I1002 03:16:09.532443 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 03:16:09 crc kubenswrapper[4775]: I1002 03:16:09.572950 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 03:16:09 crc kubenswrapper[4775]: I1002 03:16:09.574159 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 03:16:09 crc kubenswrapper[4775]: I1002 03:16:09.584640 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 03:16:09 crc kubenswrapper[4775]: I1002 03:16:09.926720 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="53d272f6-0ff6-4f2f-b386-903a04d3f3ef" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.1.81:8776/healthcheck\": read tcp 10.217.0.2:48340->10.217.1.81:8776: read: connection reset by peer" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.240649 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.381328 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-scripts\") pod \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.381659 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-config-data-custom\") pod \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.381826 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-etc-machine-id\") pod \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.381880 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm26h\" (UniqueName: \"kubernetes.io/projected/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-kube-api-access-pm26h\") pod \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.381918 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-combined-ca-bundle\") pod \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.381935 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-logs\") pod \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.382014 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-config-data\") pod \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\" (UID: \"53d272f6-0ff6-4f2f-b386-903a04d3f3ef\") " Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.382025 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "53d272f6-0ff6-4f2f-b386-903a04d3f3ef" (UID: "53d272f6-0ff6-4f2f-b386-903a04d3f3ef"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.382349 4775 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.382771 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-logs" (OuterVolumeSpecName: "logs") pod "53d272f6-0ff6-4f2f-b386-903a04d3f3ef" (UID: "53d272f6-0ff6-4f2f-b386-903a04d3f3ef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.386877 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-scripts" (OuterVolumeSpecName: "scripts") pod "53d272f6-0ff6-4f2f-b386-903a04d3f3ef" (UID: "53d272f6-0ff6-4f2f-b386-903a04d3f3ef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.387243 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "53d272f6-0ff6-4f2f-b386-903a04d3f3ef" (UID: "53d272f6-0ff6-4f2f-b386-903a04d3f3ef"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.389022 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-kube-api-access-pm26h" (OuterVolumeSpecName: "kube-api-access-pm26h") pod "53d272f6-0ff6-4f2f-b386-903a04d3f3ef" (UID: "53d272f6-0ff6-4f2f-b386-903a04d3f3ef"). InnerVolumeSpecName "kube-api-access-pm26h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.437522 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"dc936ddc-e6af-4e0a-85b0-732eddb8ca82","Type":"ContainerStarted","Data":"ebc749c1544afc8bbb73f474177d6c3dfb2ff24723abf35f04a018fe818b19d4"} Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.440727 4775 generic.go:334] "Generic (PLEG): container finished" podID="53d272f6-0ff6-4f2f-b386-903a04d3f3ef" containerID="697f8eaf8a54462a09e6c76a1581e618c690316ac6e544ba1ab4552aa24cd90e" exitCode=0 Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.440810 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"53d272f6-0ff6-4f2f-b386-903a04d3f3ef","Type":"ContainerDied","Data":"697f8eaf8a54462a09e6c76a1581e618c690316ac6e544ba1ab4552aa24cd90e"} Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.440849 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"53d272f6-0ff6-4f2f-b386-903a04d3f3ef","Type":"ContainerDied","Data":"79e30db6022f42ebc4cb7f73179c6864b1ced8053bf25f3b7b5b63dd1e339ad3"} Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.440883 4775 scope.go:117] "RemoveContainer" containerID="697f8eaf8a54462a09e6c76a1581e618c690316ac6e544ba1ab4552aa24cd90e" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.440893 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.442773 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"e763e57d-3848-40c3-ac6f-29987d3070cf","Type":"ContainerStarted","Data":"31f4ae0f27c576878dc406276d891995a341cbfbd61a732f01cd807796f8bf91"} Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.442825 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"e763e57d-3848-40c3-ac6f-29987d3070cf","Type":"ContainerStarted","Data":"4befeafc3f3af5d311f43285c74bf513e391c3241b2f768e731e252d68963450"} Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.443513 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.450304 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.453473 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53d272f6-0ff6-4f2f-b386-903a04d3f3ef" (UID: "53d272f6-0ff6-4f2f-b386-903a04d3f3ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.459836 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.477142 4775 scope.go:117] "RemoveContainer" containerID="5f96cf4b3c64f46ee14b304a352fcdfcc326262df09b1c84cb7b5ead68dba9e8" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.484036 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.484060 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.484069 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.484079 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.484087 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm26h\" (UniqueName: \"kubernetes.io/projected/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-kube-api-access-pm26h\") on node \"crc\" DevicePath \"\"" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.489995 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.70733586 podStartE2EDuration="3.489977447s" podCreationTimestamp="2025-10-02 03:16:07 +0000 UTC" firstStartedPulling="2025-10-02 03:16:08.669566093 +0000 UTC m=+5705.836310123" lastFinishedPulling="2025-10-02 03:16:09.45220766 +0000 UTC m=+5706.618951710" observedRunningTime="2025-10-02 03:16:10.471293003 +0000 UTC m=+5707.638037033" watchObservedRunningTime="2025-10-02 03:16:10.489977447 +0000 UTC m=+5707.656721487" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.491318 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-config-data" (OuterVolumeSpecName: "config-data") pod "53d272f6-0ff6-4f2f-b386-903a04d3f3ef" (UID: "53d272f6-0ff6-4f2f-b386-903a04d3f3ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.495759 4775 scope.go:117] "RemoveContainer" containerID="697f8eaf8a54462a09e6c76a1581e618c690316ac6e544ba1ab4552aa24cd90e" Oct 02 03:16:10 crc kubenswrapper[4775]: E1002 03:16:10.496461 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"697f8eaf8a54462a09e6c76a1581e618c690316ac6e544ba1ab4552aa24cd90e\": container with ID starting with 697f8eaf8a54462a09e6c76a1581e618c690316ac6e544ba1ab4552aa24cd90e not found: ID does not exist" containerID="697f8eaf8a54462a09e6c76a1581e618c690316ac6e544ba1ab4552aa24cd90e" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.496507 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"697f8eaf8a54462a09e6c76a1581e618c690316ac6e544ba1ab4552aa24cd90e"} err="failed to get container status \"697f8eaf8a54462a09e6c76a1581e618c690316ac6e544ba1ab4552aa24cd90e\": rpc error: code = NotFound desc = could not find container \"697f8eaf8a54462a09e6c76a1581e618c690316ac6e544ba1ab4552aa24cd90e\": container with ID starting with 697f8eaf8a54462a09e6c76a1581e618c690316ac6e544ba1ab4552aa24cd90e not found: ID does not exist" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.496534 4775 scope.go:117] "RemoveContainer" containerID="5f96cf4b3c64f46ee14b304a352fcdfcc326262df09b1c84cb7b5ead68dba9e8" Oct 02 03:16:10 crc kubenswrapper[4775]: E1002 03:16:10.496969 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f96cf4b3c64f46ee14b304a352fcdfcc326262df09b1c84cb7b5ead68dba9e8\": container with ID starting with 5f96cf4b3c64f46ee14b304a352fcdfcc326262df09b1c84cb7b5ead68dba9e8 not found: ID does not exist" containerID="5f96cf4b3c64f46ee14b304a352fcdfcc326262df09b1c84cb7b5ead68dba9e8" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.497000 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f96cf4b3c64f46ee14b304a352fcdfcc326262df09b1c84cb7b5ead68dba9e8"} err="failed to get container status \"5f96cf4b3c64f46ee14b304a352fcdfcc326262df09b1c84cb7b5ead68dba9e8\": rpc error: code = NotFound desc = could not find container \"5f96cf4b3c64f46ee14b304a352fcdfcc326262df09b1c84cb7b5ead68dba9e8\": container with ID starting with 5f96cf4b3c64f46ee14b304a352fcdfcc326262df09b1c84cb7b5ead68dba9e8 not found: ID does not exist" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.508161 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.585651 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53d272f6-0ff6-4f2f-b386-903a04d3f3ef-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.793811 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.824527 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.835915 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 02 03:16:10 crc kubenswrapper[4775]: E1002 03:16:10.837470 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d272f6-0ff6-4f2f-b386-903a04d3f3ef" containerName="cinder-api" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.837745 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d272f6-0ff6-4f2f-b386-903a04d3f3ef" containerName="cinder-api" Oct 02 03:16:10 crc kubenswrapper[4775]: E1002 03:16:10.837830 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d272f6-0ff6-4f2f-b386-903a04d3f3ef" containerName="cinder-api-log" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.837902 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d272f6-0ff6-4f2f-b386-903a04d3f3ef" containerName="cinder-api-log" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.838376 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d272f6-0ff6-4f2f-b386-903a04d3f3ef" containerName="cinder-api-log" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.838482 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d272f6-0ff6-4f2f-b386-903a04d3f3ef" containerName="cinder-api" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.840100 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.851968 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 02 03:16:10 crc kubenswrapper[4775]: I1002 03:16:10.857905 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.000206 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-scripts\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.000405 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-config-data-custom\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.000449 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-config-data\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.000480 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.000505 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-logs\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.000589 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.000784 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dghr9\" (UniqueName: \"kubernetes.io/projected/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-kube-api-access-dghr9\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.102324 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-config-data-custom\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.102399 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-config-data\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.102434 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.102459 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-logs\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.102508 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.102611 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dghr9\" (UniqueName: \"kubernetes.io/projected/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-kube-api-access-dghr9\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.102681 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-scripts\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.103495 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-logs\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.103597 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.107904 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.116792 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-scripts\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.118028 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-config-data-custom\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.118946 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-config-data\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.120537 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dghr9\" (UniqueName: \"kubernetes.io/projected/88a07e5c-e647-4a80-9397-fc2cc1f54fcd-kube-api-access-dghr9\") pod \"cinder-api-0\" (UID: \"88a07e5c-e647-4a80-9397-fc2cc1f54fcd\") " pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.189209 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.452267 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"dc936ddc-e6af-4e0a-85b0-732eddb8ca82","Type":"ContainerStarted","Data":"23da866edc6ae43030b9c01fce5fa158cd3a236669ddcb0623246bfe64cd6b51"} Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.478511 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.6445274100000002 podStartE2EDuration="3.478490827s" podCreationTimestamp="2025-10-02 03:16:08 +0000 UTC" firstStartedPulling="2025-10-02 03:16:09.020260109 +0000 UTC m=+5706.187004139" lastFinishedPulling="2025-10-02 03:16:09.854223466 +0000 UTC m=+5707.020967556" observedRunningTime="2025-10-02 03:16:11.47162209 +0000 UTC m=+5708.638366140" watchObservedRunningTime="2025-10-02 03:16:11.478490827 +0000 UTC m=+5708.645234877" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.649182 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 02 03:16:11 crc kubenswrapper[4775]: W1002 03:16:11.657159 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88a07e5c_e647_4a80_9397_fc2cc1f54fcd.slice/crio-34e6b621e4f9d71f2530b615e7cdeb6bc2d086d4e79a6c4ad17ed628cac22b92 WatchSource:0}: Error finding container 34e6b621e4f9d71f2530b615e7cdeb6bc2d086d4e79a6c4ad17ed628cac22b92: Status 404 returned error can't find the container with id 34e6b621e4f9d71f2530b615e7cdeb6bc2d086d4e79a6c4ad17ed628cac22b92 Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.739069 4775 scope.go:117] "RemoveContainer" containerID="2329878b563daff78e069e7606b84295cf44f86a97265abb55cd587fac699ae7" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.764280 4775 scope.go:117] "RemoveContainer" containerID="21cd184b426ef67a8be2aafaac5325aa43b84823fcbfbd2e51d22adaebadbe26" Oct 02 03:16:11 crc kubenswrapper[4775]: I1002 03:16:11.786213 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53d272f6-0ff6-4f2f-b386-903a04d3f3ef" path="/var/lib/kubelet/pods/53d272f6-0ff6-4f2f-b386-903a04d3f3ef/volumes" Oct 02 03:16:12 crc kubenswrapper[4775]: I1002 03:16:12.482556 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"88a07e5c-e647-4a80-9397-fc2cc1f54fcd","Type":"ContainerStarted","Data":"b2156a1e56db520c655f0f37eff991c2c810d8496656744aec829e4e2d270bee"} Oct 02 03:16:12 crc kubenswrapper[4775]: I1002 03:16:12.482852 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"88a07e5c-e647-4a80-9397-fc2cc1f54fcd","Type":"ContainerStarted","Data":"34e6b621e4f9d71f2530b615e7cdeb6bc2d086d4e79a6c4ad17ed628cac22b92"} Oct 02 03:16:12 crc kubenswrapper[4775]: I1002 03:16:12.940335 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:13 crc kubenswrapper[4775]: I1002 03:16:13.478360 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 02 03:16:13 crc kubenswrapper[4775]: I1002 03:16:13.492735 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"88a07e5c-e647-4a80-9397-fc2cc1f54fcd","Type":"ContainerStarted","Data":"b29baa87cfad77eea203ee6eccf495a3d4eb41958de2f2769247e7c964a6fb1b"} Oct 02 03:16:13 crc kubenswrapper[4775]: I1002 03:16:13.519746 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.519723863 podStartE2EDuration="3.519723863s" podCreationTimestamp="2025-10-02 03:16:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:16:13.511744496 +0000 UTC m=+5710.678488546" watchObservedRunningTime="2025-10-02 03:16:13.519723863 +0000 UTC m=+5710.686467913" Oct 02 03:16:13 crc kubenswrapper[4775]: I1002 03:16:13.792434 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:16:13 crc kubenswrapper[4775]: E1002 03:16:13.793149 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:16:14 crc kubenswrapper[4775]: I1002 03:16:14.508379 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 02 03:16:15 crc kubenswrapper[4775]: I1002 03:16:15.731245 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 03:16:15 crc kubenswrapper[4775]: I1002 03:16:15.835421 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 03:16:16 crc kubenswrapper[4775]: I1002 03:16:16.533134 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="2b443853-28ea-4f48-bb18-4f1adf10961a" containerName="cinder-scheduler" containerID="cri-o://01888fcf624bb0d784c04b9a18d9ae756e2805a8eafc6ab7ee3f1f6d4ac35205" gracePeriod=30 Oct 02 03:16:16 crc kubenswrapper[4775]: I1002 03:16:16.533222 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="2b443853-28ea-4f48-bb18-4f1adf10961a" containerName="probe" containerID="cri-o://9a77ab16b75e89acc319f1545b286cbeb894e8d0f4f377c5bd36b4a679a8dfe2" gracePeriod=30 Oct 02 03:16:17 crc kubenswrapper[4775]: I1002 03:16:17.549435 4775 generic.go:334] "Generic (PLEG): container finished" podID="2b443853-28ea-4f48-bb18-4f1adf10961a" containerID="9a77ab16b75e89acc319f1545b286cbeb894e8d0f4f377c5bd36b4a679a8dfe2" exitCode=0 Oct 02 03:16:17 crc kubenswrapper[4775]: I1002 03:16:17.549503 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2b443853-28ea-4f48-bb18-4f1adf10961a","Type":"ContainerDied","Data":"9a77ab16b75e89acc319f1545b286cbeb894e8d0f4f377c5bd36b4a679a8dfe2"} Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.175158 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.559276 4775 generic.go:334] "Generic (PLEG): container finished" podID="2b443853-28ea-4f48-bb18-4f1adf10961a" containerID="01888fcf624bb0d784c04b9a18d9ae756e2805a8eafc6ab7ee3f1f6d4ac35205" exitCode=0 Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.559469 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2b443853-28ea-4f48-bb18-4f1adf10961a","Type":"ContainerDied","Data":"01888fcf624bb0d784c04b9a18d9ae756e2805a8eafc6ab7ee3f1f6d4ac35205"} Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.559610 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"2b443853-28ea-4f48-bb18-4f1adf10961a","Type":"ContainerDied","Data":"9bba9ad0051e7306980ff7e54177132f7a1cdd01a776b8d83c9f56ca24e8c972"} Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.559634 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bba9ad0051e7306980ff7e54177132f7a1cdd01a776b8d83c9f56ca24e8c972" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.588501 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.594908 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b443853-28ea-4f48-bb18-4f1adf10961a-etc-machine-id\") pod \"2b443853-28ea-4f48-bb18-4f1adf10961a\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.595012 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-config-data-custom\") pod \"2b443853-28ea-4f48-bb18-4f1adf10961a\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.595020 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b443853-28ea-4f48-bb18-4f1adf10961a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2b443853-28ea-4f48-bb18-4f1adf10961a" (UID: "2b443853-28ea-4f48-bb18-4f1adf10961a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.595065 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-scripts\") pod \"2b443853-28ea-4f48-bb18-4f1adf10961a\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.595168 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-config-data\") pod \"2b443853-28ea-4f48-bb18-4f1adf10961a\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.595199 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-brkq5\" (UniqueName: \"kubernetes.io/projected/2b443853-28ea-4f48-bb18-4f1adf10961a-kube-api-access-brkq5\") pod \"2b443853-28ea-4f48-bb18-4f1adf10961a\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.595284 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-combined-ca-bundle\") pod \"2b443853-28ea-4f48-bb18-4f1adf10961a\" (UID: \"2b443853-28ea-4f48-bb18-4f1adf10961a\") " Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.595877 4775 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b443853-28ea-4f48-bb18-4f1adf10961a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.603212 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-scripts" (OuterVolumeSpecName: "scripts") pod "2b443853-28ea-4f48-bb18-4f1adf10961a" (UID: "2b443853-28ea-4f48-bb18-4f1adf10961a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.603470 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b443853-28ea-4f48-bb18-4f1adf10961a-kube-api-access-brkq5" (OuterVolumeSpecName: "kube-api-access-brkq5") pod "2b443853-28ea-4f48-bb18-4f1adf10961a" (UID: "2b443853-28ea-4f48-bb18-4f1adf10961a"). InnerVolumeSpecName "kube-api-access-brkq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.605118 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2b443853-28ea-4f48-bb18-4f1adf10961a" (UID: "2b443853-28ea-4f48-bb18-4f1adf10961a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.660115 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b443853-28ea-4f48-bb18-4f1adf10961a" (UID: "2b443853-28ea-4f48-bb18-4f1adf10961a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.673040 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.698487 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.698521 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.698533 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-brkq5\" (UniqueName: \"kubernetes.io/projected/2b443853-28ea-4f48-bb18-4f1adf10961a-kube-api-access-brkq5\") on node \"crc\" DevicePath \"\"" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.698544 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.706539 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-config-data" (OuterVolumeSpecName: "config-data") pod "2b443853-28ea-4f48-bb18-4f1adf10961a" (UID: "2b443853-28ea-4f48-bb18-4f1adf10961a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:16:18 crc kubenswrapper[4775]: I1002 03:16:18.799899 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b443853-28ea-4f48-bb18-4f1adf10961a-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.569294 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.617550 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.632105 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.643058 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 03:16:19 crc kubenswrapper[4775]: E1002 03:16:19.643549 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b443853-28ea-4f48-bb18-4f1adf10961a" containerName="probe" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.643568 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b443853-28ea-4f48-bb18-4f1adf10961a" containerName="probe" Oct 02 03:16:19 crc kubenswrapper[4775]: E1002 03:16:19.643588 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b443853-28ea-4f48-bb18-4f1adf10961a" containerName="cinder-scheduler" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.643596 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b443853-28ea-4f48-bb18-4f1adf10961a" containerName="cinder-scheduler" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.643838 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b443853-28ea-4f48-bb18-4f1adf10961a" containerName="probe" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.643859 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b443853-28ea-4f48-bb18-4f1adf10961a" containerName="cinder-scheduler" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.645112 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.650326 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.685332 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.719804 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abd6be45-e458-4dcf-94b1-39545386cfe2-scripts\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.719856 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd6be45-e458-4dcf-94b1-39545386cfe2-config-data\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.719880 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd6be45-e458-4dcf-94b1-39545386cfe2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.719911 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abd6be45-e458-4dcf-94b1-39545386cfe2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.719931 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/abd6be45-e458-4dcf-94b1-39545386cfe2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.719985 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtszh\" (UniqueName: \"kubernetes.io/projected/abd6be45-e458-4dcf-94b1-39545386cfe2-kube-api-access-rtszh\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.775479 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b443853-28ea-4f48-bb18-4f1adf10961a" path="/var/lib/kubelet/pods/2b443853-28ea-4f48-bb18-4f1adf10961a/volumes" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.822125 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/abd6be45-e458-4dcf-94b1-39545386cfe2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.822244 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtszh\" (UniqueName: \"kubernetes.io/projected/abd6be45-e458-4dcf-94b1-39545386cfe2-kube-api-access-rtszh\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.822375 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abd6be45-e458-4dcf-94b1-39545386cfe2-scripts\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.822430 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd6be45-e458-4dcf-94b1-39545386cfe2-config-data\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.822470 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd6be45-e458-4dcf-94b1-39545386cfe2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.822495 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abd6be45-e458-4dcf-94b1-39545386cfe2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.823522 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/abd6be45-e458-4dcf-94b1-39545386cfe2-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.829496 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/abd6be45-e458-4dcf-94b1-39545386cfe2-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.830859 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abd6be45-e458-4dcf-94b1-39545386cfe2-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.831537 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/abd6be45-e458-4dcf-94b1-39545386cfe2-scripts\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.834585 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/abd6be45-e458-4dcf-94b1-39545386cfe2-config-data\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:19 crc kubenswrapper[4775]: I1002 03:16:19.848278 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtszh\" (UniqueName: \"kubernetes.io/projected/abd6be45-e458-4dcf-94b1-39545386cfe2-kube-api-access-rtszh\") pod \"cinder-scheduler-0\" (UID: \"abd6be45-e458-4dcf-94b1-39545386cfe2\") " pod="openstack/cinder-scheduler-0" Oct 02 03:16:20 crc kubenswrapper[4775]: I1002 03:16:20.002465 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 02 03:16:20 crc kubenswrapper[4775]: I1002 03:16:20.543768 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 02 03:16:20 crc kubenswrapper[4775]: I1002 03:16:20.596772 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"abd6be45-e458-4dcf-94b1-39545386cfe2","Type":"ContainerStarted","Data":"f46b98698272edcd8c08e952f4ab9161ed3abaf3702d4fb37ef09cb855b6ec8b"} Oct 02 03:16:21 crc kubenswrapper[4775]: I1002 03:16:21.608192 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"abd6be45-e458-4dcf-94b1-39545386cfe2","Type":"ContainerStarted","Data":"515809c7fa629f4f2cfeb48b7d33ffb6cc73e6d6693f552ab1f68ffe8e5d740a"} Oct 02 03:16:22 crc kubenswrapper[4775]: I1002 03:16:22.623394 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"abd6be45-e458-4dcf-94b1-39545386cfe2","Type":"ContainerStarted","Data":"769eceeeeedb3e478db15bd62346eb60637851b12155b80e97f91094b87eed29"} Oct 02 03:16:22 crc kubenswrapper[4775]: I1002 03:16:22.659552 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.659526081 podStartE2EDuration="3.659526081s" podCreationTimestamp="2025-10-02 03:16:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:16:22.648650349 +0000 UTC m=+5719.815394439" watchObservedRunningTime="2025-10-02 03:16:22.659526081 +0000 UTC m=+5719.826270151" Oct 02 03:16:22 crc kubenswrapper[4775]: I1002 03:16:22.963505 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 02 03:16:25 crc kubenswrapper[4775]: I1002 03:16:25.002915 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 02 03:16:28 crc kubenswrapper[4775]: I1002 03:16:28.765465 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:16:28 crc kubenswrapper[4775]: E1002 03:16:28.767683 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:16:30 crc kubenswrapper[4775]: I1002 03:16:30.227131 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 02 03:16:42 crc kubenswrapper[4775]: I1002 03:16:42.765388 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:16:42 crc kubenswrapper[4775]: E1002 03:16:42.767480 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:16:54 crc kubenswrapper[4775]: I1002 03:16:54.087130 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-kfp5w"] Oct 02 03:16:54 crc kubenswrapper[4775]: I1002 03:16:54.103079 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-kfp5w"] Oct 02 03:16:55 crc kubenswrapper[4775]: I1002 03:16:55.791909 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a23444b7-1f4c-4927-a42e-fc8703728295" path="/var/lib/kubelet/pods/a23444b7-1f4c-4927-a42e-fc8703728295/volumes" Oct 02 03:16:57 crc kubenswrapper[4775]: I1002 03:16:57.765754 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:16:57 crc kubenswrapper[4775]: E1002 03:16:57.766463 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:17:04 crc kubenswrapper[4775]: I1002 03:17:04.047707 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-3671-account-create-szs6m"] Oct 02 03:17:04 crc kubenswrapper[4775]: I1002 03:17:04.064351 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-3671-account-create-szs6m"] Oct 02 03:17:05 crc kubenswrapper[4775]: I1002 03:17:05.782131 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e2dc5c3-2bcb-476b-9220-8643c983a17b" path="/var/lib/kubelet/pods/1e2dc5c3-2bcb-476b-9220-8643c983a17b/volumes" Oct 02 03:17:10 crc kubenswrapper[4775]: I1002 03:17:10.034709 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-nc6qh"] Oct 02 03:17:10 crc kubenswrapper[4775]: I1002 03:17:10.047887 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-nc6qh"] Oct 02 03:17:10 crc kubenswrapper[4775]: I1002 03:17:10.766601 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:17:10 crc kubenswrapper[4775]: E1002 03:17:10.767103 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:17:11 crc kubenswrapper[4775]: I1002 03:17:11.778907 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daf757e5-76fb-4aee-80c3-d362a58a977b" path="/var/lib/kubelet/pods/daf757e5-76fb-4aee-80c3-d362a58a977b/volumes" Oct 02 03:17:11 crc kubenswrapper[4775]: I1002 03:17:11.998125 4775 scope.go:117] "RemoveContainer" containerID="882b4cc388ee5d3e24b19519ebf9c9b39356accb865da6007f8a8f056098640c" Oct 02 03:17:12 crc kubenswrapper[4775]: I1002 03:17:12.045834 4775 scope.go:117] "RemoveContainer" containerID="5cd71b527c8d218a66d5ee9d6594d6ae90876f30023e50e8ba78b75de4541854" Oct 02 03:17:12 crc kubenswrapper[4775]: I1002 03:17:12.121134 4775 scope.go:117] "RemoveContainer" containerID="1ad1ca145a2617ab096b2670115314f7c65e16bd32d25181a414eecf4ead577b" Oct 02 03:17:23 crc kubenswrapper[4775]: I1002 03:17:23.781664 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:17:23 crc kubenswrapper[4775]: E1002 03:17:23.785191 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:17:24 crc kubenswrapper[4775]: I1002 03:17:24.066898 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rw6jm"] Oct 02 03:17:24 crc kubenswrapper[4775]: I1002 03:17:24.080258 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rw6jm"] Oct 02 03:17:25 crc kubenswrapper[4775]: I1002 03:17:25.790067 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1d8f136-3ac3-485b-af3a-8b0410f2b902" path="/var/lib/kubelet/pods/c1d8f136-3ac3-485b-af3a-8b0410f2b902/volumes" Oct 02 03:17:35 crc kubenswrapper[4775]: I1002 03:17:35.766256 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:17:35 crc kubenswrapper[4775]: E1002 03:17:35.769730 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:17:46 crc kubenswrapper[4775]: I1002 03:17:46.766804 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:17:46 crc kubenswrapper[4775]: E1002 03:17:46.768455 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:17:57 crc kubenswrapper[4775]: I1002 03:17:57.766054 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:17:57 crc kubenswrapper[4775]: E1002 03:17:57.767104 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:18:09 crc kubenswrapper[4775]: I1002 03:18:09.765042 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:18:09 crc kubenswrapper[4775]: E1002 03:18:09.766065 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.255618 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-b27p9"] Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.258104 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.260066 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.260493 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-gsrn6" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.272731 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-b27p9"] Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.282271 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-w6gh2"] Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.284025 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.320773 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-w6gh2"] Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.331944 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-scripts\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.332058 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-var-run-ovn\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.332088 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g82cz\" (UniqueName: \"kubernetes.io/projected/ea76f5cd-b164-49a8-9da2-30702b113bf3-kube-api-access-g82cz\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.332119 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ea76f5cd-b164-49a8-9da2-30702b113bf3-var-run\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.332149 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea76f5cd-b164-49a8-9da2-30702b113bf3-scripts\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.332224 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swxrq\" (UniqueName: \"kubernetes.io/projected/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-kube-api-access-swxrq\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.332261 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-var-run\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.332331 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ea76f5cd-b164-49a8-9da2-30702b113bf3-var-log\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.332385 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ea76f5cd-b164-49a8-9da2-30702b113bf3-var-lib\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.332416 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ea76f5cd-b164-49a8-9da2-30702b113bf3-etc-ovs\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.332437 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-var-log-ovn\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.434628 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ea76f5cd-b164-49a8-9da2-30702b113bf3-var-log\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.434738 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ea76f5cd-b164-49a8-9da2-30702b113bf3-var-lib\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.434781 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ea76f5cd-b164-49a8-9da2-30702b113bf3-etc-ovs\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.434814 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-var-log-ovn\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.434912 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-scripts\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.435015 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g82cz\" (UniqueName: \"kubernetes.io/projected/ea76f5cd-b164-49a8-9da2-30702b113bf3-kube-api-access-g82cz\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.435044 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-var-run-ovn\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.435091 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ea76f5cd-b164-49a8-9da2-30702b113bf3-var-run\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.435136 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea76f5cd-b164-49a8-9da2-30702b113bf3-scripts\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.435179 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swxrq\" (UniqueName: \"kubernetes.io/projected/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-kube-api-access-swxrq\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.435231 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-var-run\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.435736 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-var-run\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.435774 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ea76f5cd-b164-49a8-9da2-30702b113bf3-var-run\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.435855 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-var-run-ovn\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.435872 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/ea76f5cd-b164-49a8-9da2-30702b113bf3-var-lib\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.435910 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/ea76f5cd-b164-49a8-9da2-30702b113bf3-var-log\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.435928 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/ea76f5cd-b164-49a8-9da2-30702b113bf3-etc-ovs\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.435955 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-var-log-ovn\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.438492 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ea76f5cd-b164-49a8-9da2-30702b113bf3-scripts\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.440285 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-scripts\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.455954 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g82cz\" (UniqueName: \"kubernetes.io/projected/ea76f5cd-b164-49a8-9da2-30702b113bf3-kube-api-access-g82cz\") pod \"ovn-controller-ovs-w6gh2\" (UID: \"ea76f5cd-b164-49a8-9da2-30702b113bf3\") " pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.463782 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swxrq\" (UniqueName: \"kubernetes.io/projected/bef5bd3c-dc22-4fe7-b499-e8fb17d17742-kube-api-access-swxrq\") pod \"ovn-controller-b27p9\" (UID: \"bef5bd3c-dc22-4fe7-b499-e8fb17d17742\") " pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.583601 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b27p9" Oct 02 03:18:11 crc kubenswrapper[4775]: I1002 03:18:11.612249 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.189861 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-b27p9"] Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.241823 4775 scope.go:117] "RemoveContainer" containerID="3090386b30abf1216016133f8506b8d20c86c4cce95801155441acf59ebeb663" Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.460791 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-w6gh2"] Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.851319 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-gm78r"] Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.853259 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-gm78r" Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.857735 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.858383 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-gm78r"] Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.889847 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/269c1835-ec87-4102-970c-5073e78cb156-ovn-rundir\") pod \"ovn-controller-metrics-gm78r\" (UID: \"269c1835-ec87-4102-970c-5073e78cb156\") " pod="openstack/ovn-controller-metrics-gm78r" Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.889924 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/269c1835-ec87-4102-970c-5073e78cb156-config\") pod \"ovn-controller-metrics-gm78r\" (UID: \"269c1835-ec87-4102-970c-5073e78cb156\") " pod="openstack/ovn-controller-metrics-gm78r" Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.890070 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/269c1835-ec87-4102-970c-5073e78cb156-ovs-rundir\") pod \"ovn-controller-metrics-gm78r\" (UID: \"269c1835-ec87-4102-970c-5073e78cb156\") " pod="openstack/ovn-controller-metrics-gm78r" Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.890116 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgz4z\" (UniqueName: \"kubernetes.io/projected/269c1835-ec87-4102-970c-5073e78cb156-kube-api-access-qgz4z\") pod \"ovn-controller-metrics-gm78r\" (UID: \"269c1835-ec87-4102-970c-5073e78cb156\") " pod="openstack/ovn-controller-metrics-gm78r" Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.991266 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgz4z\" (UniqueName: \"kubernetes.io/projected/269c1835-ec87-4102-970c-5073e78cb156-kube-api-access-qgz4z\") pod \"ovn-controller-metrics-gm78r\" (UID: \"269c1835-ec87-4102-970c-5073e78cb156\") " pod="openstack/ovn-controller-metrics-gm78r" Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.991679 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/269c1835-ec87-4102-970c-5073e78cb156-ovn-rundir\") pod \"ovn-controller-metrics-gm78r\" (UID: \"269c1835-ec87-4102-970c-5073e78cb156\") " pod="openstack/ovn-controller-metrics-gm78r" Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.991725 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/269c1835-ec87-4102-970c-5073e78cb156-config\") pod \"ovn-controller-metrics-gm78r\" (UID: \"269c1835-ec87-4102-970c-5073e78cb156\") " pod="openstack/ovn-controller-metrics-gm78r" Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.991787 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/269c1835-ec87-4102-970c-5073e78cb156-ovs-rundir\") pod \"ovn-controller-metrics-gm78r\" (UID: \"269c1835-ec87-4102-970c-5073e78cb156\") " pod="openstack/ovn-controller-metrics-gm78r" Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.992004 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/269c1835-ec87-4102-970c-5073e78cb156-ovs-rundir\") pod \"ovn-controller-metrics-gm78r\" (UID: \"269c1835-ec87-4102-970c-5073e78cb156\") " pod="openstack/ovn-controller-metrics-gm78r" Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.992008 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/269c1835-ec87-4102-970c-5073e78cb156-ovn-rundir\") pod \"ovn-controller-metrics-gm78r\" (UID: \"269c1835-ec87-4102-970c-5073e78cb156\") " pod="openstack/ovn-controller-metrics-gm78r" Oct 02 03:18:12 crc kubenswrapper[4775]: I1002 03:18:12.992742 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/269c1835-ec87-4102-970c-5073e78cb156-config\") pod \"ovn-controller-metrics-gm78r\" (UID: \"269c1835-ec87-4102-970c-5073e78cb156\") " pod="openstack/ovn-controller-metrics-gm78r" Oct 02 03:18:13 crc kubenswrapper[4775]: I1002 03:18:13.026112 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgz4z\" (UniqueName: \"kubernetes.io/projected/269c1835-ec87-4102-970c-5073e78cb156-kube-api-access-qgz4z\") pod \"ovn-controller-metrics-gm78r\" (UID: \"269c1835-ec87-4102-970c-5073e78cb156\") " pod="openstack/ovn-controller-metrics-gm78r" Oct 02 03:18:13 crc kubenswrapper[4775]: I1002 03:18:13.051759 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-w6gh2" event={"ID":"ea76f5cd-b164-49a8-9da2-30702b113bf3","Type":"ContainerStarted","Data":"821f135be4569ed1acdac8bc56dc658c08d138f0af572270d238453be9015a9d"} Oct 02 03:18:13 crc kubenswrapper[4775]: I1002 03:18:13.051807 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-w6gh2" event={"ID":"ea76f5cd-b164-49a8-9da2-30702b113bf3","Type":"ContainerStarted","Data":"7150d191aebf08eb9aab77c60941041495e0cb42b27f09f7586b4efdeb03c467"} Oct 02 03:18:13 crc kubenswrapper[4775]: I1002 03:18:13.056815 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b27p9" event={"ID":"bef5bd3c-dc22-4fe7-b499-e8fb17d17742","Type":"ContainerStarted","Data":"a51478bd17def2e696396910b82128f2acda60b6493264c63f36432d4b4ffdb4"} Oct 02 03:18:13 crc kubenswrapper[4775]: I1002 03:18:13.056906 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b27p9" event={"ID":"bef5bd3c-dc22-4fe7-b499-e8fb17d17742","Type":"ContainerStarted","Data":"57887ecdc402f8daab9eaf2609a1434a0f9a91635554c3ec48691e4d5129271b"} Oct 02 03:18:13 crc kubenswrapper[4775]: I1002 03:18:13.057044 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-b27p9" Oct 02 03:18:13 crc kubenswrapper[4775]: I1002 03:18:13.109266 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-b27p9" podStartSLOduration=2.109235283 podStartE2EDuration="2.109235283s" podCreationTimestamp="2025-10-02 03:18:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:18:13.092663544 +0000 UTC m=+5830.259407584" watchObservedRunningTime="2025-10-02 03:18:13.109235283 +0000 UTC m=+5830.275979343" Oct 02 03:18:13 crc kubenswrapper[4775]: I1002 03:18:13.177594 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-gm78r" Oct 02 03:18:13 crc kubenswrapper[4775]: I1002 03:18:13.676556 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-gm78r"] Oct 02 03:18:14 crc kubenswrapper[4775]: I1002 03:18:14.022654 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-fvb7k"] Oct 02 03:18:14 crc kubenswrapper[4775]: I1002 03:18:14.024088 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-fvb7k" Oct 02 03:18:14 crc kubenswrapper[4775]: I1002 03:18:14.040886 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-fvb7k"] Oct 02 03:18:14 crc kubenswrapper[4775]: I1002 03:18:14.075391 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-gm78r" event={"ID":"269c1835-ec87-4102-970c-5073e78cb156","Type":"ContainerStarted","Data":"2a624b49ea4e05a3f76c24f81a67f8055d6747ab2c4c2691d72ecd1ceb665116"} Oct 02 03:18:14 crc kubenswrapper[4775]: I1002 03:18:14.075456 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-gm78r" event={"ID":"269c1835-ec87-4102-970c-5073e78cb156","Type":"ContainerStarted","Data":"9995b4c65acc88320ef70f184406e608abcf7f3200a49c1c779041d68298eda5"} Oct 02 03:18:14 crc kubenswrapper[4775]: I1002 03:18:14.077126 4775 generic.go:334] "Generic (PLEG): container finished" podID="ea76f5cd-b164-49a8-9da2-30702b113bf3" containerID="821f135be4569ed1acdac8bc56dc658c08d138f0af572270d238453be9015a9d" exitCode=0 Oct 02 03:18:14 crc kubenswrapper[4775]: I1002 03:18:14.077219 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-w6gh2" event={"ID":"ea76f5cd-b164-49a8-9da2-30702b113bf3","Type":"ContainerDied","Data":"821f135be4569ed1acdac8bc56dc658c08d138f0af572270d238453be9015a9d"} Oct 02 03:18:14 crc kubenswrapper[4775]: I1002 03:18:14.095362 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-gm78r" podStartSLOduration=2.095343132 podStartE2EDuration="2.095343132s" podCreationTimestamp="2025-10-02 03:18:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:18:14.093541566 +0000 UTC m=+5831.260285606" watchObservedRunningTime="2025-10-02 03:18:14.095343132 +0000 UTC m=+5831.262087172" Oct 02 03:18:14 crc kubenswrapper[4775]: I1002 03:18:14.127136 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47grr\" (UniqueName: \"kubernetes.io/projected/2dc1393d-8d30-4fd9-9e47-a6f5d0df2314-kube-api-access-47grr\") pod \"octavia-db-create-fvb7k\" (UID: \"2dc1393d-8d30-4fd9-9e47-a6f5d0df2314\") " pod="openstack/octavia-db-create-fvb7k" Oct 02 03:18:14 crc kubenswrapper[4775]: I1002 03:18:14.235404 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47grr\" (UniqueName: \"kubernetes.io/projected/2dc1393d-8d30-4fd9-9e47-a6f5d0df2314-kube-api-access-47grr\") pod \"octavia-db-create-fvb7k\" (UID: \"2dc1393d-8d30-4fd9-9e47-a6f5d0df2314\") " pod="openstack/octavia-db-create-fvb7k" Oct 02 03:18:14 crc kubenswrapper[4775]: I1002 03:18:14.261859 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47grr\" (UniqueName: \"kubernetes.io/projected/2dc1393d-8d30-4fd9-9e47-a6f5d0df2314-kube-api-access-47grr\") pod \"octavia-db-create-fvb7k\" (UID: \"2dc1393d-8d30-4fd9-9e47-a6f5d0df2314\") " pod="openstack/octavia-db-create-fvb7k" Oct 02 03:18:14 crc kubenswrapper[4775]: I1002 03:18:14.339691 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-fvb7k" Oct 02 03:18:14 crc kubenswrapper[4775]: I1002 03:18:14.793128 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-fvb7k"] Oct 02 03:18:15 crc kubenswrapper[4775]: I1002 03:18:15.092263 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-w6gh2" event={"ID":"ea76f5cd-b164-49a8-9da2-30702b113bf3","Type":"ContainerStarted","Data":"423a490625d28cd25e1a26167db5512ebc9298066c32891dc7d2db28ae56785b"} Oct 02 03:18:15 crc kubenswrapper[4775]: I1002 03:18:15.092765 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:15 crc kubenswrapper[4775]: I1002 03:18:15.093349 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-w6gh2" event={"ID":"ea76f5cd-b164-49a8-9da2-30702b113bf3","Type":"ContainerStarted","Data":"3ba1211586f3f9704f4540a83a9ea40beb6119401669e0f640aaae067c47f8f4"} Oct 02 03:18:15 crc kubenswrapper[4775]: I1002 03:18:15.093454 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:15 crc kubenswrapper[4775]: I1002 03:18:15.094547 4775 generic.go:334] "Generic (PLEG): container finished" podID="2dc1393d-8d30-4fd9-9e47-a6f5d0df2314" containerID="3cc458720d01fd6ac7affe6c97741c4468afd8751f16cc478f16c98fb8f9ca57" exitCode=0 Oct 02 03:18:15 crc kubenswrapper[4775]: I1002 03:18:15.094605 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-fvb7k" event={"ID":"2dc1393d-8d30-4fd9-9e47-a6f5d0df2314","Type":"ContainerDied","Data":"3cc458720d01fd6ac7affe6c97741c4468afd8751f16cc478f16c98fb8f9ca57"} Oct 02 03:18:15 crc kubenswrapper[4775]: I1002 03:18:15.094675 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-fvb7k" event={"ID":"2dc1393d-8d30-4fd9-9e47-a6f5d0df2314","Type":"ContainerStarted","Data":"4c522417ad5133529d5aafc0cdebc01c9fe783d8ea3a3ed2b580541f96effc1c"} Oct 02 03:18:15 crc kubenswrapper[4775]: I1002 03:18:15.125342 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-w6gh2" podStartSLOduration=4.125312238 podStartE2EDuration="4.125312238s" podCreationTimestamp="2025-10-02 03:18:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:18:15.115420101 +0000 UTC m=+5832.282164151" watchObservedRunningTime="2025-10-02 03:18:15.125312238 +0000 UTC m=+5832.292056278" Oct 02 03:18:16 crc kubenswrapper[4775]: I1002 03:18:16.574981 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-fvb7k" Oct 02 03:18:16 crc kubenswrapper[4775]: I1002 03:18:16.683076 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47grr\" (UniqueName: \"kubernetes.io/projected/2dc1393d-8d30-4fd9-9e47-a6f5d0df2314-kube-api-access-47grr\") pod \"2dc1393d-8d30-4fd9-9e47-a6f5d0df2314\" (UID: \"2dc1393d-8d30-4fd9-9e47-a6f5d0df2314\") " Oct 02 03:18:16 crc kubenswrapper[4775]: I1002 03:18:16.693293 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2dc1393d-8d30-4fd9-9e47-a6f5d0df2314-kube-api-access-47grr" (OuterVolumeSpecName: "kube-api-access-47grr") pod "2dc1393d-8d30-4fd9-9e47-a6f5d0df2314" (UID: "2dc1393d-8d30-4fd9-9e47-a6f5d0df2314"). InnerVolumeSpecName "kube-api-access-47grr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:18:16 crc kubenswrapper[4775]: I1002 03:18:16.786213 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47grr\" (UniqueName: \"kubernetes.io/projected/2dc1393d-8d30-4fd9-9e47-a6f5d0df2314-kube-api-access-47grr\") on node \"crc\" DevicePath \"\"" Oct 02 03:18:17 crc kubenswrapper[4775]: I1002 03:18:17.117731 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-fvb7k" event={"ID":"2dc1393d-8d30-4fd9-9e47-a6f5d0df2314","Type":"ContainerDied","Data":"4c522417ad5133529d5aafc0cdebc01c9fe783d8ea3a3ed2b580541f96effc1c"} Oct 02 03:18:17 crc kubenswrapper[4775]: I1002 03:18:17.118215 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c522417ad5133529d5aafc0cdebc01c9fe783d8ea3a3ed2b580541f96effc1c" Oct 02 03:18:17 crc kubenswrapper[4775]: I1002 03:18:17.117821 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-fvb7k" Oct 02 03:18:21 crc kubenswrapper[4775]: I1002 03:18:21.766270 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:18:21 crc kubenswrapper[4775]: E1002 03:18:21.767054 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:18:26 crc kubenswrapper[4775]: I1002 03:18:26.208169 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-e022-account-create-ggfk2"] Oct 02 03:18:26 crc kubenswrapper[4775]: E1002 03:18:26.209113 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2dc1393d-8d30-4fd9-9e47-a6f5d0df2314" containerName="mariadb-database-create" Oct 02 03:18:26 crc kubenswrapper[4775]: I1002 03:18:26.209135 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2dc1393d-8d30-4fd9-9e47-a6f5d0df2314" containerName="mariadb-database-create" Oct 02 03:18:26 crc kubenswrapper[4775]: I1002 03:18:26.209486 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2dc1393d-8d30-4fd9-9e47-a6f5d0df2314" containerName="mariadb-database-create" Oct 02 03:18:26 crc kubenswrapper[4775]: I1002 03:18:26.210564 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-e022-account-create-ggfk2" Oct 02 03:18:26 crc kubenswrapper[4775]: I1002 03:18:26.217228 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Oct 02 03:18:26 crc kubenswrapper[4775]: I1002 03:18:26.225379 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-e022-account-create-ggfk2"] Oct 02 03:18:26 crc kubenswrapper[4775]: I1002 03:18:26.308631 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5l5n\" (UniqueName: \"kubernetes.io/projected/ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad-kube-api-access-l5l5n\") pod \"octavia-e022-account-create-ggfk2\" (UID: \"ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad\") " pod="openstack/octavia-e022-account-create-ggfk2" Oct 02 03:18:26 crc kubenswrapper[4775]: I1002 03:18:26.410192 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5l5n\" (UniqueName: \"kubernetes.io/projected/ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad-kube-api-access-l5l5n\") pod \"octavia-e022-account-create-ggfk2\" (UID: \"ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad\") " pod="openstack/octavia-e022-account-create-ggfk2" Oct 02 03:18:26 crc kubenswrapper[4775]: I1002 03:18:26.452201 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5l5n\" (UniqueName: \"kubernetes.io/projected/ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad-kube-api-access-l5l5n\") pod \"octavia-e022-account-create-ggfk2\" (UID: \"ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad\") " pod="openstack/octavia-e022-account-create-ggfk2" Oct 02 03:18:26 crc kubenswrapper[4775]: I1002 03:18:26.532793 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-e022-account-create-ggfk2" Oct 02 03:18:27 crc kubenswrapper[4775]: I1002 03:18:27.049029 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-e022-account-create-ggfk2"] Oct 02 03:18:27 crc kubenswrapper[4775]: W1002 03:18:27.052429 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebbd91b0_4bb4_4edc_a7fe_2486fe8899ad.slice/crio-8e014b94be76d49e5aa96c7d0202d91c057b4ff296a29175e0a236ffa0c9f527 WatchSource:0}: Error finding container 8e014b94be76d49e5aa96c7d0202d91c057b4ff296a29175e0a236ffa0c9f527: Status 404 returned error can't find the container with id 8e014b94be76d49e5aa96c7d0202d91c057b4ff296a29175e0a236ffa0c9f527 Oct 02 03:18:27 crc kubenswrapper[4775]: I1002 03:18:27.243176 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-e022-account-create-ggfk2" event={"ID":"ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad","Type":"ContainerStarted","Data":"8e014b94be76d49e5aa96c7d0202d91c057b4ff296a29175e0a236ffa0c9f527"} Oct 02 03:18:28 crc kubenswrapper[4775]: I1002 03:18:28.258391 4775 generic.go:334] "Generic (PLEG): container finished" podID="ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad" containerID="243512ee77002e8bc897149f111abb5d5fa338b9e7fdf3d46fa9b2b4c746504e" exitCode=0 Oct 02 03:18:28 crc kubenswrapper[4775]: I1002 03:18:28.258454 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-e022-account-create-ggfk2" event={"ID":"ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad","Type":"ContainerDied","Data":"243512ee77002e8bc897149f111abb5d5fa338b9e7fdf3d46fa9b2b4c746504e"} Oct 02 03:18:29 crc kubenswrapper[4775]: I1002 03:18:29.624352 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-e022-account-create-ggfk2" Oct 02 03:18:29 crc kubenswrapper[4775]: I1002 03:18:29.696217 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5l5n\" (UniqueName: \"kubernetes.io/projected/ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad-kube-api-access-l5l5n\") pod \"ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad\" (UID: \"ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad\") " Oct 02 03:18:29 crc kubenswrapper[4775]: I1002 03:18:29.706701 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad-kube-api-access-l5l5n" (OuterVolumeSpecName: "kube-api-access-l5l5n") pod "ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad" (UID: "ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad"). InnerVolumeSpecName "kube-api-access-l5l5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:18:29 crc kubenswrapper[4775]: I1002 03:18:29.799065 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5l5n\" (UniqueName: \"kubernetes.io/projected/ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad-kube-api-access-l5l5n\") on node \"crc\" DevicePath \"\"" Oct 02 03:18:30 crc kubenswrapper[4775]: I1002 03:18:30.278277 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-e022-account-create-ggfk2" event={"ID":"ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad","Type":"ContainerDied","Data":"8e014b94be76d49e5aa96c7d0202d91c057b4ff296a29175e0a236ffa0c9f527"} Oct 02 03:18:30 crc kubenswrapper[4775]: I1002 03:18:30.278546 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e014b94be76d49e5aa96c7d0202d91c057b4ff296a29175e0a236ffa0c9f527" Oct 02 03:18:30 crc kubenswrapper[4775]: I1002 03:18:30.278407 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-e022-account-create-ggfk2" Oct 02 03:18:33 crc kubenswrapper[4775]: I1002 03:18:33.189376 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-hkplh"] Oct 02 03:18:33 crc kubenswrapper[4775]: E1002 03:18:33.190484 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad" containerName="mariadb-account-create" Oct 02 03:18:33 crc kubenswrapper[4775]: I1002 03:18:33.190508 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad" containerName="mariadb-account-create" Oct 02 03:18:33 crc kubenswrapper[4775]: I1002 03:18:33.191020 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad" containerName="mariadb-account-create" Oct 02 03:18:33 crc kubenswrapper[4775]: I1002 03:18:33.191933 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-hkplh" Oct 02 03:18:33 crc kubenswrapper[4775]: I1002 03:18:33.206400 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-hkplh"] Oct 02 03:18:33 crc kubenswrapper[4775]: I1002 03:18:33.265107 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5l5lb\" (UniqueName: \"kubernetes.io/projected/c9f570ab-6a73-47ee-8b8b-64c3f78a9d34-kube-api-access-5l5lb\") pod \"octavia-persistence-db-create-hkplh\" (UID: \"c9f570ab-6a73-47ee-8b8b-64c3f78a9d34\") " pod="openstack/octavia-persistence-db-create-hkplh" Oct 02 03:18:33 crc kubenswrapper[4775]: I1002 03:18:33.366893 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5l5lb\" (UniqueName: \"kubernetes.io/projected/c9f570ab-6a73-47ee-8b8b-64c3f78a9d34-kube-api-access-5l5lb\") pod \"octavia-persistence-db-create-hkplh\" (UID: \"c9f570ab-6a73-47ee-8b8b-64c3f78a9d34\") " pod="openstack/octavia-persistence-db-create-hkplh" Oct 02 03:18:33 crc kubenswrapper[4775]: I1002 03:18:33.388267 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5l5lb\" (UniqueName: \"kubernetes.io/projected/c9f570ab-6a73-47ee-8b8b-64c3f78a9d34-kube-api-access-5l5lb\") pod \"octavia-persistence-db-create-hkplh\" (UID: \"c9f570ab-6a73-47ee-8b8b-64c3f78a9d34\") " pod="openstack/octavia-persistence-db-create-hkplh" Oct 02 03:18:33 crc kubenswrapper[4775]: I1002 03:18:33.517324 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-hkplh" Oct 02 03:18:34 crc kubenswrapper[4775]: I1002 03:18:34.042978 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-hkplh"] Oct 02 03:18:34 crc kubenswrapper[4775]: I1002 03:18:34.329862 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-hkplh" event={"ID":"c9f570ab-6a73-47ee-8b8b-64c3f78a9d34","Type":"ContainerStarted","Data":"5dada035b80ae1e48617d87704761e1c28e4e47be96334ff0dc040c835851d7d"} Oct 02 03:18:34 crc kubenswrapper[4775]: I1002 03:18:34.329923 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-hkplh" event={"ID":"c9f570ab-6a73-47ee-8b8b-64c3f78a9d34","Type":"ContainerStarted","Data":"91bfcbac3a75ef76b518aa711d6de77b858c4ea058a6dccafbd689e6c253e58b"} Oct 02 03:18:34 crc kubenswrapper[4775]: I1002 03:18:34.364556 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-persistence-db-create-hkplh" podStartSLOduration=1.364525706 podStartE2EDuration="1.364525706s" podCreationTimestamp="2025-10-02 03:18:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:18:34.348152922 +0000 UTC m=+5851.514897002" watchObservedRunningTime="2025-10-02 03:18:34.364525706 +0000 UTC m=+5851.531269786" Oct 02 03:18:34 crc kubenswrapper[4775]: I1002 03:18:34.765434 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:18:34 crc kubenswrapper[4775]: E1002 03:18:34.766019 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:18:35 crc kubenswrapper[4775]: I1002 03:18:35.340599 4775 generic.go:334] "Generic (PLEG): container finished" podID="c9f570ab-6a73-47ee-8b8b-64c3f78a9d34" containerID="5dada035b80ae1e48617d87704761e1c28e4e47be96334ff0dc040c835851d7d" exitCode=0 Oct 02 03:18:35 crc kubenswrapper[4775]: I1002 03:18:35.341440 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-hkplh" event={"ID":"c9f570ab-6a73-47ee-8b8b-64c3f78a9d34","Type":"ContainerDied","Data":"5dada035b80ae1e48617d87704761e1c28e4e47be96334ff0dc040c835851d7d"} Oct 02 03:18:36 crc kubenswrapper[4775]: I1002 03:18:36.868137 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-hkplh" Oct 02 03:18:36 crc kubenswrapper[4775]: I1002 03:18:36.963102 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5l5lb\" (UniqueName: \"kubernetes.io/projected/c9f570ab-6a73-47ee-8b8b-64c3f78a9d34-kube-api-access-5l5lb\") pod \"c9f570ab-6a73-47ee-8b8b-64c3f78a9d34\" (UID: \"c9f570ab-6a73-47ee-8b8b-64c3f78a9d34\") " Oct 02 03:18:36 crc kubenswrapper[4775]: I1002 03:18:36.969575 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9f570ab-6a73-47ee-8b8b-64c3f78a9d34-kube-api-access-5l5lb" (OuterVolumeSpecName: "kube-api-access-5l5lb") pod "c9f570ab-6a73-47ee-8b8b-64c3f78a9d34" (UID: "c9f570ab-6a73-47ee-8b8b-64c3f78a9d34"). InnerVolumeSpecName "kube-api-access-5l5lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:18:37 crc kubenswrapper[4775]: I1002 03:18:37.065581 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5l5lb\" (UniqueName: \"kubernetes.io/projected/c9f570ab-6a73-47ee-8b8b-64c3f78a9d34-kube-api-access-5l5lb\") on node \"crc\" DevicePath \"\"" Oct 02 03:18:37 crc kubenswrapper[4775]: I1002 03:18:37.370696 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-hkplh" event={"ID":"c9f570ab-6a73-47ee-8b8b-64c3f78a9d34","Type":"ContainerDied","Data":"91bfcbac3a75ef76b518aa711d6de77b858c4ea058a6dccafbd689e6c253e58b"} Oct 02 03:18:37 crc kubenswrapper[4775]: I1002 03:18:37.370734 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91bfcbac3a75ef76b518aa711d6de77b858c4ea058a6dccafbd689e6c253e58b" Oct 02 03:18:37 crc kubenswrapper[4775]: I1002 03:18:37.370763 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-hkplh" Oct 02 03:18:44 crc kubenswrapper[4775]: I1002 03:18:44.456642 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-69a7-account-create-5hjm6"] Oct 02 03:18:44 crc kubenswrapper[4775]: E1002 03:18:44.460005 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9f570ab-6a73-47ee-8b8b-64c3f78a9d34" containerName="mariadb-database-create" Oct 02 03:18:44 crc kubenswrapper[4775]: I1002 03:18:44.460262 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9f570ab-6a73-47ee-8b8b-64c3f78a9d34" containerName="mariadb-database-create" Oct 02 03:18:44 crc kubenswrapper[4775]: I1002 03:18:44.460947 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9f570ab-6a73-47ee-8b8b-64c3f78a9d34" containerName="mariadb-database-create" Oct 02 03:18:44 crc kubenswrapper[4775]: I1002 03:18:44.463636 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-69a7-account-create-5hjm6" Oct 02 03:18:44 crc kubenswrapper[4775]: I1002 03:18:44.470438 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Oct 02 03:18:44 crc kubenswrapper[4775]: I1002 03:18:44.470511 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-69a7-account-create-5hjm6"] Oct 02 03:18:44 crc kubenswrapper[4775]: I1002 03:18:44.540403 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hz4j\" (UniqueName: \"kubernetes.io/projected/c212a5c8-90ef-45dc-9da1-762c24eafd75-kube-api-access-6hz4j\") pod \"octavia-69a7-account-create-5hjm6\" (UID: \"c212a5c8-90ef-45dc-9da1-762c24eafd75\") " pod="openstack/octavia-69a7-account-create-5hjm6" Oct 02 03:18:44 crc kubenswrapper[4775]: I1002 03:18:44.642747 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hz4j\" (UniqueName: \"kubernetes.io/projected/c212a5c8-90ef-45dc-9da1-762c24eafd75-kube-api-access-6hz4j\") pod \"octavia-69a7-account-create-5hjm6\" (UID: \"c212a5c8-90ef-45dc-9da1-762c24eafd75\") " pod="openstack/octavia-69a7-account-create-5hjm6" Oct 02 03:18:44 crc kubenswrapper[4775]: I1002 03:18:44.681317 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hz4j\" (UniqueName: \"kubernetes.io/projected/c212a5c8-90ef-45dc-9da1-762c24eafd75-kube-api-access-6hz4j\") pod \"octavia-69a7-account-create-5hjm6\" (UID: \"c212a5c8-90ef-45dc-9da1-762c24eafd75\") " pod="openstack/octavia-69a7-account-create-5hjm6" Oct 02 03:18:44 crc kubenswrapper[4775]: I1002 03:18:44.793182 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-69a7-account-create-5hjm6" Oct 02 03:18:45 crc kubenswrapper[4775]: I1002 03:18:45.331927 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-69a7-account-create-5hjm6"] Oct 02 03:18:45 crc kubenswrapper[4775]: I1002 03:18:45.483113 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-69a7-account-create-5hjm6" event={"ID":"c212a5c8-90ef-45dc-9da1-762c24eafd75","Type":"ContainerStarted","Data":"fe9e9b913ed9f1156121851fb1d806a2ffd44c084fbcadd942286e2a2659b3cb"} Oct 02 03:18:45 crc kubenswrapper[4775]: I1002 03:18:45.765551 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:18:45 crc kubenswrapper[4775]: E1002 03:18:45.766027 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.497441 4775 generic.go:334] "Generic (PLEG): container finished" podID="c212a5c8-90ef-45dc-9da1-762c24eafd75" containerID="8a22eacafcc1c5d4ea5bcf410fcf7ad2bc798f3cfabcf2fe8fdaaa7eeafd07ff" exitCode=0 Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.497534 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-69a7-account-create-5hjm6" event={"ID":"c212a5c8-90ef-45dc-9da1-762c24eafd75","Type":"ContainerDied","Data":"8a22eacafcc1c5d4ea5bcf410fcf7ad2bc798f3cfabcf2fe8fdaaa7eeafd07ff"} Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.644659 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-b27p9" podUID="bef5bd3c-dc22-4fe7-b499-e8fb17d17742" containerName="ovn-controller" probeResult="failure" output=< Oct 02 03:18:46 crc kubenswrapper[4775]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 02 03:18:46 crc kubenswrapper[4775]: > Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.690425 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.693585 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-w6gh2" Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.855006 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-b27p9-config-xvnwj"] Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.858616 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.872936 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.880576 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-b27p9-config-xvnwj"] Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.898720 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7e3a1d6b-4851-4a89-a356-c30c609bde4c-additional-scripts\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.898794 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-run-ovn\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.899530 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-run\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.899716 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e3a1d6b-4851-4a89-a356-c30c609bde4c-scripts\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.900163 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-log-ovn\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:46 crc kubenswrapper[4775]: I1002 03:18:46.900356 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4hdh\" (UniqueName: \"kubernetes.io/projected/7e3a1d6b-4851-4a89-a356-c30c609bde4c-kube-api-access-x4hdh\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.001995 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e3a1d6b-4851-4a89-a356-c30c609bde4c-scripts\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.002144 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-log-ovn\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.002212 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4hdh\" (UniqueName: \"kubernetes.io/projected/7e3a1d6b-4851-4a89-a356-c30c609bde4c-kube-api-access-x4hdh\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.002286 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7e3a1d6b-4851-4a89-a356-c30c609bde4c-additional-scripts\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.002315 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-run-ovn\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.002366 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-run\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.002750 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-run-ovn\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.002768 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-run\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.002806 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-log-ovn\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.003745 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7e3a1d6b-4851-4a89-a356-c30c609bde4c-additional-scripts\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.006182 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e3a1d6b-4851-4a89-a356-c30c609bde4c-scripts\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.025097 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4hdh\" (UniqueName: \"kubernetes.io/projected/7e3a1d6b-4851-4a89-a356-c30c609bde4c-kube-api-access-x4hdh\") pod \"ovn-controller-b27p9-config-xvnwj\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.204544 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.664018 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-b27p9-config-xvnwj"] Oct 02 03:18:47 crc kubenswrapper[4775]: W1002 03:18:47.666539 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e3a1d6b_4851_4a89_a356_c30c609bde4c.slice/crio-f81c7052eb925a46f06f0ab38a6c5b7f19317d659fb4340c34b62cf5174369d6 WatchSource:0}: Error finding container f81c7052eb925a46f06f0ab38a6c5b7f19317d659fb4340c34b62cf5174369d6: Status 404 returned error can't find the container with id f81c7052eb925a46f06f0ab38a6c5b7f19317d659fb4340c34b62cf5174369d6 Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.832847 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-69a7-account-create-5hjm6" Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.920907 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hz4j\" (UniqueName: \"kubernetes.io/projected/c212a5c8-90ef-45dc-9da1-762c24eafd75-kube-api-access-6hz4j\") pod \"c212a5c8-90ef-45dc-9da1-762c24eafd75\" (UID: \"c212a5c8-90ef-45dc-9da1-762c24eafd75\") " Oct 02 03:18:47 crc kubenswrapper[4775]: I1002 03:18:47.926144 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c212a5c8-90ef-45dc-9da1-762c24eafd75-kube-api-access-6hz4j" (OuterVolumeSpecName: "kube-api-access-6hz4j") pod "c212a5c8-90ef-45dc-9da1-762c24eafd75" (UID: "c212a5c8-90ef-45dc-9da1-762c24eafd75"). InnerVolumeSpecName "kube-api-access-6hz4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:18:48 crc kubenswrapper[4775]: I1002 03:18:48.025770 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hz4j\" (UniqueName: \"kubernetes.io/projected/c212a5c8-90ef-45dc-9da1-762c24eafd75-kube-api-access-6hz4j\") on node \"crc\" DevicePath \"\"" Oct 02 03:18:48 crc kubenswrapper[4775]: I1002 03:18:48.522482 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-69a7-account-create-5hjm6" Oct 02 03:18:48 crc kubenswrapper[4775]: I1002 03:18:48.522481 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-69a7-account-create-5hjm6" event={"ID":"c212a5c8-90ef-45dc-9da1-762c24eafd75","Type":"ContainerDied","Data":"fe9e9b913ed9f1156121851fb1d806a2ffd44c084fbcadd942286e2a2659b3cb"} Oct 02 03:18:48 crc kubenswrapper[4775]: I1002 03:18:48.523013 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe9e9b913ed9f1156121851fb1d806a2ffd44c084fbcadd942286e2a2659b3cb" Oct 02 03:18:48 crc kubenswrapper[4775]: I1002 03:18:48.526473 4775 generic.go:334] "Generic (PLEG): container finished" podID="7e3a1d6b-4851-4a89-a356-c30c609bde4c" containerID="bb9d8d149247f5f8edf1694299e0391c3bcfa5804378d2ea7f3f39e6692b3aba" exitCode=0 Oct 02 03:18:48 crc kubenswrapper[4775]: I1002 03:18:48.526511 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b27p9-config-xvnwj" event={"ID":"7e3a1d6b-4851-4a89-a356-c30c609bde4c","Type":"ContainerDied","Data":"bb9d8d149247f5f8edf1694299e0391c3bcfa5804378d2ea7f3f39e6692b3aba"} Oct 02 03:18:48 crc kubenswrapper[4775]: I1002 03:18:48.526534 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b27p9-config-xvnwj" event={"ID":"7e3a1d6b-4851-4a89-a356-c30c609bde4c","Type":"ContainerStarted","Data":"f81c7052eb925a46f06f0ab38a6c5b7f19317d659fb4340c34b62cf5174369d6"} Oct 02 03:18:49 crc kubenswrapper[4775]: I1002 03:18:49.982776 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.068930 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e3a1d6b-4851-4a89-a356-c30c609bde4c-scripts\") pod \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.069020 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-log-ovn\") pod \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.069088 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4hdh\" (UniqueName: \"kubernetes.io/projected/7e3a1d6b-4851-4a89-a356-c30c609bde4c-kube-api-access-x4hdh\") pod \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.069189 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "7e3a1d6b-4851-4a89-a356-c30c609bde4c" (UID: "7e3a1d6b-4851-4a89-a356-c30c609bde4c"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.069251 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "7e3a1d6b-4851-4a89-a356-c30c609bde4c" (UID: "7e3a1d6b-4851-4a89-a356-c30c609bde4c"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.069215 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-run-ovn\") pod \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.069306 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-run\") pod \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.069369 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7e3a1d6b-4851-4a89-a356-c30c609bde4c-additional-scripts\") pod \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\" (UID: \"7e3a1d6b-4851-4a89-a356-c30c609bde4c\") " Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.069488 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-run" (OuterVolumeSpecName: "var-run") pod "7e3a1d6b-4851-4a89-a356-c30c609bde4c" (UID: "7e3a1d6b-4851-4a89-a356-c30c609bde4c"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.069813 4775 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.069842 4775 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.069855 4775 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/7e3a1d6b-4851-4a89-a356-c30c609bde4c-var-run\") on node \"crc\" DevicePath \"\"" Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.071096 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e3a1d6b-4851-4a89-a356-c30c609bde4c-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "7e3a1d6b-4851-4a89-a356-c30c609bde4c" (UID: "7e3a1d6b-4851-4a89-a356-c30c609bde4c"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.071682 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e3a1d6b-4851-4a89-a356-c30c609bde4c-scripts" (OuterVolumeSpecName: "scripts") pod "7e3a1d6b-4851-4a89-a356-c30c609bde4c" (UID: "7e3a1d6b-4851-4a89-a356-c30c609bde4c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.077074 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e3a1d6b-4851-4a89-a356-c30c609bde4c-kube-api-access-x4hdh" (OuterVolumeSpecName: "kube-api-access-x4hdh") pod "7e3a1d6b-4851-4a89-a356-c30c609bde4c" (UID: "7e3a1d6b-4851-4a89-a356-c30c609bde4c"). InnerVolumeSpecName "kube-api-access-x4hdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.171898 4775 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/7e3a1d6b-4851-4a89-a356-c30c609bde4c-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.171983 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e3a1d6b-4851-4a89-a356-c30c609bde4c-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.172000 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4hdh\" (UniqueName: \"kubernetes.io/projected/7e3a1d6b-4851-4a89-a356-c30c609bde4c-kube-api-access-x4hdh\") on node \"crc\" DevicePath \"\"" Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.558873 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b27p9-config-xvnwj" event={"ID":"7e3a1d6b-4851-4a89-a356-c30c609bde4c","Type":"ContainerDied","Data":"f81c7052eb925a46f06f0ab38a6c5b7f19317d659fb4340c34b62cf5174369d6"} Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.559285 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f81c7052eb925a46f06f0ab38a6c5b7f19317d659fb4340c34b62cf5174369d6" Oct 02 03:18:50 crc kubenswrapper[4775]: I1002 03:18:50.559028 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b27p9-config-xvnwj" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.102717 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-b27p9-config-xvnwj"] Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.120730 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-b27p9-config-xvnwj"] Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.372342 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-9669b56-5jpsb"] Oct 02 03:18:51 crc kubenswrapper[4775]: E1002 03:18:51.372811 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c212a5c8-90ef-45dc-9da1-762c24eafd75" containerName="mariadb-account-create" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.372893 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c212a5c8-90ef-45dc-9da1-762c24eafd75" containerName="mariadb-account-create" Oct 02 03:18:51 crc kubenswrapper[4775]: E1002 03:18:51.372974 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e3a1d6b-4851-4a89-a356-c30c609bde4c" containerName="ovn-config" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.373038 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e3a1d6b-4851-4a89-a356-c30c609bde4c" containerName="ovn-config" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.373280 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e3a1d6b-4851-4a89-a356-c30c609bde4c" containerName="ovn-config" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.373352 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c212a5c8-90ef-45dc-9da1-762c24eafd75" containerName="mariadb-account-create" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.374721 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.383135 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.383919 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-h2njb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.384144 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.390435 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-9669b56-5jpsb"] Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.505663 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e534bc54-984a-4661-9e75-f898b8c78d33-config-data-merged\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.505713 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e534bc54-984a-4661-9e75-f898b8c78d33-combined-ca-bundle\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.505854 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e534bc54-984a-4661-9e75-f898b8c78d33-scripts\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.505880 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/e534bc54-984a-4661-9e75-f898b8c78d33-octavia-run\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.505908 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e534bc54-984a-4661-9e75-f898b8c78d33-config-data\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.607608 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e534bc54-984a-4661-9e75-f898b8c78d33-scripts\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.607657 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/e534bc54-984a-4661-9e75-f898b8c78d33-octavia-run\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.607687 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e534bc54-984a-4661-9e75-f898b8c78d33-config-data\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.607731 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e534bc54-984a-4661-9e75-f898b8c78d33-config-data-merged\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.607754 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e534bc54-984a-4661-9e75-f898b8c78d33-combined-ca-bundle\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.608777 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e534bc54-984a-4661-9e75-f898b8c78d33-config-data-merged\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.608846 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/e534bc54-984a-4661-9e75-f898b8c78d33-octavia-run\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.618738 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e534bc54-984a-4661-9e75-f898b8c78d33-scripts\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.618751 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e534bc54-984a-4661-9e75-f898b8c78d33-combined-ca-bundle\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.618873 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e534bc54-984a-4661-9e75-f898b8c78d33-config-data\") pod \"octavia-api-9669b56-5jpsb\" (UID: \"e534bc54-984a-4661-9e75-f898b8c78d33\") " pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.676339 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-b27p9" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.694554 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:18:51 crc kubenswrapper[4775]: I1002 03:18:51.774850 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e3a1d6b-4851-4a89-a356-c30c609bde4c" path="/var/lib/kubelet/pods/7e3a1d6b-4851-4a89-a356-c30c609bde4c/volumes" Oct 02 03:18:52 crc kubenswrapper[4775]: I1002 03:18:52.248622 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-9669b56-5jpsb"] Oct 02 03:18:52 crc kubenswrapper[4775]: I1002 03:18:52.575307 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-9669b56-5jpsb" event={"ID":"e534bc54-984a-4661-9e75-f898b8c78d33","Type":"ContainerStarted","Data":"02996d4f156136847f4f9e5d1cb58d03eb62b37717f97422fdd0f71b177a4ae3"} Oct 02 03:18:58 crc kubenswrapper[4775]: I1002 03:18:58.765487 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:18:58 crc kubenswrapper[4775]: E1002 03:18:58.766210 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:19:01 crc kubenswrapper[4775]: I1002 03:19:01.666179 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-9669b56-5jpsb" event={"ID":"e534bc54-984a-4661-9e75-f898b8c78d33","Type":"ContainerStarted","Data":"9287c3c6919c1fe1e88d760f79381f88a7bfb24226394a0191c682a51236a0f7"} Oct 02 03:19:02 crc kubenswrapper[4775]: I1002 03:19:02.675898 4775 generic.go:334] "Generic (PLEG): container finished" podID="e534bc54-984a-4661-9e75-f898b8c78d33" containerID="9287c3c6919c1fe1e88d760f79381f88a7bfb24226394a0191c682a51236a0f7" exitCode=0 Oct 02 03:19:02 crc kubenswrapper[4775]: I1002 03:19:02.675947 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-9669b56-5jpsb" event={"ID":"e534bc54-984a-4661-9e75-f898b8c78d33","Type":"ContainerDied","Data":"9287c3c6919c1fe1e88d760f79381f88a7bfb24226394a0191c682a51236a0f7"} Oct 02 03:19:03 crc kubenswrapper[4775]: I1002 03:19:03.693720 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-9669b56-5jpsb" event={"ID":"e534bc54-984a-4661-9e75-f898b8c78d33","Type":"ContainerStarted","Data":"3beb8a30f226e76782f5502111a6e460c858367c6a1bdec33571bf27fea7495d"} Oct 02 03:19:03 crc kubenswrapper[4775]: I1002 03:19:03.693994 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-9669b56-5jpsb" event={"ID":"e534bc54-984a-4661-9e75-f898b8c78d33","Type":"ContainerStarted","Data":"ccf802975d6380f06940975af408308c5c7261b309f90b99a5d901e336fbd1b8"} Oct 02 03:19:03 crc kubenswrapper[4775]: I1002 03:19:03.695072 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:19:03 crc kubenswrapper[4775]: I1002 03:19:03.695100 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:19:03 crc kubenswrapper[4775]: I1002 03:19:03.738333 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-9669b56-5jpsb" podStartSLOduration=3.584565306 podStartE2EDuration="12.738314976s" podCreationTimestamp="2025-10-02 03:18:51 +0000 UTC" firstStartedPulling="2025-10-02 03:18:52.250735921 +0000 UTC m=+5869.417479961" lastFinishedPulling="2025-10-02 03:19:01.404485551 +0000 UTC m=+5878.571229631" observedRunningTime="2025-10-02 03:19:03.719864398 +0000 UTC m=+5880.886608438" watchObservedRunningTime="2025-10-02 03:19:03.738314976 +0000 UTC m=+5880.905059016" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.013257 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-9l9kr"] Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.015599 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-9l9kr"] Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.015689 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.042028 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.042062 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.042028 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.058062 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d1cf6eca-9072-43a2-8827-d2e0b90134b5-hm-ports\") pod \"octavia-rsyslog-9l9kr\" (UID: \"d1cf6eca-9072-43a2-8827-d2e0b90134b5\") " pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.058328 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d1cf6eca-9072-43a2-8827-d2e0b90134b5-config-data-merged\") pod \"octavia-rsyslog-9l9kr\" (UID: \"d1cf6eca-9072-43a2-8827-d2e0b90134b5\") " pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.058369 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1cf6eca-9072-43a2-8827-d2e0b90134b5-config-data\") pod \"octavia-rsyslog-9l9kr\" (UID: \"d1cf6eca-9072-43a2-8827-d2e0b90134b5\") " pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.058438 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1cf6eca-9072-43a2-8827-d2e0b90134b5-scripts\") pod \"octavia-rsyslog-9l9kr\" (UID: \"d1cf6eca-9072-43a2-8827-d2e0b90134b5\") " pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.159923 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1cf6eca-9072-43a2-8827-d2e0b90134b5-scripts\") pod \"octavia-rsyslog-9l9kr\" (UID: \"d1cf6eca-9072-43a2-8827-d2e0b90134b5\") " pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.160130 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d1cf6eca-9072-43a2-8827-d2e0b90134b5-hm-ports\") pod \"octavia-rsyslog-9l9kr\" (UID: \"d1cf6eca-9072-43a2-8827-d2e0b90134b5\") " pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.160158 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d1cf6eca-9072-43a2-8827-d2e0b90134b5-config-data-merged\") pod \"octavia-rsyslog-9l9kr\" (UID: \"d1cf6eca-9072-43a2-8827-d2e0b90134b5\") " pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.160199 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1cf6eca-9072-43a2-8827-d2e0b90134b5-config-data\") pod \"octavia-rsyslog-9l9kr\" (UID: \"d1cf6eca-9072-43a2-8827-d2e0b90134b5\") " pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.160611 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/d1cf6eca-9072-43a2-8827-d2e0b90134b5-config-data-merged\") pod \"octavia-rsyslog-9l9kr\" (UID: \"d1cf6eca-9072-43a2-8827-d2e0b90134b5\") " pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.161576 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/d1cf6eca-9072-43a2-8827-d2e0b90134b5-hm-ports\") pod \"octavia-rsyslog-9l9kr\" (UID: \"d1cf6eca-9072-43a2-8827-d2e0b90134b5\") " pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.165716 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1cf6eca-9072-43a2-8827-d2e0b90134b5-scripts\") pod \"octavia-rsyslog-9l9kr\" (UID: \"d1cf6eca-9072-43a2-8827-d2e0b90134b5\") " pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.176345 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1cf6eca-9072-43a2-8827-d2e0b90134b5-config-data\") pod \"octavia-rsyslog-9l9kr\" (UID: \"d1cf6eca-9072-43a2-8827-d2e0b90134b5\") " pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.356709 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.935151 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-9l9kr"] Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.997273 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-6hrw2"] Oct 02 03:19:09 crc kubenswrapper[4775]: I1002 03:19:09.999076 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-6hrw2" Oct 02 03:19:10 crc kubenswrapper[4775]: I1002 03:19:10.003095 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Oct 02 03:19:10 crc kubenswrapper[4775]: I1002 03:19:10.009864 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-6hrw2"] Oct 02 03:19:10 crc kubenswrapper[4775]: I1002 03:19:10.077177 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/a500c203-7f7e-4aee-96f8-1a4b579b446f-amphora-image\") pod \"octavia-image-upload-59f8cff499-6hrw2\" (UID: \"a500c203-7f7e-4aee-96f8-1a4b579b446f\") " pod="openstack/octavia-image-upload-59f8cff499-6hrw2" Oct 02 03:19:10 crc kubenswrapper[4775]: I1002 03:19:10.077462 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a500c203-7f7e-4aee-96f8-1a4b579b446f-httpd-config\") pod \"octavia-image-upload-59f8cff499-6hrw2\" (UID: \"a500c203-7f7e-4aee-96f8-1a4b579b446f\") " pod="openstack/octavia-image-upload-59f8cff499-6hrw2" Oct 02 03:19:10 crc kubenswrapper[4775]: I1002 03:19:10.179016 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a500c203-7f7e-4aee-96f8-1a4b579b446f-httpd-config\") pod \"octavia-image-upload-59f8cff499-6hrw2\" (UID: \"a500c203-7f7e-4aee-96f8-1a4b579b446f\") " pod="openstack/octavia-image-upload-59f8cff499-6hrw2" Oct 02 03:19:10 crc kubenswrapper[4775]: I1002 03:19:10.179081 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/a500c203-7f7e-4aee-96f8-1a4b579b446f-amphora-image\") pod \"octavia-image-upload-59f8cff499-6hrw2\" (UID: \"a500c203-7f7e-4aee-96f8-1a4b579b446f\") " pod="openstack/octavia-image-upload-59f8cff499-6hrw2" Oct 02 03:19:10 crc kubenswrapper[4775]: I1002 03:19:10.179620 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/a500c203-7f7e-4aee-96f8-1a4b579b446f-amphora-image\") pod \"octavia-image-upload-59f8cff499-6hrw2\" (UID: \"a500c203-7f7e-4aee-96f8-1a4b579b446f\") " pod="openstack/octavia-image-upload-59f8cff499-6hrw2" Oct 02 03:19:10 crc kubenswrapper[4775]: I1002 03:19:10.184625 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a500c203-7f7e-4aee-96f8-1a4b579b446f-httpd-config\") pod \"octavia-image-upload-59f8cff499-6hrw2\" (UID: \"a500c203-7f7e-4aee-96f8-1a4b579b446f\") " pod="openstack/octavia-image-upload-59f8cff499-6hrw2" Oct 02 03:19:10 crc kubenswrapper[4775]: I1002 03:19:10.322400 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-6hrw2" Oct 02 03:19:10 crc kubenswrapper[4775]: I1002 03:19:10.742794 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-6hrw2"] Oct 02 03:19:10 crc kubenswrapper[4775]: I1002 03:19:10.766382 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:19:10 crc kubenswrapper[4775]: E1002 03:19:10.766568 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:19:10 crc kubenswrapper[4775]: I1002 03:19:10.779909 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-9l9kr" event={"ID":"d1cf6eca-9072-43a2-8827-d2e0b90134b5","Type":"ContainerStarted","Data":"b01a324e484f415b0a7f98101fd131e49d164c1f30eaca5792ac3b3b618cf45a"} Oct 02 03:19:10 crc kubenswrapper[4775]: I1002 03:19:10.781470 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-6hrw2" event={"ID":"a500c203-7f7e-4aee-96f8-1a4b579b446f","Type":"ContainerStarted","Data":"ee289759a80e4b07d71a8956c8471e317261e2582bcd5aec281c53c2dbfe7bb8"} Oct 02 03:19:11 crc kubenswrapper[4775]: I1002 03:19:11.790863 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-9l9kr" event={"ID":"d1cf6eca-9072-43a2-8827-d2e0b90134b5","Type":"ContainerStarted","Data":"1b539df55c12384d80eed4f4b5d3e6d00de12a709ceff3ccc28b3cfa72fa970b"} Oct 02 03:19:13 crc kubenswrapper[4775]: I1002 03:19:13.811523 4775 generic.go:334] "Generic (PLEG): container finished" podID="d1cf6eca-9072-43a2-8827-d2e0b90134b5" containerID="1b539df55c12384d80eed4f4b5d3e6d00de12a709ceff3ccc28b3cfa72fa970b" exitCode=0 Oct 02 03:19:13 crc kubenswrapper[4775]: I1002 03:19:13.811616 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-9l9kr" event={"ID":"d1cf6eca-9072-43a2-8827-d2e0b90134b5","Type":"ContainerDied","Data":"1b539df55c12384d80eed4f4b5d3e6d00de12a709ceff3ccc28b3cfa72fa970b"} Oct 02 03:19:15 crc kubenswrapper[4775]: I1002 03:19:15.835699 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-9l9kr" event={"ID":"d1cf6eca-9072-43a2-8827-d2e0b90134b5","Type":"ContainerStarted","Data":"37b26036ecabded630c71caf7bf64546d4f334228d3d2e277c74c4da9f5afc47"} Oct 02 03:19:15 crc kubenswrapper[4775]: I1002 03:19:15.836150 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:15 crc kubenswrapper[4775]: I1002 03:19:15.854184 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-9l9kr" podStartSLOduration=2.824733266 podStartE2EDuration="7.854165581s" podCreationTimestamp="2025-10-02 03:19:08 +0000 UTC" firstStartedPulling="2025-10-02 03:19:09.939336796 +0000 UTC m=+5887.106080836" lastFinishedPulling="2025-10-02 03:19:14.968769111 +0000 UTC m=+5892.135513151" observedRunningTime="2025-10-02 03:19:15.852358074 +0000 UTC m=+5893.019102134" watchObservedRunningTime="2025-10-02 03:19:15.854165581 +0000 UTC m=+5893.020909621" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.132255 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-5vw8g"] Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.134142 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.136913 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.146737 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-5vw8g"] Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.211887 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-combined-ca-bundle\") pod \"octavia-db-sync-5vw8g\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.212236 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-config-data\") pod \"octavia-db-sync-5vw8g\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.212590 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-scripts\") pod \"octavia-db-sync-5vw8g\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.212749 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-config-data-merged\") pod \"octavia-db-sync-5vw8g\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.315418 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-config-data\") pod \"octavia-db-sync-5vw8g\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.315558 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-scripts\") pod \"octavia-db-sync-5vw8g\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.315614 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-config-data-merged\") pod \"octavia-db-sync-5vw8g\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.315828 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-combined-ca-bundle\") pod \"octavia-db-sync-5vw8g\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.316594 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-config-data-merged\") pod \"octavia-db-sync-5vw8g\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.321149 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-scripts\") pod \"octavia-db-sync-5vw8g\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.321408 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-config-data\") pod \"octavia-db-sync-5vw8g\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.333530 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-combined-ca-bundle\") pod \"octavia-db-sync-5vw8g\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:16 crc kubenswrapper[4775]: I1002 03:19:16.463835 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:21 crc kubenswrapper[4775]: I1002 03:19:21.399439 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-5vw8g"] Oct 02 03:19:21 crc kubenswrapper[4775]: I1002 03:19:21.896679 4775 generic.go:334] "Generic (PLEG): container finished" podID="7b44ecb2-8bf4-4c25-841b-3885dbbd0318" containerID="5f56e5ffaef8e27022e4e93c19e7765239c07b45e7a61987b5e7c35cfa99c796" exitCode=0 Oct 02 03:19:21 crc kubenswrapper[4775]: I1002 03:19:21.896796 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-5vw8g" event={"ID":"7b44ecb2-8bf4-4c25-841b-3885dbbd0318","Type":"ContainerDied","Data":"5f56e5ffaef8e27022e4e93c19e7765239c07b45e7a61987b5e7c35cfa99c796"} Oct 02 03:19:21 crc kubenswrapper[4775]: I1002 03:19:21.897029 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-5vw8g" event={"ID":"7b44ecb2-8bf4-4c25-841b-3885dbbd0318","Type":"ContainerStarted","Data":"1787d79b22a51455af71baca6c008f741e540476ba51d386daa7e89ee8908726"} Oct 02 03:19:21 crc kubenswrapper[4775]: I1002 03:19:21.902794 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-6hrw2" event={"ID":"a500c203-7f7e-4aee-96f8-1a4b579b446f","Type":"ContainerStarted","Data":"5669ba481699298add6989b26c76822744f92d1d0cefe3571e055c7c5bb5b711"} Oct 02 03:19:22 crc kubenswrapper[4775]: I1002 03:19:22.915141 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-5vw8g" event={"ID":"7b44ecb2-8bf4-4c25-841b-3885dbbd0318","Type":"ContainerStarted","Data":"f243ea03d6cd70473c444c5b6cfa5bf7f79cb3ced96e5c68bc0c93d48e079f4e"} Oct 02 03:19:22 crc kubenswrapper[4775]: I1002 03:19:22.942139 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-5vw8g" podStartSLOduration=6.942124439 podStartE2EDuration="6.942124439s" podCreationTimestamp="2025-10-02 03:19:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:19:22.935384194 +0000 UTC m=+5900.102128234" watchObservedRunningTime="2025-10-02 03:19:22.942124439 +0000 UTC m=+5900.108868479" Oct 02 03:19:24 crc kubenswrapper[4775]: I1002 03:19:24.413593 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-9l9kr" Oct 02 03:19:25 crc kubenswrapper[4775]: I1002 03:19:25.510667 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:19:25 crc kubenswrapper[4775]: I1002 03:19:25.604806 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-9669b56-5jpsb" Oct 02 03:19:25 crc kubenswrapper[4775]: I1002 03:19:25.769921 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:19:25 crc kubenswrapper[4775]: E1002 03:19:25.770180 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:19:25 crc kubenswrapper[4775]: I1002 03:19:25.962218 4775 generic.go:334] "Generic (PLEG): container finished" podID="a500c203-7f7e-4aee-96f8-1a4b579b446f" containerID="5669ba481699298add6989b26c76822744f92d1d0cefe3571e055c7c5bb5b711" exitCode=0 Oct 02 03:19:25 crc kubenswrapper[4775]: I1002 03:19:25.962263 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-6hrw2" event={"ID":"a500c203-7f7e-4aee-96f8-1a4b579b446f","Type":"ContainerDied","Data":"5669ba481699298add6989b26c76822744f92d1d0cefe3571e055c7c5bb5b711"} Oct 02 03:19:26 crc kubenswrapper[4775]: E1002 03:19:26.712864 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b44ecb2_8bf4_4c25_841b_3885dbbd0318.slice/crio-conmon-f243ea03d6cd70473c444c5b6cfa5bf7f79cb3ced96e5c68bc0c93d48e079f4e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b44ecb2_8bf4_4c25_841b_3885dbbd0318.slice/crio-f243ea03d6cd70473c444c5b6cfa5bf7f79cb3ced96e5c68bc0c93d48e079f4e.scope\": RecentStats: unable to find data in memory cache]" Oct 02 03:19:26 crc kubenswrapper[4775]: I1002 03:19:26.975985 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-5vw8g" event={"ID":"7b44ecb2-8bf4-4c25-841b-3885dbbd0318","Type":"ContainerDied","Data":"f243ea03d6cd70473c444c5b6cfa5bf7f79cb3ced96e5c68bc0c93d48e079f4e"} Oct 02 03:19:26 crc kubenswrapper[4775]: I1002 03:19:26.975943 4775 generic.go:334] "Generic (PLEG): container finished" podID="7b44ecb2-8bf4-4c25-841b-3885dbbd0318" containerID="f243ea03d6cd70473c444c5b6cfa5bf7f79cb3ced96e5c68bc0c93d48e079f4e" exitCode=0 Oct 02 03:19:27 crc kubenswrapper[4775]: I1002 03:19:27.992249 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-6hrw2" event={"ID":"a500c203-7f7e-4aee-96f8-1a4b579b446f","Type":"ContainerStarted","Data":"f66f200331f09ec3cbdc3abc93020ba3c90b1e54688dc2b30baee39f9fc1b7de"} Oct 02 03:19:28 crc kubenswrapper[4775]: I1002 03:19:28.028143 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-6hrw2" podStartSLOduration=2.459149112 podStartE2EDuration="19.028118229s" podCreationTimestamp="2025-10-02 03:19:09 +0000 UTC" firstStartedPulling="2025-10-02 03:19:10.755727338 +0000 UTC m=+5887.922471378" lastFinishedPulling="2025-10-02 03:19:27.324696425 +0000 UTC m=+5904.491440495" observedRunningTime="2025-10-02 03:19:28.023381216 +0000 UTC m=+5905.190125266" watchObservedRunningTime="2025-10-02 03:19:28.028118229 +0000 UTC m=+5905.194862309" Oct 02 03:19:28 crc kubenswrapper[4775]: I1002 03:19:28.497366 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:28 crc kubenswrapper[4775]: I1002 03:19:28.607162 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-config-data\") pod \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " Oct 02 03:19:28 crc kubenswrapper[4775]: I1002 03:19:28.607215 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-combined-ca-bundle\") pod \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " Oct 02 03:19:28 crc kubenswrapper[4775]: I1002 03:19:28.607363 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-config-data-merged\") pod \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " Oct 02 03:19:28 crc kubenswrapper[4775]: I1002 03:19:28.607412 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-scripts\") pod \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\" (UID: \"7b44ecb2-8bf4-4c25-841b-3885dbbd0318\") " Oct 02 03:19:28 crc kubenswrapper[4775]: I1002 03:19:28.613123 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-config-data" (OuterVolumeSpecName: "config-data") pod "7b44ecb2-8bf4-4c25-841b-3885dbbd0318" (UID: "7b44ecb2-8bf4-4c25-841b-3885dbbd0318"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:19:28 crc kubenswrapper[4775]: I1002 03:19:28.616522 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-scripts" (OuterVolumeSpecName: "scripts") pod "7b44ecb2-8bf4-4c25-841b-3885dbbd0318" (UID: "7b44ecb2-8bf4-4c25-841b-3885dbbd0318"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:19:28 crc kubenswrapper[4775]: I1002 03:19:28.628626 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "7b44ecb2-8bf4-4c25-841b-3885dbbd0318" (UID: "7b44ecb2-8bf4-4c25-841b-3885dbbd0318"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:19:28 crc kubenswrapper[4775]: I1002 03:19:28.639522 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b44ecb2-8bf4-4c25-841b-3885dbbd0318" (UID: "7b44ecb2-8bf4-4c25-841b-3885dbbd0318"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:19:28 crc kubenswrapper[4775]: I1002 03:19:28.709180 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:19:28 crc kubenswrapper[4775]: I1002 03:19:28.709220 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:19:28 crc kubenswrapper[4775]: I1002 03:19:28.709232 4775 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-config-data-merged\") on node \"crc\" DevicePath \"\"" Oct 02 03:19:28 crc kubenswrapper[4775]: I1002 03:19:28.709240 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b44ecb2-8bf4-4c25-841b-3885dbbd0318-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:19:29 crc kubenswrapper[4775]: I1002 03:19:29.005934 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-5vw8g" event={"ID":"7b44ecb2-8bf4-4c25-841b-3885dbbd0318","Type":"ContainerDied","Data":"1787d79b22a51455af71baca6c008f741e540476ba51d386daa7e89ee8908726"} Oct 02 03:19:29 crc kubenswrapper[4775]: I1002 03:19:29.005985 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1787d79b22a51455af71baca6c008f741e540476ba51d386daa7e89ee8908726" Oct 02 03:19:29 crc kubenswrapper[4775]: I1002 03:19:29.006038 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-5vw8g" Oct 02 03:19:40 crc kubenswrapper[4775]: I1002 03:19:40.765993 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:19:40 crc kubenswrapper[4775]: E1002 03:19:40.767290 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:19:49 crc kubenswrapper[4775]: I1002 03:19:49.069418 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-zwvbk"] Oct 02 03:19:49 crc kubenswrapper[4775]: I1002 03:19:49.081392 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-zwvbk"] Oct 02 03:19:49 crc kubenswrapper[4775]: I1002 03:19:49.781749 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02d4d565-63c4-4522-a2c5-daa9dc836057" path="/var/lib/kubelet/pods/02d4d565-63c4-4522-a2c5-daa9dc836057/volumes" Oct 02 03:19:55 crc kubenswrapper[4775]: I1002 03:19:55.766609 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:19:55 crc kubenswrapper[4775]: E1002 03:19:55.767659 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:19:58 crc kubenswrapper[4775]: I1002 03:19:58.320839 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-6hrw2"] Oct 02 03:19:58 crc kubenswrapper[4775]: I1002 03:19:58.321431 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-59f8cff499-6hrw2" podUID="a500c203-7f7e-4aee-96f8-1a4b579b446f" containerName="octavia-amphora-httpd" containerID="cri-o://f66f200331f09ec3cbdc3abc93020ba3c90b1e54688dc2b30baee39f9fc1b7de" gracePeriod=30 Oct 02 03:19:58 crc kubenswrapper[4775]: I1002 03:19:58.997893 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-6hrw2" Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.190364 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a500c203-7f7e-4aee-96f8-1a4b579b446f-httpd-config\") pod \"a500c203-7f7e-4aee-96f8-1a4b579b446f\" (UID: \"a500c203-7f7e-4aee-96f8-1a4b579b446f\") " Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.190552 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/a500c203-7f7e-4aee-96f8-1a4b579b446f-amphora-image\") pod \"a500c203-7f7e-4aee-96f8-1a4b579b446f\" (UID: \"a500c203-7f7e-4aee-96f8-1a4b579b446f\") " Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.219595 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a500c203-7f7e-4aee-96f8-1a4b579b446f-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "a500c203-7f7e-4aee-96f8-1a4b579b446f" (UID: "a500c203-7f7e-4aee-96f8-1a4b579b446f"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.236016 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a500c203-7f7e-4aee-96f8-1a4b579b446f-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "a500c203-7f7e-4aee-96f8-1a4b579b446f" (UID: "a500c203-7f7e-4aee-96f8-1a4b579b446f"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.293179 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a500c203-7f7e-4aee-96f8-1a4b579b446f-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.293235 4775 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/a500c203-7f7e-4aee-96f8-1a4b579b446f-amphora-image\") on node \"crc\" DevicePath \"\"" Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.396623 4775 generic.go:334] "Generic (PLEG): container finished" podID="a500c203-7f7e-4aee-96f8-1a4b579b446f" containerID="f66f200331f09ec3cbdc3abc93020ba3c90b1e54688dc2b30baee39f9fc1b7de" exitCode=0 Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.396699 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-6hrw2" event={"ID":"a500c203-7f7e-4aee-96f8-1a4b579b446f","Type":"ContainerDied","Data":"f66f200331f09ec3cbdc3abc93020ba3c90b1e54688dc2b30baee39f9fc1b7de"} Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.396729 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-6hrw2" Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.396766 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-6hrw2" event={"ID":"a500c203-7f7e-4aee-96f8-1a4b579b446f","Type":"ContainerDied","Data":"ee289759a80e4b07d71a8956c8471e317261e2582bcd5aec281c53c2dbfe7bb8"} Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.396805 4775 scope.go:117] "RemoveContainer" containerID="f66f200331f09ec3cbdc3abc93020ba3c90b1e54688dc2b30baee39f9fc1b7de" Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.431330 4775 scope.go:117] "RemoveContainer" containerID="5669ba481699298add6989b26c76822744f92d1d0cefe3571e055c7c5bb5b711" Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.457172 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-6hrw2"] Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.472056 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-6hrw2"] Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.476380 4775 scope.go:117] "RemoveContainer" containerID="f66f200331f09ec3cbdc3abc93020ba3c90b1e54688dc2b30baee39f9fc1b7de" Oct 02 03:19:59 crc kubenswrapper[4775]: E1002 03:19:59.477124 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f66f200331f09ec3cbdc3abc93020ba3c90b1e54688dc2b30baee39f9fc1b7de\": container with ID starting with f66f200331f09ec3cbdc3abc93020ba3c90b1e54688dc2b30baee39f9fc1b7de not found: ID does not exist" containerID="f66f200331f09ec3cbdc3abc93020ba3c90b1e54688dc2b30baee39f9fc1b7de" Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.477179 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f66f200331f09ec3cbdc3abc93020ba3c90b1e54688dc2b30baee39f9fc1b7de"} err="failed to get container status \"f66f200331f09ec3cbdc3abc93020ba3c90b1e54688dc2b30baee39f9fc1b7de\": rpc error: code = NotFound desc = could not find container \"f66f200331f09ec3cbdc3abc93020ba3c90b1e54688dc2b30baee39f9fc1b7de\": container with ID starting with f66f200331f09ec3cbdc3abc93020ba3c90b1e54688dc2b30baee39f9fc1b7de not found: ID does not exist" Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.477210 4775 scope.go:117] "RemoveContainer" containerID="5669ba481699298add6989b26c76822744f92d1d0cefe3571e055c7c5bb5b711" Oct 02 03:19:59 crc kubenswrapper[4775]: E1002 03:19:59.477695 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5669ba481699298add6989b26c76822744f92d1d0cefe3571e055c7c5bb5b711\": container with ID starting with 5669ba481699298add6989b26c76822744f92d1d0cefe3571e055c7c5bb5b711 not found: ID does not exist" containerID="5669ba481699298add6989b26c76822744f92d1d0cefe3571e055c7c5bb5b711" Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.477831 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5669ba481699298add6989b26c76822744f92d1d0cefe3571e055c7c5bb5b711"} err="failed to get container status \"5669ba481699298add6989b26c76822744f92d1d0cefe3571e055c7c5bb5b711\": rpc error: code = NotFound desc = could not find container \"5669ba481699298add6989b26c76822744f92d1d0cefe3571e055c7c5bb5b711\": container with ID starting with 5669ba481699298add6989b26c76822744f92d1d0cefe3571e055c7c5bb5b711 not found: ID does not exist" Oct 02 03:19:59 crc kubenswrapper[4775]: I1002 03:19:59.775453 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a500c203-7f7e-4aee-96f8-1a4b579b446f" path="/var/lib/kubelet/pods/a500c203-7f7e-4aee-96f8-1a4b579b446f/volumes" Oct 02 03:20:00 crc kubenswrapper[4775]: I1002 03:20:00.040823 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-c027-account-create-rn5zs"] Oct 02 03:20:00 crc kubenswrapper[4775]: I1002 03:20:00.055635 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-c027-account-create-rn5zs"] Oct 02 03:20:01 crc kubenswrapper[4775]: I1002 03:20:01.778935 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc0b72bd-79d3-4195-b494-d2c783115fd7" path="/var/lib/kubelet/pods/fc0b72bd-79d3-4195-b494-d2c783115fd7/volumes" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.564073 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-dcpms"] Oct 02 03:20:05 crc kubenswrapper[4775]: E1002 03:20:05.564899 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a500c203-7f7e-4aee-96f8-1a4b579b446f" containerName="init" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.564911 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a500c203-7f7e-4aee-96f8-1a4b579b446f" containerName="init" Oct 02 03:20:05 crc kubenswrapper[4775]: E1002 03:20:05.564931 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b44ecb2-8bf4-4c25-841b-3885dbbd0318" containerName="octavia-db-sync" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.564936 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b44ecb2-8bf4-4c25-841b-3885dbbd0318" containerName="octavia-db-sync" Oct 02 03:20:05 crc kubenswrapper[4775]: E1002 03:20:05.564947 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b44ecb2-8bf4-4c25-841b-3885dbbd0318" containerName="init" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.564968 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b44ecb2-8bf4-4c25-841b-3885dbbd0318" containerName="init" Oct 02 03:20:05 crc kubenswrapper[4775]: E1002 03:20:05.564981 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a500c203-7f7e-4aee-96f8-1a4b579b446f" containerName="octavia-amphora-httpd" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.564987 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a500c203-7f7e-4aee-96f8-1a4b579b446f" containerName="octavia-amphora-httpd" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.565157 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a500c203-7f7e-4aee-96f8-1a4b579b446f" containerName="octavia-amphora-httpd" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.565169 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b44ecb2-8bf4-4c25-841b-3885dbbd0318" containerName="octavia-db-sync" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.566144 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-dcpms" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.568996 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.589119 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-dcpms"] Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.732693 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/462e5b8f-d3ae-4073-ba5f-18a1d992b634-amphora-image\") pod \"octavia-image-upload-59f8cff499-dcpms\" (UID: \"462e5b8f-d3ae-4073-ba5f-18a1d992b634\") " pod="openstack/octavia-image-upload-59f8cff499-dcpms" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.733113 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/462e5b8f-d3ae-4073-ba5f-18a1d992b634-httpd-config\") pod \"octavia-image-upload-59f8cff499-dcpms\" (UID: \"462e5b8f-d3ae-4073-ba5f-18a1d992b634\") " pod="openstack/octavia-image-upload-59f8cff499-dcpms" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.835237 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/462e5b8f-d3ae-4073-ba5f-18a1d992b634-amphora-image\") pod \"octavia-image-upload-59f8cff499-dcpms\" (UID: \"462e5b8f-d3ae-4073-ba5f-18a1d992b634\") " pod="openstack/octavia-image-upload-59f8cff499-dcpms" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.835512 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/462e5b8f-d3ae-4073-ba5f-18a1d992b634-httpd-config\") pod \"octavia-image-upload-59f8cff499-dcpms\" (UID: \"462e5b8f-d3ae-4073-ba5f-18a1d992b634\") " pod="openstack/octavia-image-upload-59f8cff499-dcpms" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.836016 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/462e5b8f-d3ae-4073-ba5f-18a1d992b634-amphora-image\") pod \"octavia-image-upload-59f8cff499-dcpms\" (UID: \"462e5b8f-d3ae-4073-ba5f-18a1d992b634\") " pod="openstack/octavia-image-upload-59f8cff499-dcpms" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.842485 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/462e5b8f-d3ae-4073-ba5f-18a1d992b634-httpd-config\") pod \"octavia-image-upload-59f8cff499-dcpms\" (UID: \"462e5b8f-d3ae-4073-ba5f-18a1d992b634\") " pod="openstack/octavia-image-upload-59f8cff499-dcpms" Oct 02 03:20:05 crc kubenswrapper[4775]: I1002 03:20:05.892436 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-dcpms" Oct 02 03:20:06 crc kubenswrapper[4775]: I1002 03:20:06.039158 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-5758k"] Oct 02 03:20:06 crc kubenswrapper[4775]: I1002 03:20:06.050026 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-5758k"] Oct 02 03:20:06 crc kubenswrapper[4775]: I1002 03:20:06.375452 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-dcpms"] Oct 02 03:20:06 crc kubenswrapper[4775]: I1002 03:20:06.489124 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-dcpms" event={"ID":"462e5b8f-d3ae-4073-ba5f-18a1d992b634","Type":"ContainerStarted","Data":"f3b34d998a085ca97960bbbe89e67d9aa542e72bdc3d2339ac2b13817521367d"} Oct 02 03:20:07 crc kubenswrapper[4775]: I1002 03:20:07.503444 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-dcpms" event={"ID":"462e5b8f-d3ae-4073-ba5f-18a1d992b634","Type":"ContainerStarted","Data":"db78d822afa857385471e481e2d7e2b1e178ee94d160b1489a9fa4d869f9ddbb"} Oct 02 03:20:07 crc kubenswrapper[4775]: I1002 03:20:07.780476 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f14ca3c4-6278-41a9-94b2-29045c809785" path="/var/lib/kubelet/pods/f14ca3c4-6278-41a9-94b2-29045c809785/volumes" Oct 02 03:20:08 crc kubenswrapper[4775]: I1002 03:20:08.765674 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:20:08 crc kubenswrapper[4775]: E1002 03:20:08.766488 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:20:10 crc kubenswrapper[4775]: I1002 03:20:10.547000 4775 generic.go:334] "Generic (PLEG): container finished" podID="462e5b8f-d3ae-4073-ba5f-18a1d992b634" containerID="db78d822afa857385471e481e2d7e2b1e178ee94d160b1489a9fa4d869f9ddbb" exitCode=0 Oct 02 03:20:10 crc kubenswrapper[4775]: I1002 03:20:10.547111 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-dcpms" event={"ID":"462e5b8f-d3ae-4073-ba5f-18a1d992b634","Type":"ContainerDied","Data":"db78d822afa857385471e481e2d7e2b1e178ee94d160b1489a9fa4d869f9ddbb"} Oct 02 03:20:12 crc kubenswrapper[4775]: I1002 03:20:12.387480 4775 scope.go:117] "RemoveContainer" containerID="bae7647a6403874672d4069f85fa0fc18c870bf6bea9eaa4a263f8242571c0a9" Oct 02 03:20:12 crc kubenswrapper[4775]: I1002 03:20:12.411420 4775 scope.go:117] "RemoveContainer" containerID="bda9b3638a12955f1d1a462b8f97fd8d02dde1dce8c49dd422d50ab5bdc5b5d6" Oct 02 03:20:12 crc kubenswrapper[4775]: I1002 03:20:12.469206 4775 scope.go:117] "RemoveContainer" containerID="9b3a5a3dbe93481567b2f28b1194c37ea150030e6f7f1589415a4fae7a391161" Oct 02 03:20:12 crc kubenswrapper[4775]: I1002 03:20:12.509934 4775 scope.go:117] "RemoveContainer" containerID="253c2b542479ae7d61337ab667bbb0c152706bf9aa25f6a81c70f8682aa874d9" Oct 02 03:20:12 crc kubenswrapper[4775]: I1002 03:20:12.557133 4775 scope.go:117] "RemoveContainer" containerID="45ebcf3bd424f58ba6a18ad63bead97953a421eb5bbee73f38099235edb4495b" Oct 02 03:20:12 crc kubenswrapper[4775]: I1002 03:20:12.572467 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-dcpms" event={"ID":"462e5b8f-d3ae-4073-ba5f-18a1d992b634","Type":"ContainerStarted","Data":"cb2963e70eff985a9e729a3d4194d0c6dee703d7f2d7b73df131e941c8f2b88b"} Oct 02 03:20:12 crc kubenswrapper[4775]: I1002 03:20:12.605912 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-dcpms" podStartSLOduration=2.2274595440000002 podStartE2EDuration="7.605890471s" podCreationTimestamp="2025-10-02 03:20:05 +0000 UTC" firstStartedPulling="2025-10-02 03:20:06.392841141 +0000 UTC m=+5943.559585191" lastFinishedPulling="2025-10-02 03:20:11.771272068 +0000 UTC m=+5948.938016118" observedRunningTime="2025-10-02 03:20:12.58846621 +0000 UTC m=+5949.755210270" watchObservedRunningTime="2025-10-02 03:20:12.605890471 +0000 UTC m=+5949.772634512" Oct 02 03:20:12 crc kubenswrapper[4775]: I1002 03:20:12.627283 4775 scope.go:117] "RemoveContainer" containerID="7d10971f75f7a0ea376bc95cdd84d53c288f132218505d3ac1eab456caabfd77" Oct 02 03:20:16 crc kubenswrapper[4775]: I1002 03:20:16.792810 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-vrtnf"] Oct 02 03:20:16 crc kubenswrapper[4775]: I1002 03:20:16.796034 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:16 crc kubenswrapper[4775]: I1002 03:20:16.798758 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Oct 02 03:20:16 crc kubenswrapper[4775]: I1002 03:20:16.799034 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Oct 02 03:20:16 crc kubenswrapper[4775]: I1002 03:20:16.799124 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Oct 02 03:20:16 crc kubenswrapper[4775]: I1002 03:20:16.814027 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-vrtnf"] Oct 02 03:20:16 crc kubenswrapper[4775]: I1002 03:20:16.980357 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c72a379-38b8-4152-80f9-49c83f7b8719-combined-ca-bundle\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:16 crc kubenswrapper[4775]: I1002 03:20:16.980485 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2c72a379-38b8-4152-80f9-49c83f7b8719-config-data-merged\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:16 crc kubenswrapper[4775]: I1002 03:20:16.980720 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c72a379-38b8-4152-80f9-49c83f7b8719-scripts\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:16 crc kubenswrapper[4775]: I1002 03:20:16.980906 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2c72a379-38b8-4152-80f9-49c83f7b8719-hm-ports\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:16 crc kubenswrapper[4775]: I1002 03:20:16.981068 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/2c72a379-38b8-4152-80f9-49c83f7b8719-amphora-certs\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:16 crc kubenswrapper[4775]: I1002 03:20:16.981259 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c72a379-38b8-4152-80f9-49c83f7b8719-config-data\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:17 crc kubenswrapper[4775]: I1002 03:20:17.082743 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c72a379-38b8-4152-80f9-49c83f7b8719-combined-ca-bundle\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:17 crc kubenswrapper[4775]: I1002 03:20:17.082824 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2c72a379-38b8-4152-80f9-49c83f7b8719-config-data-merged\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:17 crc kubenswrapper[4775]: I1002 03:20:17.082880 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c72a379-38b8-4152-80f9-49c83f7b8719-scripts\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:17 crc kubenswrapper[4775]: I1002 03:20:17.082915 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2c72a379-38b8-4152-80f9-49c83f7b8719-hm-ports\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:17 crc kubenswrapper[4775]: I1002 03:20:17.082975 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/2c72a379-38b8-4152-80f9-49c83f7b8719-amphora-certs\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:17 crc kubenswrapper[4775]: I1002 03:20:17.083020 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c72a379-38b8-4152-80f9-49c83f7b8719-config-data\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:17 crc kubenswrapper[4775]: I1002 03:20:17.085133 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/2c72a379-38b8-4152-80f9-49c83f7b8719-config-data-merged\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:17 crc kubenswrapper[4775]: I1002 03:20:17.086139 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/2c72a379-38b8-4152-80f9-49c83f7b8719-hm-ports\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:17 crc kubenswrapper[4775]: I1002 03:20:17.094841 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/2c72a379-38b8-4152-80f9-49c83f7b8719-amphora-certs\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:17 crc kubenswrapper[4775]: I1002 03:20:17.095286 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2c72a379-38b8-4152-80f9-49c83f7b8719-combined-ca-bundle\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:17 crc kubenswrapper[4775]: I1002 03:20:17.095823 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2c72a379-38b8-4152-80f9-49c83f7b8719-config-data\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:17 crc kubenswrapper[4775]: I1002 03:20:17.104341 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2c72a379-38b8-4152-80f9-49c83f7b8719-scripts\") pod \"octavia-healthmanager-vrtnf\" (UID: \"2c72a379-38b8-4152-80f9-49c83f7b8719\") " pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:17 crc kubenswrapper[4775]: I1002 03:20:17.129296 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:17 crc kubenswrapper[4775]: I1002 03:20:17.738081 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-vrtnf"] Oct 02 03:20:18 crc kubenswrapper[4775]: I1002 03:20:18.676044 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-vrtnf" event={"ID":"2c72a379-38b8-4152-80f9-49c83f7b8719","Type":"ContainerStarted","Data":"b252180a42caef7ceae6ee1709d7b63749efa330dbaedc2d3d907acd81872270"} Oct 02 03:20:18 crc kubenswrapper[4775]: I1002 03:20:18.676307 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-vrtnf" event={"ID":"2c72a379-38b8-4152-80f9-49c83f7b8719","Type":"ContainerStarted","Data":"89bc43defd34157e510c8136d4b4470c446bedc603685dfa0c528289c1d3fe50"} Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.338254 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-4tlkz"] Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.340110 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.342396 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.342588 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.350257 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-4tlkz"] Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.530541 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26dc7016-0e75-4919-957a-feac2af12ab0-combined-ca-bundle\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.530591 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26dc7016-0e75-4919-957a-feac2af12ab0-scripts\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.530629 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/26dc7016-0e75-4919-957a-feac2af12ab0-hm-ports\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.530940 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/26dc7016-0e75-4919-957a-feac2af12ab0-amphora-certs\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.531059 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/26dc7016-0e75-4919-957a-feac2af12ab0-config-data-merged\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.531110 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26dc7016-0e75-4919-957a-feac2af12ab0-config-data\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.632892 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/26dc7016-0e75-4919-957a-feac2af12ab0-amphora-certs\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.632961 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/26dc7016-0e75-4919-957a-feac2af12ab0-config-data-merged\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.632987 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26dc7016-0e75-4919-957a-feac2af12ab0-config-data\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.633055 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26dc7016-0e75-4919-957a-feac2af12ab0-combined-ca-bundle\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.633091 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26dc7016-0e75-4919-957a-feac2af12ab0-scripts\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.633124 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/26dc7016-0e75-4919-957a-feac2af12ab0-hm-ports\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.633736 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/26dc7016-0e75-4919-957a-feac2af12ab0-config-data-merged\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.634640 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/26dc7016-0e75-4919-957a-feac2af12ab0-hm-ports\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.640727 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26dc7016-0e75-4919-957a-feac2af12ab0-scripts\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.652603 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26dc7016-0e75-4919-957a-feac2af12ab0-config-data\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.653532 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26dc7016-0e75-4919-957a-feac2af12ab0-combined-ca-bundle\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.654128 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/26dc7016-0e75-4919-957a-feac2af12ab0-amphora-certs\") pod \"octavia-housekeeping-4tlkz\" (UID: \"26dc7016-0e75-4919-957a-feac2af12ab0\") " pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:19 crc kubenswrapper[4775]: I1002 03:20:19.659299 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.201884 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-twql4"] Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.205161 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.219216 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-4tlkz"] Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.227073 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-twql4"] Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.346626 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c9789ad-1c3f-40eb-a012-96551808b45d-catalog-content\") pod \"redhat-marketplace-twql4\" (UID: \"1c9789ad-1c3f-40eb-a012-96551808b45d\") " pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.346686 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25d54\" (UniqueName: \"kubernetes.io/projected/1c9789ad-1c3f-40eb-a012-96551808b45d-kube-api-access-25d54\") pod \"redhat-marketplace-twql4\" (UID: \"1c9789ad-1c3f-40eb-a012-96551808b45d\") " pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.346719 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c9789ad-1c3f-40eb-a012-96551808b45d-utilities\") pod \"redhat-marketplace-twql4\" (UID: \"1c9789ad-1c3f-40eb-a012-96551808b45d\") " pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.448499 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c9789ad-1c3f-40eb-a012-96551808b45d-catalog-content\") pod \"redhat-marketplace-twql4\" (UID: \"1c9789ad-1c3f-40eb-a012-96551808b45d\") " pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.448561 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25d54\" (UniqueName: \"kubernetes.io/projected/1c9789ad-1c3f-40eb-a012-96551808b45d-kube-api-access-25d54\") pod \"redhat-marketplace-twql4\" (UID: \"1c9789ad-1c3f-40eb-a012-96551808b45d\") " pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.448596 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c9789ad-1c3f-40eb-a012-96551808b45d-utilities\") pod \"redhat-marketplace-twql4\" (UID: \"1c9789ad-1c3f-40eb-a012-96551808b45d\") " pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.449028 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c9789ad-1c3f-40eb-a012-96551808b45d-catalog-content\") pod \"redhat-marketplace-twql4\" (UID: \"1c9789ad-1c3f-40eb-a012-96551808b45d\") " pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.449094 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c9789ad-1c3f-40eb-a012-96551808b45d-utilities\") pod \"redhat-marketplace-twql4\" (UID: \"1c9789ad-1c3f-40eb-a012-96551808b45d\") " pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.469136 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25d54\" (UniqueName: \"kubernetes.io/projected/1c9789ad-1c3f-40eb-a012-96551808b45d-kube-api-access-25d54\") pod \"redhat-marketplace-twql4\" (UID: \"1c9789ad-1c3f-40eb-a012-96551808b45d\") " pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.585084 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.699643 4775 generic.go:334] "Generic (PLEG): container finished" podID="2c72a379-38b8-4152-80f9-49c83f7b8719" containerID="b252180a42caef7ceae6ee1709d7b63749efa330dbaedc2d3d907acd81872270" exitCode=0 Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.699850 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-vrtnf" event={"ID":"2c72a379-38b8-4152-80f9-49c83f7b8719","Type":"ContainerDied","Data":"b252180a42caef7ceae6ee1709d7b63749efa330dbaedc2d3d907acd81872270"} Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.701677 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-4tlkz" event={"ID":"26dc7016-0e75-4919-957a-feac2af12ab0","Type":"ContainerStarted","Data":"dd19926c6e6982b7ed8d03a7d5b1f2961df0f7641c3f53f3307ad0cce1d77152"} Oct 02 03:20:20 crc kubenswrapper[4775]: I1002 03:20:20.764802 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:20:20 crc kubenswrapper[4775]: E1002 03:20:20.765166 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:20:21 crc kubenswrapper[4775]: I1002 03:20:21.043527 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-twql4"] Oct 02 03:20:21 crc kubenswrapper[4775]: I1002 03:20:21.714226 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-vrtnf" event={"ID":"2c72a379-38b8-4152-80f9-49c83f7b8719","Type":"ContainerStarted","Data":"82714dd324ecd77038b03ba4587dbc38e27a36876d303e3ef5564dcfdd45a367"} Oct 02 03:20:21 crc kubenswrapper[4775]: I1002 03:20:21.715757 4775 generic.go:334] "Generic (PLEG): container finished" podID="1c9789ad-1c3f-40eb-a012-96551808b45d" containerID="d583a650144c43c7a2365cac80e6df9d96dc6e2df101164ab2f393c2b4493f68" exitCode=0 Oct 02 03:20:21 crc kubenswrapper[4775]: I1002 03:20:21.715783 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:21 crc kubenswrapper[4775]: I1002 03:20:21.715850 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-twql4" event={"ID":"1c9789ad-1c3f-40eb-a012-96551808b45d","Type":"ContainerDied","Data":"d583a650144c43c7a2365cac80e6df9d96dc6e2df101164ab2f393c2b4493f68"} Oct 02 03:20:21 crc kubenswrapper[4775]: I1002 03:20:21.715877 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-twql4" event={"ID":"1c9789ad-1c3f-40eb-a012-96551808b45d","Type":"ContainerStarted","Data":"9f46fc5cef38f3bdbde9ccee457909b99ecadfbc25015fc31f89455e95da0541"} Oct 02 03:20:21 crc kubenswrapper[4775]: I1002 03:20:21.732933 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-vrtnf" podStartSLOduration=5.732920588 podStartE2EDuration="5.732920588s" podCreationTimestamp="2025-10-02 03:20:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:20:21.732089987 +0000 UTC m=+5958.898834027" watchObservedRunningTime="2025-10-02 03:20:21.732920588 +0000 UTC m=+5958.899664628" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.492268 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-crg8r"] Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.495082 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.497006 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.497359 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.527875 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-crg8r"] Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.599606 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/516737d3-2942-4df1-861e-a4e0dbaccbf7-config-data\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.599666 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/516737d3-2942-4df1-861e-a4e0dbaccbf7-config-data-merged\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.600348 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/516737d3-2942-4df1-861e-a4e0dbaccbf7-amphora-certs\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.601047 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/516737d3-2942-4df1-861e-a4e0dbaccbf7-combined-ca-bundle\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.601095 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/516737d3-2942-4df1-861e-a4e0dbaccbf7-scripts\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.601118 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/516737d3-2942-4df1-861e-a4e0dbaccbf7-hm-ports\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.706518 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/516737d3-2942-4df1-861e-a4e0dbaccbf7-config-data\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.706593 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/516737d3-2942-4df1-861e-a4e0dbaccbf7-config-data-merged\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.706642 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/516737d3-2942-4df1-861e-a4e0dbaccbf7-amphora-certs\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.706907 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/516737d3-2942-4df1-861e-a4e0dbaccbf7-combined-ca-bundle\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.706987 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/516737d3-2942-4df1-861e-a4e0dbaccbf7-scripts\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.707021 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/516737d3-2942-4df1-861e-a4e0dbaccbf7-hm-ports\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.708099 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/516737d3-2942-4df1-861e-a4e0dbaccbf7-config-data-merged\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.708818 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/516737d3-2942-4df1-861e-a4e0dbaccbf7-hm-ports\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.713298 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/516737d3-2942-4df1-861e-a4e0dbaccbf7-combined-ca-bundle\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.714061 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/516737d3-2942-4df1-861e-a4e0dbaccbf7-config-data\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.714584 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/516737d3-2942-4df1-861e-a4e0dbaccbf7-amphora-certs\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.716041 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/516737d3-2942-4df1-861e-a4e0dbaccbf7-scripts\") pod \"octavia-worker-crg8r\" (UID: \"516737d3-2942-4df1-861e-a4e0dbaccbf7\") " pod="openstack/octavia-worker-crg8r" Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.728257 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-4tlkz" event={"ID":"26dc7016-0e75-4919-957a-feac2af12ab0","Type":"ContainerStarted","Data":"5586fc7e07b91e960a4883e3ae990a1255b3f98d546858b73d801cca4e7575d4"} Oct 02 03:20:22 crc kubenswrapper[4775]: I1002 03:20:22.854268 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-crg8r" Oct 02 03:20:23 crc kubenswrapper[4775]: I1002 03:20:23.381838 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-crg8r"] Oct 02 03:20:23 crc kubenswrapper[4775]: I1002 03:20:23.740684 4775 generic.go:334] "Generic (PLEG): container finished" podID="26dc7016-0e75-4919-957a-feac2af12ab0" containerID="5586fc7e07b91e960a4883e3ae990a1255b3f98d546858b73d801cca4e7575d4" exitCode=0 Oct 02 03:20:23 crc kubenswrapper[4775]: I1002 03:20:23.740757 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-4tlkz" event={"ID":"26dc7016-0e75-4919-957a-feac2af12ab0","Type":"ContainerDied","Data":"5586fc7e07b91e960a4883e3ae990a1255b3f98d546858b73d801cca4e7575d4"} Oct 02 03:20:23 crc kubenswrapper[4775]: I1002 03:20:23.748990 4775 generic.go:334] "Generic (PLEG): container finished" podID="1c9789ad-1c3f-40eb-a012-96551808b45d" containerID="d815f68fc61e213a829827c4ddb6cc31707b17787817cd510899aefb70c243b1" exitCode=0 Oct 02 03:20:23 crc kubenswrapper[4775]: I1002 03:20:23.749195 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-twql4" event={"ID":"1c9789ad-1c3f-40eb-a012-96551808b45d","Type":"ContainerDied","Data":"d815f68fc61e213a829827c4ddb6cc31707b17787817cd510899aefb70c243b1"} Oct 02 03:20:23 crc kubenswrapper[4775]: I1002 03:20:23.760811 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-crg8r" event={"ID":"516737d3-2942-4df1-861e-a4e0dbaccbf7","Type":"ContainerStarted","Data":"5890aef0c9eb98952fe45e5831778152135bea4a7e51a82722e826e9246f64f0"} Oct 02 03:20:24 crc kubenswrapper[4775]: I1002 03:20:24.772612 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-4tlkz" event={"ID":"26dc7016-0e75-4919-957a-feac2af12ab0","Type":"ContainerStarted","Data":"d8142369c3ca0eadf1b3d90d1e622bee8307455306350021ad3fc805c1eb0163"} Oct 02 03:20:24 crc kubenswrapper[4775]: I1002 03:20:24.772914 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:24 crc kubenswrapper[4775]: I1002 03:20:24.785158 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-twql4" event={"ID":"1c9789ad-1c3f-40eb-a012-96551808b45d","Type":"ContainerStarted","Data":"19dd2ff55f62b3ee2fd45cc93499c107214cb224273c4fd5e8763ce97579347c"} Oct 02 03:20:24 crc kubenswrapper[4775]: I1002 03:20:24.792064 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-4tlkz" podStartSLOduration=4.035014414 podStartE2EDuration="5.792047696s" podCreationTimestamp="2025-10-02 03:20:19 +0000 UTC" firstStartedPulling="2025-10-02 03:20:20.23229749 +0000 UTC m=+5957.399041540" lastFinishedPulling="2025-10-02 03:20:21.989330782 +0000 UTC m=+5959.156074822" observedRunningTime="2025-10-02 03:20:24.78871535 +0000 UTC m=+5961.955459400" watchObservedRunningTime="2025-10-02 03:20:24.792047696 +0000 UTC m=+5961.958791736" Oct 02 03:20:24 crc kubenswrapper[4775]: I1002 03:20:24.809429 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-twql4" podStartSLOduration=2.533359216 podStartE2EDuration="4.809410706s" podCreationTimestamp="2025-10-02 03:20:20 +0000 UTC" firstStartedPulling="2025-10-02 03:20:21.952812925 +0000 UTC m=+5959.119556965" lastFinishedPulling="2025-10-02 03:20:24.228864405 +0000 UTC m=+5961.395608455" observedRunningTime="2025-10-02 03:20:24.803155954 +0000 UTC m=+5961.969899994" watchObservedRunningTime="2025-10-02 03:20:24.809410706 +0000 UTC m=+5961.976154746" Oct 02 03:20:25 crc kubenswrapper[4775]: I1002 03:20:25.798662 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-crg8r" event={"ID":"516737d3-2942-4df1-861e-a4e0dbaccbf7","Type":"ContainerStarted","Data":"1fdfe19578c987711062f2c88e4c5b8510d7e55d2b7145fed3a2470bef2604bf"} Oct 02 03:20:26 crc kubenswrapper[4775]: I1002 03:20:26.811902 4775 generic.go:334] "Generic (PLEG): container finished" podID="516737d3-2942-4df1-861e-a4e0dbaccbf7" containerID="1fdfe19578c987711062f2c88e4c5b8510d7e55d2b7145fed3a2470bef2604bf" exitCode=0 Oct 02 03:20:26 crc kubenswrapper[4775]: I1002 03:20:26.811991 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-crg8r" event={"ID":"516737d3-2942-4df1-861e-a4e0dbaccbf7","Type":"ContainerDied","Data":"1fdfe19578c987711062f2c88e4c5b8510d7e55d2b7145fed3a2470bef2604bf"} Oct 02 03:20:27 crc kubenswrapper[4775]: I1002 03:20:27.823100 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-crg8r" event={"ID":"516737d3-2942-4df1-861e-a4e0dbaccbf7","Type":"ContainerStarted","Data":"5490608b7a2b3b2e97affcd310d2b05aedb8934bcb08f6fc09bd38a288e1417d"} Oct 02 03:20:27 crc kubenswrapper[4775]: I1002 03:20:27.823396 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-crg8r" Oct 02 03:20:27 crc kubenswrapper[4775]: I1002 03:20:27.853951 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-crg8r" podStartSLOduration=4.365262686 podStartE2EDuration="5.853930325s" podCreationTimestamp="2025-10-02 03:20:22 +0000 UTC" firstStartedPulling="2025-10-02 03:20:23.388366899 +0000 UTC m=+5960.555110949" lastFinishedPulling="2025-10-02 03:20:24.877034548 +0000 UTC m=+5962.043778588" observedRunningTime="2025-10-02 03:20:27.843488875 +0000 UTC m=+5965.010232915" watchObservedRunningTime="2025-10-02 03:20:27.853930325 +0000 UTC m=+5965.020674385" Oct 02 03:20:30 crc kubenswrapper[4775]: I1002 03:20:30.585699 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:30 crc kubenswrapper[4775]: I1002 03:20:30.586468 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:30 crc kubenswrapper[4775]: I1002 03:20:30.664087 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:30 crc kubenswrapper[4775]: I1002 03:20:30.928415 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:31 crc kubenswrapper[4775]: I1002 03:20:31.002265 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-twql4"] Oct 02 03:20:32 crc kubenswrapper[4775]: I1002 03:20:32.175320 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-vrtnf" Oct 02 03:20:32 crc kubenswrapper[4775]: I1002 03:20:32.878804 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-twql4" podUID="1c9789ad-1c3f-40eb-a012-96551808b45d" containerName="registry-server" containerID="cri-o://19dd2ff55f62b3ee2fd45cc93499c107214cb224273c4fd5e8763ce97579347c" gracePeriod=2 Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.050313 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-66vnq"] Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.058765 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-66vnq"] Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.453296 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.549709 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c9789ad-1c3f-40eb-a012-96551808b45d-catalog-content\") pod \"1c9789ad-1c3f-40eb-a012-96551808b45d\" (UID: \"1c9789ad-1c3f-40eb-a012-96551808b45d\") " Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.551051 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c9789ad-1c3f-40eb-a012-96551808b45d-utilities\") pod \"1c9789ad-1c3f-40eb-a012-96551808b45d\" (UID: \"1c9789ad-1c3f-40eb-a012-96551808b45d\") " Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.551245 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-25d54\" (UniqueName: \"kubernetes.io/projected/1c9789ad-1c3f-40eb-a012-96551808b45d-kube-api-access-25d54\") pod \"1c9789ad-1c3f-40eb-a012-96551808b45d\" (UID: \"1c9789ad-1c3f-40eb-a012-96551808b45d\") " Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.551943 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c9789ad-1c3f-40eb-a012-96551808b45d-utilities" (OuterVolumeSpecName: "utilities") pod "1c9789ad-1c3f-40eb-a012-96551808b45d" (UID: "1c9789ad-1c3f-40eb-a012-96551808b45d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.560548 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c9789ad-1c3f-40eb-a012-96551808b45d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c9789ad-1c3f-40eb-a012-96551808b45d" (UID: "1c9789ad-1c3f-40eb-a012-96551808b45d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.562987 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c9789ad-1c3f-40eb-a012-96551808b45d-kube-api-access-25d54" (OuterVolumeSpecName: "kube-api-access-25d54") pod "1c9789ad-1c3f-40eb-a012-96551808b45d" (UID: "1c9789ad-1c3f-40eb-a012-96551808b45d"). InnerVolumeSpecName "kube-api-access-25d54". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.653265 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c9789ad-1c3f-40eb-a012-96551808b45d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.653297 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c9789ad-1c3f-40eb-a012-96551808b45d-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.653309 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-25d54\" (UniqueName: \"kubernetes.io/projected/1c9789ad-1c3f-40eb-a012-96551808b45d-kube-api-access-25d54\") on node \"crc\" DevicePath \"\"" Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.778944 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:20:33 crc kubenswrapper[4775]: E1002 03:20:33.779522 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.782516 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3050a03c-11f3-416c-a630-a75c621e949e" path="/var/lib/kubelet/pods/3050a03c-11f3-416c-a630-a75c621e949e/volumes" Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.891142 4775 generic.go:334] "Generic (PLEG): container finished" podID="1c9789ad-1c3f-40eb-a012-96551808b45d" containerID="19dd2ff55f62b3ee2fd45cc93499c107214cb224273c4fd5e8763ce97579347c" exitCode=0 Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.891201 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-twql4" event={"ID":"1c9789ad-1c3f-40eb-a012-96551808b45d","Type":"ContainerDied","Data":"19dd2ff55f62b3ee2fd45cc93499c107214cb224273c4fd5e8763ce97579347c"} Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.891241 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-twql4" event={"ID":"1c9789ad-1c3f-40eb-a012-96551808b45d","Type":"ContainerDied","Data":"9f46fc5cef38f3bdbde9ccee457909b99ecadfbc25015fc31f89455e95da0541"} Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.891268 4775 scope.go:117] "RemoveContainer" containerID="19dd2ff55f62b3ee2fd45cc93499c107214cb224273c4fd5e8763ce97579347c" Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.891480 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-twql4" Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.923504 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-twql4"] Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.925807 4775 scope.go:117] "RemoveContainer" containerID="d815f68fc61e213a829827c4ddb6cc31707b17787817cd510899aefb70c243b1" Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.932813 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-twql4"] Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.966210 4775 scope.go:117] "RemoveContainer" containerID="d583a650144c43c7a2365cac80e6df9d96dc6e2df101164ab2f393c2b4493f68" Oct 02 03:20:33 crc kubenswrapper[4775]: I1002 03:20:33.999902 4775 scope.go:117] "RemoveContainer" containerID="19dd2ff55f62b3ee2fd45cc93499c107214cb224273c4fd5e8763ce97579347c" Oct 02 03:20:34 crc kubenswrapper[4775]: E1002 03:20:34.005482 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19dd2ff55f62b3ee2fd45cc93499c107214cb224273c4fd5e8763ce97579347c\": container with ID starting with 19dd2ff55f62b3ee2fd45cc93499c107214cb224273c4fd5e8763ce97579347c not found: ID does not exist" containerID="19dd2ff55f62b3ee2fd45cc93499c107214cb224273c4fd5e8763ce97579347c" Oct 02 03:20:34 crc kubenswrapper[4775]: I1002 03:20:34.005554 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19dd2ff55f62b3ee2fd45cc93499c107214cb224273c4fd5e8763ce97579347c"} err="failed to get container status \"19dd2ff55f62b3ee2fd45cc93499c107214cb224273c4fd5e8763ce97579347c\": rpc error: code = NotFound desc = could not find container \"19dd2ff55f62b3ee2fd45cc93499c107214cb224273c4fd5e8763ce97579347c\": container with ID starting with 19dd2ff55f62b3ee2fd45cc93499c107214cb224273c4fd5e8763ce97579347c not found: ID does not exist" Oct 02 03:20:34 crc kubenswrapper[4775]: I1002 03:20:34.005585 4775 scope.go:117] "RemoveContainer" containerID="d815f68fc61e213a829827c4ddb6cc31707b17787817cd510899aefb70c243b1" Oct 02 03:20:34 crc kubenswrapper[4775]: E1002 03:20:34.006217 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d815f68fc61e213a829827c4ddb6cc31707b17787817cd510899aefb70c243b1\": container with ID starting with d815f68fc61e213a829827c4ddb6cc31707b17787817cd510899aefb70c243b1 not found: ID does not exist" containerID="d815f68fc61e213a829827c4ddb6cc31707b17787817cd510899aefb70c243b1" Oct 02 03:20:34 crc kubenswrapper[4775]: I1002 03:20:34.006244 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d815f68fc61e213a829827c4ddb6cc31707b17787817cd510899aefb70c243b1"} err="failed to get container status \"d815f68fc61e213a829827c4ddb6cc31707b17787817cd510899aefb70c243b1\": rpc error: code = NotFound desc = could not find container \"d815f68fc61e213a829827c4ddb6cc31707b17787817cd510899aefb70c243b1\": container with ID starting with d815f68fc61e213a829827c4ddb6cc31707b17787817cd510899aefb70c243b1 not found: ID does not exist" Oct 02 03:20:34 crc kubenswrapper[4775]: I1002 03:20:34.006265 4775 scope.go:117] "RemoveContainer" containerID="d583a650144c43c7a2365cac80e6df9d96dc6e2df101164ab2f393c2b4493f68" Oct 02 03:20:34 crc kubenswrapper[4775]: E1002 03:20:34.006541 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d583a650144c43c7a2365cac80e6df9d96dc6e2df101164ab2f393c2b4493f68\": container with ID starting with d583a650144c43c7a2365cac80e6df9d96dc6e2df101164ab2f393c2b4493f68 not found: ID does not exist" containerID="d583a650144c43c7a2365cac80e6df9d96dc6e2df101164ab2f393c2b4493f68" Oct 02 03:20:34 crc kubenswrapper[4775]: I1002 03:20:34.006558 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d583a650144c43c7a2365cac80e6df9d96dc6e2df101164ab2f393c2b4493f68"} err="failed to get container status \"d583a650144c43c7a2365cac80e6df9d96dc6e2df101164ab2f393c2b4493f68\": rpc error: code = NotFound desc = could not find container \"d583a650144c43c7a2365cac80e6df9d96dc6e2df101164ab2f393c2b4493f68\": container with ID starting with d583a650144c43c7a2365cac80e6df9d96dc6e2df101164ab2f393c2b4493f68 not found: ID does not exist" Oct 02 03:20:34 crc kubenswrapper[4775]: I1002 03:20:34.691242 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-4tlkz" Oct 02 03:20:35 crc kubenswrapper[4775]: I1002 03:20:35.779460 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c9789ad-1c3f-40eb-a012-96551808b45d" path="/var/lib/kubelet/pods/1c9789ad-1c3f-40eb-a012-96551808b45d/volumes" Oct 02 03:20:37 crc kubenswrapper[4775]: I1002 03:20:37.894630 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-crg8r" Oct 02 03:20:43 crc kubenswrapper[4775]: I1002 03:20:43.053999 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6871-account-create-tprh9"] Oct 02 03:20:43 crc kubenswrapper[4775]: I1002 03:20:43.067400 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6871-account-create-tprh9"] Oct 02 03:20:43 crc kubenswrapper[4775]: I1002 03:20:43.783700 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="236c95d5-4c86-40e1-8be3-ddf52e119401" path="/var/lib/kubelet/pods/236c95d5-4c86-40e1-8be3-ddf52e119401/volumes" Oct 02 03:20:47 crc kubenswrapper[4775]: I1002 03:20:47.768240 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:20:48 crc kubenswrapper[4775]: I1002 03:20:48.101529 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"c4930148cbbbfb59b99fb94d29b145a85b1cfbffe3a70db63ae0fd48271fb062"} Oct 02 03:20:52 crc kubenswrapper[4775]: I1002 03:20:52.047821 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-q9nq4"] Oct 02 03:20:52 crc kubenswrapper[4775]: I1002 03:20:52.059522 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-q9nq4"] Oct 02 03:20:53 crc kubenswrapper[4775]: I1002 03:20:53.780682 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56c0f90d-dd66-4bfd-9db2-52fcefd3919d" path="/var/lib/kubelet/pods/56c0f90d-dd66-4bfd-9db2-52fcefd3919d/volumes" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.596081 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nfrbj"] Oct 02 03:21:11 crc kubenswrapper[4775]: E1002 03:21:11.598144 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9789ad-1c3f-40eb-a012-96551808b45d" containerName="extract-content" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.598358 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9789ad-1c3f-40eb-a012-96551808b45d" containerName="extract-content" Oct 02 03:21:11 crc kubenswrapper[4775]: E1002 03:21:11.598392 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9789ad-1c3f-40eb-a012-96551808b45d" containerName="registry-server" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.598403 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9789ad-1c3f-40eb-a012-96551808b45d" containerName="registry-server" Oct 02 03:21:11 crc kubenswrapper[4775]: E1002 03:21:11.598468 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c9789ad-1c3f-40eb-a012-96551808b45d" containerName="extract-utilities" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.598479 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c9789ad-1c3f-40eb-a012-96551808b45d" containerName="extract-utilities" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.598895 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c9789ad-1c3f-40eb-a012-96551808b45d" containerName="registry-server" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.601309 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.610789 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nfrbj"] Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.801158 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkgwp\" (UniqueName: \"kubernetes.io/projected/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-kube-api-access-lkgwp\") pod \"community-operators-nfrbj\" (UID: \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\") " pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.801260 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-catalog-content\") pod \"community-operators-nfrbj\" (UID: \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\") " pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.801322 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-utilities\") pod \"community-operators-nfrbj\" (UID: \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\") " pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.903218 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-utilities\") pod \"community-operators-nfrbj\" (UID: \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\") " pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.903321 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkgwp\" (UniqueName: \"kubernetes.io/projected/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-kube-api-access-lkgwp\") pod \"community-operators-nfrbj\" (UID: \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\") " pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.903419 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-catalog-content\") pod \"community-operators-nfrbj\" (UID: \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\") " pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.903917 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-catalog-content\") pod \"community-operators-nfrbj\" (UID: \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\") " pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.904153 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-utilities\") pod \"community-operators-nfrbj\" (UID: \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\") " pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.927616 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkgwp\" (UniqueName: \"kubernetes.io/projected/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-kube-api-access-lkgwp\") pod \"community-operators-nfrbj\" (UID: \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\") " pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:11 crc kubenswrapper[4775]: I1002 03:21:11.953241 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:12 crc kubenswrapper[4775]: I1002 03:21:12.460455 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nfrbj"] Oct 02 03:21:12 crc kubenswrapper[4775]: I1002 03:21:12.789684 4775 scope.go:117] "RemoveContainer" containerID="e905bf68bfd9cdbb864c08467b80a04191c06dfad7ceaa3f15650d5c6ee2c5dd" Oct 02 03:21:12 crc kubenswrapper[4775]: I1002 03:21:12.824640 4775 scope.go:117] "RemoveContainer" containerID="2e7aa961aab766d93eaa70f5c14beaf43a65f60065d58a23ab6696ce743dfcb1" Oct 02 03:21:12 crc kubenswrapper[4775]: I1002 03:21:12.845630 4775 scope.go:117] "RemoveContainer" containerID="f2ab1daf84a1043ef3c9cd1424fc25b5d2a9bb1398312ed18b0edd2a5951414d" Oct 02 03:21:13 crc kubenswrapper[4775]: I1002 03:21:13.392580 4775 generic.go:334] "Generic (PLEG): container finished" podID="974bcb71-bf1f-47b4-aff4-dfc09a83f99a" containerID="863299895ea7f35f3ebf53e76d572a655f6b5a3cacafd55c4223a68db251bee7" exitCode=0 Oct 02 03:21:13 crc kubenswrapper[4775]: I1002 03:21:13.392644 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nfrbj" event={"ID":"974bcb71-bf1f-47b4-aff4-dfc09a83f99a","Type":"ContainerDied","Data":"863299895ea7f35f3ebf53e76d572a655f6b5a3cacafd55c4223a68db251bee7"} Oct 02 03:21:13 crc kubenswrapper[4775]: I1002 03:21:13.392686 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nfrbj" event={"ID":"974bcb71-bf1f-47b4-aff4-dfc09a83f99a","Type":"ContainerStarted","Data":"043942f7981dda90324c976f848a8cb39ef2a6d2c7e2b105b975f0a8186d8d41"} Oct 02 03:21:13 crc kubenswrapper[4775]: I1002 03:21:13.395841 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 03:21:14 crc kubenswrapper[4775]: I1002 03:21:14.407312 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nfrbj" event={"ID":"974bcb71-bf1f-47b4-aff4-dfc09a83f99a","Type":"ContainerStarted","Data":"c1cef518d3e293c8cd4398846427128a88188b12134e54231980a9540f1eb350"} Oct 02 03:21:15 crc kubenswrapper[4775]: I1002 03:21:15.433220 4775 generic.go:334] "Generic (PLEG): container finished" podID="974bcb71-bf1f-47b4-aff4-dfc09a83f99a" containerID="c1cef518d3e293c8cd4398846427128a88188b12134e54231980a9540f1eb350" exitCode=0 Oct 02 03:21:15 crc kubenswrapper[4775]: I1002 03:21:15.433531 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nfrbj" event={"ID":"974bcb71-bf1f-47b4-aff4-dfc09a83f99a","Type":"ContainerDied","Data":"c1cef518d3e293c8cd4398846427128a88188b12134e54231980a9540f1eb350"} Oct 02 03:21:16 crc kubenswrapper[4775]: I1002 03:21:16.448061 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nfrbj" event={"ID":"974bcb71-bf1f-47b4-aff4-dfc09a83f99a","Type":"ContainerStarted","Data":"adc8cdf6151f351b0757cc8db4c395808282814fdb61eac145f1ea1f79d35216"} Oct 02 03:21:16 crc kubenswrapper[4775]: I1002 03:21:16.471725 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nfrbj" podStartSLOduration=2.995746112 podStartE2EDuration="5.471689909s" podCreationTimestamp="2025-10-02 03:21:11 +0000 UTC" firstStartedPulling="2025-10-02 03:21:13.395468469 +0000 UTC m=+6010.562212539" lastFinishedPulling="2025-10-02 03:21:15.871412256 +0000 UTC m=+6013.038156336" observedRunningTime="2025-10-02 03:21:16.467557352 +0000 UTC m=+6013.634301462" watchObservedRunningTime="2025-10-02 03:21:16.471689909 +0000 UTC m=+6013.638433979" Oct 02 03:21:21 crc kubenswrapper[4775]: I1002 03:21:21.953421 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:21 crc kubenswrapper[4775]: I1002 03:21:21.954110 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:23 crc kubenswrapper[4775]: I1002 03:21:23.023518 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-nfrbj" podUID="974bcb71-bf1f-47b4-aff4-dfc09a83f99a" containerName="registry-server" probeResult="failure" output=< Oct 02 03:21:23 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 03:21:23 crc kubenswrapper[4775]: > Oct 02 03:21:28 crc kubenswrapper[4775]: I1002 03:21:28.468432 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kztcb"] Oct 02 03:21:28 crc kubenswrapper[4775]: I1002 03:21:28.471844 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:28 crc kubenswrapper[4775]: I1002 03:21:28.529292 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/311723d8-6456-4824-9be4-db62813b2c4b-utilities\") pod \"certified-operators-kztcb\" (UID: \"311723d8-6456-4824-9be4-db62813b2c4b\") " pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:28 crc kubenswrapper[4775]: I1002 03:21:28.529398 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/311723d8-6456-4824-9be4-db62813b2c4b-catalog-content\") pod \"certified-operators-kztcb\" (UID: \"311723d8-6456-4824-9be4-db62813b2c4b\") " pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:28 crc kubenswrapper[4775]: I1002 03:21:28.529453 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqpfn\" (UniqueName: \"kubernetes.io/projected/311723d8-6456-4824-9be4-db62813b2c4b-kube-api-access-hqpfn\") pod \"certified-operators-kztcb\" (UID: \"311723d8-6456-4824-9be4-db62813b2c4b\") " pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:28 crc kubenswrapper[4775]: I1002 03:21:28.535035 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kztcb"] Oct 02 03:21:28 crc kubenswrapper[4775]: I1002 03:21:28.631343 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/311723d8-6456-4824-9be4-db62813b2c4b-catalog-content\") pod \"certified-operators-kztcb\" (UID: \"311723d8-6456-4824-9be4-db62813b2c4b\") " pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:28 crc kubenswrapper[4775]: I1002 03:21:28.631635 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqpfn\" (UniqueName: \"kubernetes.io/projected/311723d8-6456-4824-9be4-db62813b2c4b-kube-api-access-hqpfn\") pod \"certified-operators-kztcb\" (UID: \"311723d8-6456-4824-9be4-db62813b2c4b\") " pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:28 crc kubenswrapper[4775]: I1002 03:21:28.631790 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/311723d8-6456-4824-9be4-db62813b2c4b-utilities\") pod \"certified-operators-kztcb\" (UID: \"311723d8-6456-4824-9be4-db62813b2c4b\") " pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:28 crc kubenswrapper[4775]: I1002 03:21:28.632259 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/311723d8-6456-4824-9be4-db62813b2c4b-catalog-content\") pod \"certified-operators-kztcb\" (UID: \"311723d8-6456-4824-9be4-db62813b2c4b\") " pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:28 crc kubenswrapper[4775]: I1002 03:21:28.632386 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/311723d8-6456-4824-9be4-db62813b2c4b-utilities\") pod \"certified-operators-kztcb\" (UID: \"311723d8-6456-4824-9be4-db62813b2c4b\") " pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:28 crc kubenswrapper[4775]: I1002 03:21:28.658730 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqpfn\" (UniqueName: \"kubernetes.io/projected/311723d8-6456-4824-9be4-db62813b2c4b-kube-api-access-hqpfn\") pod \"certified-operators-kztcb\" (UID: \"311723d8-6456-4824-9be4-db62813b2c4b\") " pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:28 crc kubenswrapper[4775]: I1002 03:21:28.846319 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:29 crc kubenswrapper[4775]: I1002 03:21:29.388061 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kztcb"] Oct 02 03:21:29 crc kubenswrapper[4775]: I1002 03:21:29.608591 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kztcb" event={"ID":"311723d8-6456-4824-9be4-db62813b2c4b","Type":"ContainerStarted","Data":"fdf41d70b539190cc67e3177c03271e7b936c27a6020b846a5097d38f3b8be67"} Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.659348 4775 generic.go:334] "Generic (PLEG): container finished" podID="311723d8-6456-4824-9be4-db62813b2c4b" containerID="6778be0ee010740f7735362e649b07d58103e88f5646e5ca3e3c557c08b056dc" exitCode=0 Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.659680 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kztcb" event={"ID":"311723d8-6456-4824-9be4-db62813b2c4b","Type":"ContainerDied","Data":"6778be0ee010740f7735362e649b07d58103e88f5646e5ca3e3c557c08b056dc"} Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.771630 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d58758bb5-g5nvs"] Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.773481 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.775348 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.777558 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.778136 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.782119 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-bmk5n" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.786878 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d58758bb5-g5nvs"] Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.824153 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.824373 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="74fd9de0-3d01-420e-bdc7-35223a06a96d" containerName="glance-log" containerID="cri-o://f1fc960cdde4c98e2edafe7b23cba75f4c05e9d0a8ce499568d09c54eb0a2aeb" gracePeriod=30 Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.824644 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="74fd9de0-3d01-420e-bdc7-35223a06a96d" containerName="glance-httpd" containerID="cri-o://96f0f4427f4e8c46c7165cd632656c506d1272458cf0f0318c2ab2d30d34c098" gracePeriod=30 Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.837107 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-75d677f769-kctl2"] Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.842168 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.860151 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-75d677f769-kctl2"] Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.881087 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60a04c3a-7911-4229-ad11-ce2e56c7c994-config-data\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.881135 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h97hp\" (UniqueName: \"kubernetes.io/projected/60a04c3a-7911-4229-ad11-ce2e56c7c994-kube-api-access-h97hp\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.881980 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60a04c3a-7911-4229-ad11-ce2e56c7c994-horizon-secret-key\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.882065 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60a04c3a-7911-4229-ad11-ce2e56c7c994-scripts\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.882174 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60a04c3a-7911-4229-ad11-ce2e56c7c994-logs\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.911781 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.912046 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a00190d2-bf43-401c-b86c-e52bcbfefcd9" containerName="glance-log" containerID="cri-o://04ee5d869350eecf5f6e6c74e5d7903373815ccb3da1ac919cf2e49347f2eb65" gracePeriod=30 Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.912468 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a00190d2-bf43-401c-b86c-e52bcbfefcd9" containerName="glance-httpd" containerID="cri-o://a3547ba71c1286ae422447403c12040e5231d0ada533c72ef9f64a4e03f8f587" gracePeriod=30 Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.983719 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60a04c3a-7911-4229-ad11-ce2e56c7c994-horizon-secret-key\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.983798 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60a04c3a-7911-4229-ad11-ce2e56c7c994-scripts\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.983857 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/69e1b98c-f22d-44a9-a1bf-69a83ee62930-horizon-secret-key\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.983890 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60a04c3a-7911-4229-ad11-ce2e56c7c994-logs\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.983937 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69e1b98c-f22d-44a9-a1bf-69a83ee62930-scripts\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.984001 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ldkd\" (UniqueName: \"kubernetes.io/projected/69e1b98c-f22d-44a9-a1bf-69a83ee62930-kube-api-access-8ldkd\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.984100 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69e1b98c-f22d-44a9-a1bf-69a83ee62930-logs\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.984165 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60a04c3a-7911-4229-ad11-ce2e56c7c994-config-data\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.984188 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h97hp\" (UniqueName: \"kubernetes.io/projected/60a04c3a-7911-4229-ad11-ce2e56c7c994-kube-api-access-h97hp\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.984211 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69e1b98c-f22d-44a9-a1bf-69a83ee62930-config-data\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.984766 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60a04c3a-7911-4229-ad11-ce2e56c7c994-scripts\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.984884 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60a04c3a-7911-4229-ad11-ce2e56c7c994-logs\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.985878 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60a04c3a-7911-4229-ad11-ce2e56c7c994-config-data\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:30 crc kubenswrapper[4775]: I1002 03:21:30.988761 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60a04c3a-7911-4229-ad11-ce2e56c7c994-horizon-secret-key\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.002198 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h97hp\" (UniqueName: \"kubernetes.io/projected/60a04c3a-7911-4229-ad11-ce2e56c7c994-kube-api-access-h97hp\") pod \"horizon-7d58758bb5-g5nvs\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.085581 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69e1b98c-f22d-44a9-a1bf-69a83ee62930-scripts\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.085636 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ldkd\" (UniqueName: \"kubernetes.io/projected/69e1b98c-f22d-44a9-a1bf-69a83ee62930-kube-api-access-8ldkd\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.085692 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69e1b98c-f22d-44a9-a1bf-69a83ee62930-logs\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.085742 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69e1b98c-f22d-44a9-a1bf-69a83ee62930-config-data\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.085793 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/69e1b98c-f22d-44a9-a1bf-69a83ee62930-horizon-secret-key\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.086794 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69e1b98c-f22d-44a9-a1bf-69a83ee62930-logs\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.087404 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69e1b98c-f22d-44a9-a1bf-69a83ee62930-scripts\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.087456 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69e1b98c-f22d-44a9-a1bf-69a83ee62930-config-data\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.090655 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/69e1b98c-f22d-44a9-a1bf-69a83ee62930-horizon-secret-key\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.093438 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.111680 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ldkd\" (UniqueName: \"kubernetes.io/projected/69e1b98c-f22d-44a9-a1bf-69a83ee62930-kube-api-access-8ldkd\") pod \"horizon-75d677f769-kctl2\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.161494 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.471564 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-75d677f769-kctl2"] Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.516059 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-678b44477-rgx9c"] Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.517505 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.524872 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-678b44477-rgx9c"] Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.610214 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d58758bb5-g5nvs"] Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.670694 4775 generic.go:334] "Generic (PLEG): container finished" podID="74fd9de0-3d01-420e-bdc7-35223a06a96d" containerID="f1fc960cdde4c98e2edafe7b23cba75f4c05e9d0a8ce499568d09c54eb0a2aeb" exitCode=143 Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.670776 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74fd9de0-3d01-420e-bdc7-35223a06a96d","Type":"ContainerDied","Data":"f1fc960cdde4c98e2edafe7b23cba75f4c05e9d0a8ce499568d09c54eb0a2aeb"} Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.672861 4775 generic.go:334] "Generic (PLEG): container finished" podID="311723d8-6456-4824-9be4-db62813b2c4b" containerID="a2fe1422939d79e2468fe7a06d7f6751596b6f2b9f349efdc411f33285879018" exitCode=0 Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.672914 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kztcb" event={"ID":"311723d8-6456-4824-9be4-db62813b2c4b","Type":"ContainerDied","Data":"a2fe1422939d79e2468fe7a06d7f6751596b6f2b9f349efdc411f33285879018"} Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.678042 4775 generic.go:334] "Generic (PLEG): container finished" podID="a00190d2-bf43-401c-b86c-e52bcbfefcd9" containerID="04ee5d869350eecf5f6e6c74e5d7903373815ccb3da1ac919cf2e49347f2eb65" exitCode=143 Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.678083 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a00190d2-bf43-401c-b86c-e52bcbfefcd9","Type":"ContainerDied","Data":"04ee5d869350eecf5f6e6c74e5d7903373815ccb3da1ac919cf2e49347f2eb65"} Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.697601 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98rfd\" (UniqueName: \"kubernetes.io/projected/fa5bfdee-6816-4e3a-b12d-e720bc834be8-kube-api-access-98rfd\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.697647 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fa5bfdee-6816-4e3a-b12d-e720bc834be8-horizon-secret-key\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.697709 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa5bfdee-6816-4e3a-b12d-e720bc834be8-logs\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.697736 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa5bfdee-6816-4e3a-b12d-e720bc834be8-scripts\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.697755 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa5bfdee-6816-4e3a-b12d-e720bc834be8-config-data\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: W1002 03:21:31.703730 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69e1b98c_f22d_44a9_a1bf_69a83ee62930.slice/crio-bcf15176b520860b10d059a19d513fdb60a91f5f0241a3919c69ebcebf87f6f3 WatchSource:0}: Error finding container bcf15176b520860b10d059a19d513fdb60a91f5f0241a3919c69ebcebf87f6f3: Status 404 returned error can't find the container with id bcf15176b520860b10d059a19d513fdb60a91f5f0241a3919c69ebcebf87f6f3 Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.707806 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-75d677f769-kctl2"] Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.799995 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98rfd\" (UniqueName: \"kubernetes.io/projected/fa5bfdee-6816-4e3a-b12d-e720bc834be8-kube-api-access-98rfd\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.800045 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fa5bfdee-6816-4e3a-b12d-e720bc834be8-horizon-secret-key\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.800111 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa5bfdee-6816-4e3a-b12d-e720bc834be8-logs\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.800163 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa5bfdee-6816-4e3a-b12d-e720bc834be8-scripts\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.801877 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa5bfdee-6816-4e3a-b12d-e720bc834be8-logs\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.800354 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa5bfdee-6816-4e3a-b12d-e720bc834be8-config-data\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.802119 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa5bfdee-6816-4e3a-b12d-e720bc834be8-scripts\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.803107 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa5bfdee-6816-4e3a-b12d-e720bc834be8-config-data\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.805819 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fa5bfdee-6816-4e3a-b12d-e720bc834be8-horizon-secret-key\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.821117 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98rfd\" (UniqueName: \"kubernetes.io/projected/fa5bfdee-6816-4e3a-b12d-e720bc834be8-kube-api-access-98rfd\") pod \"horizon-678b44477-rgx9c\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:31 crc kubenswrapper[4775]: I1002 03:21:31.834248 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:32 crc kubenswrapper[4775]: I1002 03:21:32.017742 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:32 crc kubenswrapper[4775]: I1002 03:21:32.080857 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:32 crc kubenswrapper[4775]: I1002 03:21:32.297755 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-678b44477-rgx9c"] Oct 02 03:21:32 crc kubenswrapper[4775]: I1002 03:21:32.691109 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75d677f769-kctl2" event={"ID":"69e1b98c-f22d-44a9-a1bf-69a83ee62930","Type":"ContainerStarted","Data":"bcf15176b520860b10d059a19d513fdb60a91f5f0241a3919c69ebcebf87f6f3"} Oct 02 03:21:32 crc kubenswrapper[4775]: I1002 03:21:32.692184 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-678b44477-rgx9c" event={"ID":"fa5bfdee-6816-4e3a-b12d-e720bc834be8","Type":"ContainerStarted","Data":"6d1790dd50c313380fcfbf2748cb8a7f86cad990fbd85bca6d3e049d188aa807"} Oct 02 03:21:32 crc kubenswrapper[4775]: I1002 03:21:32.696857 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d58758bb5-g5nvs" event={"ID":"60a04c3a-7911-4229-ad11-ce2e56c7c994","Type":"ContainerStarted","Data":"99f0b00c8e106dc11c623933257d1d86c4ec44c91e1e26677ff9131c690df746"} Oct 02 03:21:32 crc kubenswrapper[4775]: I1002 03:21:32.699682 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kztcb" event={"ID":"311723d8-6456-4824-9be4-db62813b2c4b","Type":"ContainerStarted","Data":"837a88f8148ea3e01389b9602b57ea4ae3bbc39f72a6ecca45b582835fe8a9b4"} Oct 02 03:21:32 crc kubenswrapper[4775]: I1002 03:21:32.720781 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kztcb" podStartSLOduration=3.317913672 podStartE2EDuration="4.720764908s" podCreationTimestamp="2025-10-02 03:21:28 +0000 UTC" firstStartedPulling="2025-10-02 03:21:30.664229476 +0000 UTC m=+6027.830973516" lastFinishedPulling="2025-10-02 03:21:32.067080712 +0000 UTC m=+6029.233824752" observedRunningTime="2025-10-02 03:21:32.714517916 +0000 UTC m=+6029.881261966" watchObservedRunningTime="2025-10-02 03:21:32.720764908 +0000 UTC m=+6029.887508948" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.040126 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-8n4dw"] Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.050174 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-8n4dw"] Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.414992 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nfrbj"] Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.415429 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nfrbj" podUID="974bcb71-bf1f-47b4-aff4-dfc09a83f99a" containerName="registry-server" containerID="cri-o://adc8cdf6151f351b0757cc8db4c395808282814fdb61eac145f1ea1f79d35216" gracePeriod=2 Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.643607 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.649932 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.723841 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-config-data\") pod \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.723910 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a00190d2-bf43-401c-b86c-e52bcbfefcd9-logs\") pod \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.723936 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74fd9de0-3d01-420e-bdc7-35223a06a96d-logs\") pod \"74fd9de0-3d01-420e-bdc7-35223a06a96d\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.724058 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/74fd9de0-3d01-420e-bdc7-35223a06a96d-ceph\") pod \"74fd9de0-3d01-420e-bdc7-35223a06a96d\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.724094 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74fd9de0-3d01-420e-bdc7-35223a06a96d-httpd-run\") pod \"74fd9de0-3d01-420e-bdc7-35223a06a96d\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.724109 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-scripts\") pod \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.724138 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-combined-ca-bundle\") pod \"74fd9de0-3d01-420e-bdc7-35223a06a96d\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.724157 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qkvj\" (UniqueName: \"kubernetes.io/projected/74fd9de0-3d01-420e-bdc7-35223a06a96d-kube-api-access-2qkvj\") pod \"74fd9de0-3d01-420e-bdc7-35223a06a96d\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.724297 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jm6b\" (UniqueName: \"kubernetes.io/projected/a00190d2-bf43-401c-b86c-e52bcbfefcd9-kube-api-access-9jm6b\") pod \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.724339 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a00190d2-bf43-401c-b86c-e52bcbfefcd9-ceph\") pod \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.724354 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-scripts\") pod \"74fd9de0-3d01-420e-bdc7-35223a06a96d\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.724458 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a00190d2-bf43-401c-b86c-e52bcbfefcd9-httpd-run\") pod \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.724480 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-combined-ca-bundle\") pod \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\" (UID: \"a00190d2-bf43-401c-b86c-e52bcbfefcd9\") " Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.724552 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-config-data\") pod \"74fd9de0-3d01-420e-bdc7-35223a06a96d\" (UID: \"74fd9de0-3d01-420e-bdc7-35223a06a96d\") " Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.727376 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74fd9de0-3d01-420e-bdc7-35223a06a96d-logs" (OuterVolumeSpecName: "logs") pod "74fd9de0-3d01-420e-bdc7-35223a06a96d" (UID: "74fd9de0-3d01-420e-bdc7-35223a06a96d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.727788 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a00190d2-bf43-401c-b86c-e52bcbfefcd9-logs" (OuterVolumeSpecName: "logs") pod "a00190d2-bf43-401c-b86c-e52bcbfefcd9" (UID: "a00190d2-bf43-401c-b86c-e52bcbfefcd9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.728128 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a00190d2-bf43-401c-b86c-e52bcbfefcd9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a00190d2-bf43-401c-b86c-e52bcbfefcd9" (UID: "a00190d2-bf43-401c-b86c-e52bcbfefcd9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.729664 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74fd9de0-3d01-420e-bdc7-35223a06a96d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "74fd9de0-3d01-420e-bdc7-35223a06a96d" (UID: "74fd9de0-3d01-420e-bdc7-35223a06a96d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.756887 4775 generic.go:334] "Generic (PLEG): container finished" podID="974bcb71-bf1f-47b4-aff4-dfc09a83f99a" containerID="adc8cdf6151f351b0757cc8db4c395808282814fdb61eac145f1ea1f79d35216" exitCode=0 Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.756985 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nfrbj" event={"ID":"974bcb71-bf1f-47b4-aff4-dfc09a83f99a","Type":"ContainerDied","Data":"adc8cdf6151f351b0757cc8db4c395808282814fdb61eac145f1ea1f79d35216"} Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.759697 4775 generic.go:334] "Generic (PLEG): container finished" podID="a00190d2-bf43-401c-b86c-e52bcbfefcd9" containerID="a3547ba71c1286ae422447403c12040e5231d0ada533c72ef9f64a4e03f8f587" exitCode=0 Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.759729 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a00190d2-bf43-401c-b86c-e52bcbfefcd9","Type":"ContainerDied","Data":"a3547ba71c1286ae422447403c12040e5231d0ada533c72ef9f64a4e03f8f587"} Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.759745 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a00190d2-bf43-401c-b86c-e52bcbfefcd9","Type":"ContainerDied","Data":"348e3915a85af1aa181aac83792d404f0f198309b9fde5cdf586f1e1913d7498"} Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.759761 4775 scope.go:117] "RemoveContainer" containerID="a3547ba71c1286ae422447403c12040e5231d0ada533c72ef9f64a4e03f8f587" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.759884 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.770904 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74fd9de0-3d01-420e-bdc7-35223a06a96d-ceph" (OuterVolumeSpecName: "ceph") pod "74fd9de0-3d01-420e-bdc7-35223a06a96d" (UID: "74fd9de0-3d01-420e-bdc7-35223a06a96d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.773709 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74fd9de0-3d01-420e-bdc7-35223a06a96d-kube-api-access-2qkvj" (OuterVolumeSpecName: "kube-api-access-2qkvj") pod "74fd9de0-3d01-420e-bdc7-35223a06a96d" (UID: "74fd9de0-3d01-420e-bdc7-35223a06a96d"). InnerVolumeSpecName "kube-api-access-2qkvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.774315 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a00190d2-bf43-401c-b86c-e52bcbfefcd9-ceph" (OuterVolumeSpecName: "ceph") pod "a00190d2-bf43-401c-b86c-e52bcbfefcd9" (UID: "a00190d2-bf43-401c-b86c-e52bcbfefcd9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.774422 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a00190d2-bf43-401c-b86c-e52bcbfefcd9-kube-api-access-9jm6b" (OuterVolumeSpecName: "kube-api-access-9jm6b") pod "a00190d2-bf43-401c-b86c-e52bcbfefcd9" (UID: "a00190d2-bf43-401c-b86c-e52bcbfefcd9"). InnerVolumeSpecName "kube-api-access-9jm6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.775899 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-scripts" (OuterVolumeSpecName: "scripts") pod "74fd9de0-3d01-420e-bdc7-35223a06a96d" (UID: "74fd9de0-3d01-420e-bdc7-35223a06a96d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.777573 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-scripts" (OuterVolumeSpecName: "scripts") pod "a00190d2-bf43-401c-b86c-e52bcbfefcd9" (UID: "a00190d2-bf43-401c-b86c-e52bcbfefcd9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.782375 4775 generic.go:334] "Generic (PLEG): container finished" podID="74fd9de0-3d01-420e-bdc7-35223a06a96d" containerID="96f0f4427f4e8c46c7165cd632656c506d1272458cf0f0318c2ab2d30d34c098" exitCode=0 Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.782434 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74fd9de0-3d01-420e-bdc7-35223a06a96d","Type":"ContainerDied","Data":"96f0f4427f4e8c46c7165cd632656c506d1272458cf0f0318c2ab2d30d34c098"} Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.782463 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"74fd9de0-3d01-420e-bdc7-35223a06a96d","Type":"ContainerDied","Data":"e88874dd460a17bf43c17f31b2a7dd929bbac257764cc69446615aaf9aa4176a"} Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.782511 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.823936 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-config-data" (OuterVolumeSpecName: "config-data") pod "a00190d2-bf43-401c-b86c-e52bcbfefcd9" (UID: "a00190d2-bf43-401c-b86c-e52bcbfefcd9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.829002 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a00190d2-bf43-401c-b86c-e52bcbfefcd9-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.829099 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.829150 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a00190d2-bf43-401c-b86c-e52bcbfefcd9-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.829200 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74fd9de0-3d01-420e-bdc7-35223a06a96d-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.829248 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/74fd9de0-3d01-420e-bdc7-35223a06a96d-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.829297 4775 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74fd9de0-3d01-420e-bdc7-35223a06a96d-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.829355 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.829404 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qkvj\" (UniqueName: \"kubernetes.io/projected/74fd9de0-3d01-420e-bdc7-35223a06a96d-kube-api-access-2qkvj\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.829454 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jm6b\" (UniqueName: \"kubernetes.io/projected/a00190d2-bf43-401c-b86c-e52bcbfefcd9-kube-api-access-9jm6b\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.829509 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a00190d2-bf43-401c-b86c-e52bcbfefcd9-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.829558 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.850528 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-config-data" (OuterVolumeSpecName: "config-data") pod "74fd9de0-3d01-420e-bdc7-35223a06a96d" (UID: "74fd9de0-3d01-420e-bdc7-35223a06a96d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.853832 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74fd9de0-3d01-420e-bdc7-35223a06a96d" (UID: "74fd9de0-3d01-420e-bdc7-35223a06a96d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.856040 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a00190d2-bf43-401c-b86c-e52bcbfefcd9" (UID: "a00190d2-bf43-401c-b86c-e52bcbfefcd9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.931696 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.931729 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a00190d2-bf43-401c-b86c-e52bcbfefcd9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:34 crc kubenswrapper[4775]: I1002 03:21:34.931738 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74fd9de0-3d01-420e-bdc7-35223a06a96d-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.164693 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.178761 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.199305 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.210708 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.226197 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:21:35 crc kubenswrapper[4775]: E1002 03:21:35.226572 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74fd9de0-3d01-420e-bdc7-35223a06a96d" containerName="glance-log" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.226589 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="74fd9de0-3d01-420e-bdc7-35223a06a96d" containerName="glance-log" Oct 02 03:21:35 crc kubenswrapper[4775]: E1002 03:21:35.226610 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74fd9de0-3d01-420e-bdc7-35223a06a96d" containerName="glance-httpd" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.226618 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="74fd9de0-3d01-420e-bdc7-35223a06a96d" containerName="glance-httpd" Oct 02 03:21:35 crc kubenswrapper[4775]: E1002 03:21:35.226637 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00190d2-bf43-401c-b86c-e52bcbfefcd9" containerName="glance-log" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.226644 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00190d2-bf43-401c-b86c-e52bcbfefcd9" containerName="glance-log" Oct 02 03:21:35 crc kubenswrapper[4775]: E1002 03:21:35.226661 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00190d2-bf43-401c-b86c-e52bcbfefcd9" containerName="glance-httpd" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.226667 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00190d2-bf43-401c-b86c-e52bcbfefcd9" containerName="glance-httpd" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.226852 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="74fd9de0-3d01-420e-bdc7-35223a06a96d" containerName="glance-httpd" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.226867 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a00190d2-bf43-401c-b86c-e52bcbfefcd9" containerName="glance-httpd" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.226886 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="74fd9de0-3d01-420e-bdc7-35223a06a96d" containerName="glance-log" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.226899 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="a00190d2-bf43-401c-b86c-e52bcbfefcd9" containerName="glance-log" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.227939 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.230532 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.231225 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.231334 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ngf6p" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.245362 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.262226 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.273632 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.276772 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.287648 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.342776 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c541a988-a6cf-4625-bc69-5d114ee4aaa8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.343163 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55qrk\" (UniqueName: \"kubernetes.io/projected/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-kube-api-access-55qrk\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.343201 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c541a988-a6cf-4625-bc69-5d114ee4aaa8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.343241 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.343259 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-ceph\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.343282 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8762\" (UniqueName: \"kubernetes.io/projected/c541a988-a6cf-4625-bc69-5d114ee4aaa8-kube-api-access-w8762\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.343300 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c541a988-a6cf-4625-bc69-5d114ee4aaa8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.343340 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c541a988-a6cf-4625-bc69-5d114ee4aaa8-logs\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.343359 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.343390 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-scripts\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.343408 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c541a988-a6cf-4625-bc69-5d114ee4aaa8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.343439 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-logs\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.343469 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c541a988-a6cf-4625-bc69-5d114ee4aaa8-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.343490 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-config-data\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.445163 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.445260 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-ceph\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.445318 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8762\" (UniqueName: \"kubernetes.io/projected/c541a988-a6cf-4625-bc69-5d114ee4aaa8-kube-api-access-w8762\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.445336 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c541a988-a6cf-4625-bc69-5d114ee4aaa8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.445892 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c541a988-a6cf-4625-bc69-5d114ee4aaa8-logs\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.445921 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.445973 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-scripts\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.445993 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c541a988-a6cf-4625-bc69-5d114ee4aaa8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.446025 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-logs\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.446058 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c541a988-a6cf-4625-bc69-5d114ee4aaa8-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.446082 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-config-data\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.446106 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c541a988-a6cf-4625-bc69-5d114ee4aaa8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.446133 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55qrk\" (UniqueName: \"kubernetes.io/projected/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-kube-api-access-55qrk\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.446160 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c541a988-a6cf-4625-bc69-5d114ee4aaa8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.446674 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c541a988-a6cf-4625-bc69-5d114ee4aaa8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.446840 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.447056 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-logs\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.447229 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c541a988-a6cf-4625-bc69-5d114ee4aaa8-logs\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.451489 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c541a988-a6cf-4625-bc69-5d114ee4aaa8-ceph\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.451942 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-ceph\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.452516 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c541a988-a6cf-4625-bc69-5d114ee4aaa8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.453357 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-config-data\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.453419 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-scripts\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.456822 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c541a988-a6cf-4625-bc69-5d114ee4aaa8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.457444 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.457841 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c541a988-a6cf-4625-bc69-5d114ee4aaa8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.459850 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8762\" (UniqueName: \"kubernetes.io/projected/c541a988-a6cf-4625-bc69-5d114ee4aaa8-kube-api-access-w8762\") pod \"glance-default-internal-api-0\" (UID: \"c541a988-a6cf-4625-bc69-5d114ee4aaa8\") " pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.462688 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55qrk\" (UniqueName: \"kubernetes.io/projected/08ba6c24-402e-4dc1-b71a-47cb167ac6f2-kube-api-access-55qrk\") pod \"glance-default-external-api-0\" (UID: \"08ba6c24-402e-4dc1-b71a-47cb167ac6f2\") " pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.554513 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.605710 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.784676 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28461030-4574-48b2-bc21-b782267b22fd" path="/var/lib/kubelet/pods/28461030-4574-48b2-bc21-b782267b22fd/volumes" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.785611 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74fd9de0-3d01-420e-bdc7-35223a06a96d" path="/var/lib/kubelet/pods/74fd9de0-3d01-420e-bdc7-35223a06a96d/volumes" Oct 02 03:21:35 crc kubenswrapper[4775]: I1002 03:21:35.789779 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a00190d2-bf43-401c-b86c-e52bcbfefcd9" path="/var/lib/kubelet/pods/a00190d2-bf43-401c-b86c-e52bcbfefcd9/volumes" Oct 02 03:21:38 crc kubenswrapper[4775]: I1002 03:21:38.847130 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:38 crc kubenswrapper[4775]: I1002 03:21:38.847782 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:38 crc kubenswrapper[4775]: I1002 03:21:38.898501 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:39 crc kubenswrapper[4775]: I1002 03:21:39.929748 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.152196 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kztcb"] Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.223367 4775 scope.go:117] "RemoveContainer" containerID="04ee5d869350eecf5f6e6c74e5d7903373815ccb3da1ac919cf2e49347f2eb65" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.485163 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.529162 4775 scope.go:117] "RemoveContainer" containerID="a3547ba71c1286ae422447403c12040e5231d0ada533c72ef9f64a4e03f8f587" Oct 02 03:21:40 crc kubenswrapper[4775]: E1002 03:21:40.529683 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3547ba71c1286ae422447403c12040e5231d0ada533c72ef9f64a4e03f8f587\": container with ID starting with a3547ba71c1286ae422447403c12040e5231d0ada533c72ef9f64a4e03f8f587 not found: ID does not exist" containerID="a3547ba71c1286ae422447403c12040e5231d0ada533c72ef9f64a4e03f8f587" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.529713 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3547ba71c1286ae422447403c12040e5231d0ada533c72ef9f64a4e03f8f587"} err="failed to get container status \"a3547ba71c1286ae422447403c12040e5231d0ada533c72ef9f64a4e03f8f587\": rpc error: code = NotFound desc = could not find container \"a3547ba71c1286ae422447403c12040e5231d0ada533c72ef9f64a4e03f8f587\": container with ID starting with a3547ba71c1286ae422447403c12040e5231d0ada533c72ef9f64a4e03f8f587 not found: ID does not exist" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.529741 4775 scope.go:117] "RemoveContainer" containerID="04ee5d869350eecf5f6e6c74e5d7903373815ccb3da1ac919cf2e49347f2eb65" Oct 02 03:21:40 crc kubenswrapper[4775]: E1002 03:21:40.530136 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04ee5d869350eecf5f6e6c74e5d7903373815ccb3da1ac919cf2e49347f2eb65\": container with ID starting with 04ee5d869350eecf5f6e6c74e5d7903373815ccb3da1ac919cf2e49347f2eb65 not found: ID does not exist" containerID="04ee5d869350eecf5f6e6c74e5d7903373815ccb3da1ac919cf2e49347f2eb65" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.530176 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04ee5d869350eecf5f6e6c74e5d7903373815ccb3da1ac919cf2e49347f2eb65"} err="failed to get container status \"04ee5d869350eecf5f6e6c74e5d7903373815ccb3da1ac919cf2e49347f2eb65\": rpc error: code = NotFound desc = could not find container \"04ee5d869350eecf5f6e6c74e5d7903373815ccb3da1ac919cf2e49347f2eb65\": container with ID starting with 04ee5d869350eecf5f6e6c74e5d7903373815ccb3da1ac919cf2e49347f2eb65 not found: ID does not exist" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.530211 4775 scope.go:117] "RemoveContainer" containerID="96f0f4427f4e8c46c7165cd632656c506d1272458cf0f0318c2ab2d30d34c098" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.559070 4775 scope.go:117] "RemoveContainer" containerID="f1fc960cdde4c98e2edafe7b23cba75f4c05e9d0a8ce499568d09c54eb0a2aeb" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.569905 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-utilities\") pod \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\" (UID: \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\") " Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.569939 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-catalog-content\") pod \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\" (UID: \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\") " Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.570035 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkgwp\" (UniqueName: \"kubernetes.io/projected/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-kube-api-access-lkgwp\") pod \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\" (UID: \"974bcb71-bf1f-47b4-aff4-dfc09a83f99a\") " Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.570941 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-utilities" (OuterVolumeSpecName: "utilities") pod "974bcb71-bf1f-47b4-aff4-dfc09a83f99a" (UID: "974bcb71-bf1f-47b4-aff4-dfc09a83f99a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.578123 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-kube-api-access-lkgwp" (OuterVolumeSpecName: "kube-api-access-lkgwp") pod "974bcb71-bf1f-47b4-aff4-dfc09a83f99a" (UID: "974bcb71-bf1f-47b4-aff4-dfc09a83f99a"). InnerVolumeSpecName "kube-api-access-lkgwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.586322 4775 scope.go:117] "RemoveContainer" containerID="96f0f4427f4e8c46c7165cd632656c506d1272458cf0f0318c2ab2d30d34c098" Oct 02 03:21:40 crc kubenswrapper[4775]: E1002 03:21:40.586810 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96f0f4427f4e8c46c7165cd632656c506d1272458cf0f0318c2ab2d30d34c098\": container with ID starting with 96f0f4427f4e8c46c7165cd632656c506d1272458cf0f0318c2ab2d30d34c098 not found: ID does not exist" containerID="96f0f4427f4e8c46c7165cd632656c506d1272458cf0f0318c2ab2d30d34c098" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.586848 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96f0f4427f4e8c46c7165cd632656c506d1272458cf0f0318c2ab2d30d34c098"} err="failed to get container status \"96f0f4427f4e8c46c7165cd632656c506d1272458cf0f0318c2ab2d30d34c098\": rpc error: code = NotFound desc = could not find container \"96f0f4427f4e8c46c7165cd632656c506d1272458cf0f0318c2ab2d30d34c098\": container with ID starting with 96f0f4427f4e8c46c7165cd632656c506d1272458cf0f0318c2ab2d30d34c098 not found: ID does not exist" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.586887 4775 scope.go:117] "RemoveContainer" containerID="f1fc960cdde4c98e2edafe7b23cba75f4c05e9d0a8ce499568d09c54eb0a2aeb" Oct 02 03:21:40 crc kubenswrapper[4775]: E1002 03:21:40.587218 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1fc960cdde4c98e2edafe7b23cba75f4c05e9d0a8ce499568d09c54eb0a2aeb\": container with ID starting with f1fc960cdde4c98e2edafe7b23cba75f4c05e9d0a8ce499568d09c54eb0a2aeb not found: ID does not exist" containerID="f1fc960cdde4c98e2edafe7b23cba75f4c05e9d0a8ce499568d09c54eb0a2aeb" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.587237 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1fc960cdde4c98e2edafe7b23cba75f4c05e9d0a8ce499568d09c54eb0a2aeb"} err="failed to get container status \"f1fc960cdde4c98e2edafe7b23cba75f4c05e9d0a8ce499568d09c54eb0a2aeb\": rpc error: code = NotFound desc = could not find container \"f1fc960cdde4c98e2edafe7b23cba75f4c05e9d0a8ce499568d09c54eb0a2aeb\": container with ID starting with f1fc960cdde4c98e2edafe7b23cba75f4c05e9d0a8ce499568d09c54eb0a2aeb not found: ID does not exist" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.630378 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "974bcb71-bf1f-47b4-aff4-dfc09a83f99a" (UID: "974bcb71-bf1f-47b4-aff4-dfc09a83f99a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.672869 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.672896 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.672910 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkgwp\" (UniqueName: \"kubernetes.io/projected/974bcb71-bf1f-47b4-aff4-dfc09a83f99a-kube-api-access-lkgwp\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.856453 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-678b44477-rgx9c" event={"ID":"fa5bfdee-6816-4e3a-b12d-e720bc834be8","Type":"ContainerStarted","Data":"a9a1f6e63b4665842f12b06154913dd76656efb6b696a752765227d6f9efd426"} Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.856516 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-678b44477-rgx9c" event={"ID":"fa5bfdee-6816-4e3a-b12d-e720bc834be8","Type":"ContainerStarted","Data":"5dcf8616f73f8f2882adde2b4a02ca43c6fab0c9724bf4733d7d04335ce88e6c"} Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.873666 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d58758bb5-g5nvs" event={"ID":"60a04c3a-7911-4229-ad11-ce2e56c7c994","Type":"ContainerStarted","Data":"08daed43b5b2c3af8cd8d2ed33d2ab6c8d4215a06a0afa0b06eac6c17223cc76"} Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.882299 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nfrbj" event={"ID":"974bcb71-bf1f-47b4-aff4-dfc09a83f99a","Type":"ContainerDied","Data":"043942f7981dda90324c976f848a8cb39ef2a6d2c7e2b105b975f0a8186d8d41"} Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.882379 4775 scope.go:117] "RemoveContainer" containerID="adc8cdf6151f351b0757cc8db4c395808282814fdb61eac145f1ea1f79d35216" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.882398 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nfrbj" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.888541 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75d677f769-kctl2" event={"ID":"69e1b98c-f22d-44a9-a1bf-69a83ee62930","Type":"ContainerStarted","Data":"64326c41f019909555f9a29048ce5f52f58aa35c55b23859039ccdd6fecbaf50"} Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.888602 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75d677f769-kctl2" event={"ID":"69e1b98c-f22d-44a9-a1bf-69a83ee62930","Type":"ContainerStarted","Data":"093c7b5016f58ac468b44bcde981ff59684d781a1871d62c5b8003687c1c1678"} Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.888851 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-75d677f769-kctl2" podUID="69e1b98c-f22d-44a9-a1bf-69a83ee62930" containerName="horizon-log" containerID="cri-o://093c7b5016f58ac468b44bcde981ff59684d781a1871d62c5b8003687c1c1678" gracePeriod=30 Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.889326 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-75d677f769-kctl2" podUID="69e1b98c-f22d-44a9-a1bf-69a83ee62930" containerName="horizon" containerID="cri-o://64326c41f019909555f9a29048ce5f52f58aa35c55b23859039ccdd6fecbaf50" gracePeriod=30 Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.900332 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-678b44477-rgx9c" podStartSLOduration=1.876495263 podStartE2EDuration="9.900309678s" podCreationTimestamp="2025-10-02 03:21:31 +0000 UTC" firstStartedPulling="2025-10-02 03:21:32.320781815 +0000 UTC m=+6029.487525855" lastFinishedPulling="2025-10-02 03:21:40.34459621 +0000 UTC m=+6037.511340270" observedRunningTime="2025-10-02 03:21:40.88226498 +0000 UTC m=+6038.049009020" watchObservedRunningTime="2025-10-02 03:21:40.900309678 +0000 UTC m=+6038.067053728" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.912514 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-75d677f769-kctl2" podStartSLOduration=2.273974283 podStartE2EDuration="10.912494824s" podCreationTimestamp="2025-10-02 03:21:30 +0000 UTC" firstStartedPulling="2025-10-02 03:21:31.706083199 +0000 UTC m=+6028.872827239" lastFinishedPulling="2025-10-02 03:21:40.34460374 +0000 UTC m=+6037.511347780" observedRunningTime="2025-10-02 03:21:40.909810394 +0000 UTC m=+6038.076554464" watchObservedRunningTime="2025-10-02 03:21:40.912494824 +0000 UTC m=+6038.079238864" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.919811 4775 scope.go:117] "RemoveContainer" containerID="c1cef518d3e293c8cd4398846427128a88188b12134e54231980a9540f1eb350" Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.939183 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nfrbj"] Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.951042 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nfrbj"] Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.956473 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 02 03:21:40 crc kubenswrapper[4775]: I1002 03:21:40.963813 4775 scope.go:117] "RemoveContainer" containerID="863299895ea7f35f3ebf53e76d572a655f6b5a3cacafd55c4223a68db251bee7" Oct 02 03:21:40 crc kubenswrapper[4775]: W1002 03:21:40.972072 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08ba6c24_402e_4dc1_b71a_47cb167ac6f2.slice/crio-128aed669c2097536844beb0f9e11f22abd540cd9dc2f4e8bc3430baa3620431 WatchSource:0}: Error finding container 128aed669c2097536844beb0f9e11f22abd540cd9dc2f4e8bc3430baa3620431: Status 404 returned error can't find the container with id 128aed669c2097536844beb0f9e11f22abd540cd9dc2f4e8bc3430baa3620431 Oct 02 03:21:41 crc kubenswrapper[4775]: I1002 03:21:41.008046 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 02 03:21:41 crc kubenswrapper[4775]: W1002 03:21:41.015029 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc541a988_a6cf_4625_bc69_5d114ee4aaa8.slice/crio-c7647d8474b396d5dde724e370cfd00218f82cd1208cdfbc2a57f69acdff1d86 WatchSource:0}: Error finding container c7647d8474b396d5dde724e370cfd00218f82cd1208cdfbc2a57f69acdff1d86: Status 404 returned error can't find the container with id c7647d8474b396d5dde724e370cfd00218f82cd1208cdfbc2a57f69acdff1d86 Oct 02 03:21:41 crc kubenswrapper[4775]: I1002 03:21:41.162678 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:21:41 crc kubenswrapper[4775]: I1002 03:21:41.782075 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="974bcb71-bf1f-47b4-aff4-dfc09a83f99a" path="/var/lib/kubelet/pods/974bcb71-bf1f-47b4-aff4-dfc09a83f99a/volumes" Oct 02 03:21:41 crc kubenswrapper[4775]: I1002 03:21:41.837194 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:41 crc kubenswrapper[4775]: I1002 03:21:41.837260 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.034461 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d58758bb5-g5nvs" event={"ID":"60a04c3a-7911-4229-ad11-ce2e56c7c994","Type":"ContainerStarted","Data":"c906ec36c1f6343283414031f74bada9b06dccd28fcf9b63b7e95d29c96e0463"} Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.038714 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"08ba6c24-402e-4dc1-b71a-47cb167ac6f2","Type":"ContainerStarted","Data":"cfeda1888ef2f78094194396120e9228337ddb4a5f92a8cc53d49cdd1cd6481c"} Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.038754 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"08ba6c24-402e-4dc1-b71a-47cb167ac6f2","Type":"ContainerStarted","Data":"128aed669c2097536844beb0f9e11f22abd540cd9dc2f4e8bc3430baa3620431"} Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.054124 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c541a988-a6cf-4625-bc69-5d114ee4aaa8","Type":"ContainerStarted","Data":"dace83d87de8e48c41f6d1e02dfe76b9df8f25a4afbe770bf4f98ae00ad2ad6c"} Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.054162 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c541a988-a6cf-4625-bc69-5d114ee4aaa8","Type":"ContainerStarted","Data":"c7647d8474b396d5dde724e370cfd00218f82cd1208cdfbc2a57f69acdff1d86"} Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.054296 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kztcb" podUID="311723d8-6456-4824-9be4-db62813b2c4b" containerName="registry-server" containerID="cri-o://837a88f8148ea3e01389b9602b57ea4ae3bbc39f72a6ecca45b582835fe8a9b4" gracePeriod=2 Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.070643 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7d58758bb5-g5nvs" podStartSLOduration=3.354766574 podStartE2EDuration="12.070620379s" podCreationTimestamp="2025-10-02 03:21:30 +0000 UTC" firstStartedPulling="2025-10-02 03:21:31.667845088 +0000 UTC m=+6028.834589128" lastFinishedPulling="2025-10-02 03:21:40.383698883 +0000 UTC m=+6037.550442933" observedRunningTime="2025-10-02 03:21:42.05523563 +0000 UTC m=+6039.221979680" watchObservedRunningTime="2025-10-02 03:21:42.070620379 +0000 UTC m=+6039.237364409" Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.674870 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.850944 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/311723d8-6456-4824-9be4-db62813b2c4b-utilities\") pod \"311723d8-6456-4824-9be4-db62813b2c4b\" (UID: \"311723d8-6456-4824-9be4-db62813b2c4b\") " Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.851123 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/311723d8-6456-4824-9be4-db62813b2c4b-catalog-content\") pod \"311723d8-6456-4824-9be4-db62813b2c4b\" (UID: \"311723d8-6456-4824-9be4-db62813b2c4b\") " Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.851229 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqpfn\" (UniqueName: \"kubernetes.io/projected/311723d8-6456-4824-9be4-db62813b2c4b-kube-api-access-hqpfn\") pod \"311723d8-6456-4824-9be4-db62813b2c4b\" (UID: \"311723d8-6456-4824-9be4-db62813b2c4b\") " Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.852664 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/311723d8-6456-4824-9be4-db62813b2c4b-utilities" (OuterVolumeSpecName: "utilities") pod "311723d8-6456-4824-9be4-db62813b2c4b" (UID: "311723d8-6456-4824-9be4-db62813b2c4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.860081 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/311723d8-6456-4824-9be4-db62813b2c4b-kube-api-access-hqpfn" (OuterVolumeSpecName: "kube-api-access-hqpfn") pod "311723d8-6456-4824-9be4-db62813b2c4b" (UID: "311723d8-6456-4824-9be4-db62813b2c4b"). InnerVolumeSpecName "kube-api-access-hqpfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.934468 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/311723d8-6456-4824-9be4-db62813b2c4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "311723d8-6456-4824-9be4-db62813b2c4b" (UID: "311723d8-6456-4824-9be4-db62813b2c4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.955004 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqpfn\" (UniqueName: \"kubernetes.io/projected/311723d8-6456-4824-9be4-db62813b2c4b-kube-api-access-hqpfn\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.955037 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/311723d8-6456-4824-9be4-db62813b2c4b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:42 crc kubenswrapper[4775]: I1002 03:21:42.955051 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/311723d8-6456-4824-9be4-db62813b2c4b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.071402 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"08ba6c24-402e-4dc1-b71a-47cb167ac6f2","Type":"ContainerStarted","Data":"99ca8151f580fca5c28487c3fbd9d0bc4fca9b57b63232da86f1e0b7bbaef890"} Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.075415 4775 generic.go:334] "Generic (PLEG): container finished" podID="311723d8-6456-4824-9be4-db62813b2c4b" containerID="837a88f8148ea3e01389b9602b57ea4ae3bbc39f72a6ecca45b582835fe8a9b4" exitCode=0 Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.075477 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kztcb" event={"ID":"311723d8-6456-4824-9be4-db62813b2c4b","Type":"ContainerDied","Data":"837a88f8148ea3e01389b9602b57ea4ae3bbc39f72a6ecca45b582835fe8a9b4"} Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.075508 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kztcb" Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.075531 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kztcb" event={"ID":"311723d8-6456-4824-9be4-db62813b2c4b","Type":"ContainerDied","Data":"fdf41d70b539190cc67e3177c03271e7b936c27a6020b846a5097d38f3b8be67"} Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.075578 4775 scope.go:117] "RemoveContainer" containerID="837a88f8148ea3e01389b9602b57ea4ae3bbc39f72a6ecca45b582835fe8a9b4" Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.079620 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"c541a988-a6cf-4625-bc69-5d114ee4aaa8","Type":"ContainerStarted","Data":"7e3171e03f81b6fe1b30370e6227ca285b571711e7caf853b823fa3d927a542e"} Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.117589 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.117567514 podStartE2EDuration="8.117567514s" podCreationTimestamp="2025-10-02 03:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:21:43.099440934 +0000 UTC m=+6040.266185034" watchObservedRunningTime="2025-10-02 03:21:43.117567514 +0000 UTC m=+6040.284311564" Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.133088 4775 scope.go:117] "RemoveContainer" containerID="a2fe1422939d79e2468fe7a06d7f6751596b6f2b9f349efdc411f33285879018" Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.143708 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.143691601 podStartE2EDuration="8.143691601s" podCreationTimestamp="2025-10-02 03:21:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:21:43.135996202 +0000 UTC m=+6040.302740272" watchObservedRunningTime="2025-10-02 03:21:43.143691601 +0000 UTC m=+6040.310435641" Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.181822 4775 scope.go:117] "RemoveContainer" containerID="6778be0ee010740f7735362e649b07d58103e88f5646e5ca3e3c557c08b056dc" Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.185115 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kztcb"] Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.194699 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kztcb"] Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.225853 4775 scope.go:117] "RemoveContainer" containerID="837a88f8148ea3e01389b9602b57ea4ae3bbc39f72a6ecca45b582835fe8a9b4" Oct 02 03:21:43 crc kubenswrapper[4775]: E1002 03:21:43.226402 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"837a88f8148ea3e01389b9602b57ea4ae3bbc39f72a6ecca45b582835fe8a9b4\": container with ID starting with 837a88f8148ea3e01389b9602b57ea4ae3bbc39f72a6ecca45b582835fe8a9b4 not found: ID does not exist" containerID="837a88f8148ea3e01389b9602b57ea4ae3bbc39f72a6ecca45b582835fe8a9b4" Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.226452 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"837a88f8148ea3e01389b9602b57ea4ae3bbc39f72a6ecca45b582835fe8a9b4"} err="failed to get container status \"837a88f8148ea3e01389b9602b57ea4ae3bbc39f72a6ecca45b582835fe8a9b4\": rpc error: code = NotFound desc = could not find container \"837a88f8148ea3e01389b9602b57ea4ae3bbc39f72a6ecca45b582835fe8a9b4\": container with ID starting with 837a88f8148ea3e01389b9602b57ea4ae3bbc39f72a6ecca45b582835fe8a9b4 not found: ID does not exist" Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.226485 4775 scope.go:117] "RemoveContainer" containerID="a2fe1422939d79e2468fe7a06d7f6751596b6f2b9f349efdc411f33285879018" Oct 02 03:21:43 crc kubenswrapper[4775]: E1002 03:21:43.228486 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2fe1422939d79e2468fe7a06d7f6751596b6f2b9f349efdc411f33285879018\": container with ID starting with a2fe1422939d79e2468fe7a06d7f6751596b6f2b9f349efdc411f33285879018 not found: ID does not exist" containerID="a2fe1422939d79e2468fe7a06d7f6751596b6f2b9f349efdc411f33285879018" Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.228524 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2fe1422939d79e2468fe7a06d7f6751596b6f2b9f349efdc411f33285879018"} err="failed to get container status \"a2fe1422939d79e2468fe7a06d7f6751596b6f2b9f349efdc411f33285879018\": rpc error: code = NotFound desc = could not find container \"a2fe1422939d79e2468fe7a06d7f6751596b6f2b9f349efdc411f33285879018\": container with ID starting with a2fe1422939d79e2468fe7a06d7f6751596b6f2b9f349efdc411f33285879018 not found: ID does not exist" Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.228546 4775 scope.go:117] "RemoveContainer" containerID="6778be0ee010740f7735362e649b07d58103e88f5646e5ca3e3c557c08b056dc" Oct 02 03:21:43 crc kubenswrapper[4775]: E1002 03:21:43.228900 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6778be0ee010740f7735362e649b07d58103e88f5646e5ca3e3c557c08b056dc\": container with ID starting with 6778be0ee010740f7735362e649b07d58103e88f5646e5ca3e3c557c08b056dc not found: ID does not exist" containerID="6778be0ee010740f7735362e649b07d58103e88f5646e5ca3e3c557c08b056dc" Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.228961 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6778be0ee010740f7735362e649b07d58103e88f5646e5ca3e3c557c08b056dc"} err="failed to get container status \"6778be0ee010740f7735362e649b07d58103e88f5646e5ca3e3c557c08b056dc\": rpc error: code = NotFound desc = could not find container \"6778be0ee010740f7735362e649b07d58103e88f5646e5ca3e3c557c08b056dc\": container with ID starting with 6778be0ee010740f7735362e649b07d58103e88f5646e5ca3e3c557c08b056dc not found: ID does not exist" Oct 02 03:21:43 crc kubenswrapper[4775]: I1002 03:21:43.804439 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="311723d8-6456-4824-9be4-db62813b2c4b" path="/var/lib/kubelet/pods/311723d8-6456-4824-9be4-db62813b2c4b/volumes" Oct 02 03:21:45 crc kubenswrapper[4775]: I1002 03:21:45.034489 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-78b4-account-create-qprsk"] Oct 02 03:21:45 crc kubenswrapper[4775]: I1002 03:21:45.043837 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-78b4-account-create-qprsk"] Oct 02 03:21:45 crc kubenswrapper[4775]: I1002 03:21:45.555038 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 03:21:45 crc kubenswrapper[4775]: I1002 03:21:45.556371 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 02 03:21:45 crc kubenswrapper[4775]: I1002 03:21:45.587508 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 03:21:45 crc kubenswrapper[4775]: I1002 03:21:45.606559 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 03:21:45 crc kubenswrapper[4775]: I1002 03:21:45.606599 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 02 03:21:45 crc kubenswrapper[4775]: I1002 03:21:45.616301 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 02 03:21:45 crc kubenswrapper[4775]: I1002 03:21:45.655697 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 03:21:45 crc kubenswrapper[4775]: I1002 03:21:45.682452 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 02 03:21:45 crc kubenswrapper[4775]: I1002 03:21:45.780015 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae0119d1-e4f7-4bcf-b47c-2650f169a4e7" path="/var/lib/kubelet/pods/ae0119d1-e4f7-4bcf-b47c-2650f169a4e7/volumes" Oct 02 03:21:46 crc kubenswrapper[4775]: I1002 03:21:46.115001 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 03:21:46 crc kubenswrapper[4775]: I1002 03:21:46.115032 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 03:21:46 crc kubenswrapper[4775]: I1002 03:21:46.115041 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 02 03:21:46 crc kubenswrapper[4775]: I1002 03:21:46.115049 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 02 03:21:48 crc kubenswrapper[4775]: I1002 03:21:48.067481 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 03:21:48 crc kubenswrapper[4775]: I1002 03:21:48.138546 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 03:21:48 crc kubenswrapper[4775]: I1002 03:21:48.280681 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 03:21:48 crc kubenswrapper[4775]: I1002 03:21:48.954001 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 02 03:21:49 crc kubenswrapper[4775]: I1002 03:21:49.145254 4775 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 02 03:21:49 crc kubenswrapper[4775]: I1002 03:21:49.159571 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 02 03:21:51 crc kubenswrapper[4775]: I1002 03:21:51.094474 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:51 crc kubenswrapper[4775]: I1002 03:21:51.094968 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:21:51 crc kubenswrapper[4775]: I1002 03:21:51.095835 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d58758bb5-g5nvs" podUID="60a04c3a-7911-4229-ad11-ce2e56c7c994" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.112:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.112:8080: connect: connection refused" Oct 02 03:21:51 crc kubenswrapper[4775]: I1002 03:21:51.837100 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-678b44477-rgx9c" podUID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 02 03:21:53 crc kubenswrapper[4775]: I1002 03:21:53.053618 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-sgm76"] Oct 02 03:21:53 crc kubenswrapper[4775]: I1002 03:21:53.063378 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-sgm76"] Oct 02 03:21:53 crc kubenswrapper[4775]: I1002 03:21:53.789150 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a599a786-6215-46c8-8628-34f2cd3aba4f" path="/var/lib/kubelet/pods/a599a786-6215-46c8-8628-34f2cd3aba4f/volumes" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.393398 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fw8sc"] Oct 02 03:21:59 crc kubenswrapper[4775]: E1002 03:21:59.394360 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="974bcb71-bf1f-47b4-aff4-dfc09a83f99a" containerName="extract-utilities" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.394373 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="974bcb71-bf1f-47b4-aff4-dfc09a83f99a" containerName="extract-utilities" Oct 02 03:21:59 crc kubenswrapper[4775]: E1002 03:21:59.394384 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="311723d8-6456-4824-9be4-db62813b2c4b" containerName="registry-server" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.394390 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="311723d8-6456-4824-9be4-db62813b2c4b" containerName="registry-server" Oct 02 03:21:59 crc kubenswrapper[4775]: E1002 03:21:59.394410 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="974bcb71-bf1f-47b4-aff4-dfc09a83f99a" containerName="extract-content" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.394441 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="974bcb71-bf1f-47b4-aff4-dfc09a83f99a" containerName="extract-content" Oct 02 03:21:59 crc kubenswrapper[4775]: E1002 03:21:59.394457 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="311723d8-6456-4824-9be4-db62813b2c4b" containerName="extract-utilities" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.394463 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="311723d8-6456-4824-9be4-db62813b2c4b" containerName="extract-utilities" Oct 02 03:21:59 crc kubenswrapper[4775]: E1002 03:21:59.394477 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="974bcb71-bf1f-47b4-aff4-dfc09a83f99a" containerName="registry-server" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.394483 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="974bcb71-bf1f-47b4-aff4-dfc09a83f99a" containerName="registry-server" Oct 02 03:21:59 crc kubenswrapper[4775]: E1002 03:21:59.394499 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="311723d8-6456-4824-9be4-db62813b2c4b" containerName="extract-content" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.394505 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="311723d8-6456-4824-9be4-db62813b2c4b" containerName="extract-content" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.394689 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="974bcb71-bf1f-47b4-aff4-dfc09a83f99a" containerName="registry-server" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.394699 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="311723d8-6456-4824-9be4-db62813b2c4b" containerName="registry-server" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.396223 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.415664 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fw8sc"] Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.450623 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218a6647-9a67-46cf-ac2e-29d9d7adaa26-utilities\") pod \"redhat-operators-fw8sc\" (UID: \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\") " pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.450712 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xxvx\" (UniqueName: \"kubernetes.io/projected/218a6647-9a67-46cf-ac2e-29d9d7adaa26-kube-api-access-6xxvx\") pod \"redhat-operators-fw8sc\" (UID: \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\") " pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.451784 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218a6647-9a67-46cf-ac2e-29d9d7adaa26-catalog-content\") pod \"redhat-operators-fw8sc\" (UID: \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\") " pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.554138 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218a6647-9a67-46cf-ac2e-29d9d7adaa26-catalog-content\") pod \"redhat-operators-fw8sc\" (UID: \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\") " pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.554295 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218a6647-9a67-46cf-ac2e-29d9d7adaa26-utilities\") pod \"redhat-operators-fw8sc\" (UID: \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\") " pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.554335 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xxvx\" (UniqueName: \"kubernetes.io/projected/218a6647-9a67-46cf-ac2e-29d9d7adaa26-kube-api-access-6xxvx\") pod \"redhat-operators-fw8sc\" (UID: \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\") " pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.554619 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218a6647-9a67-46cf-ac2e-29d9d7adaa26-catalog-content\") pod \"redhat-operators-fw8sc\" (UID: \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\") " pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.554663 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218a6647-9a67-46cf-ac2e-29d9d7adaa26-utilities\") pod \"redhat-operators-fw8sc\" (UID: \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\") " pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.584298 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xxvx\" (UniqueName: \"kubernetes.io/projected/218a6647-9a67-46cf-ac2e-29d9d7adaa26-kube-api-access-6xxvx\") pod \"redhat-operators-fw8sc\" (UID: \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\") " pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:21:59 crc kubenswrapper[4775]: I1002 03:21:59.716655 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:22:00 crc kubenswrapper[4775]: I1002 03:22:00.260297 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fw8sc"] Oct 02 03:22:00 crc kubenswrapper[4775]: W1002 03:22:00.262334 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod218a6647_9a67_46cf_ac2e_29d9d7adaa26.slice/crio-033c5b82ecebb9fc39bdae4fa16c322433319138f587a6f9b145e74ae7dae90c WatchSource:0}: Error finding container 033c5b82ecebb9fc39bdae4fa16c322433319138f587a6f9b145e74ae7dae90c: Status 404 returned error can't find the container with id 033c5b82ecebb9fc39bdae4fa16c322433319138f587a6f9b145e74ae7dae90c Oct 02 03:22:01 crc kubenswrapper[4775]: I1002 03:22:01.271885 4775 generic.go:334] "Generic (PLEG): container finished" podID="218a6647-9a67-46cf-ac2e-29d9d7adaa26" containerID="a88f400c69ba6b4b1c7cb188da2701101bcbb79bb0eaeca6f24d93769e544c46" exitCode=0 Oct 02 03:22:01 crc kubenswrapper[4775]: I1002 03:22:01.272367 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fw8sc" event={"ID":"218a6647-9a67-46cf-ac2e-29d9d7adaa26","Type":"ContainerDied","Data":"a88f400c69ba6b4b1c7cb188da2701101bcbb79bb0eaeca6f24d93769e544c46"} Oct 02 03:22:01 crc kubenswrapper[4775]: I1002 03:22:01.272942 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fw8sc" event={"ID":"218a6647-9a67-46cf-ac2e-29d9d7adaa26","Type":"ContainerStarted","Data":"033c5b82ecebb9fc39bdae4fa16c322433319138f587a6f9b145e74ae7dae90c"} Oct 02 03:22:02 crc kubenswrapper[4775]: I1002 03:22:02.781640 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:22:03 crc kubenswrapper[4775]: I1002 03:22:03.308743 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fw8sc" event={"ID":"218a6647-9a67-46cf-ac2e-29d9d7adaa26","Type":"ContainerStarted","Data":"d4a3591df683883562265c916e04b36c0804c69e3b09d993150e8f888bf2beed"} Oct 02 03:22:03 crc kubenswrapper[4775]: I1002 03:22:03.486390 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:22:04 crc kubenswrapper[4775]: I1002 03:22:04.401159 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:22:05 crc kubenswrapper[4775]: I1002 03:22:05.057139 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:22:05 crc kubenswrapper[4775]: I1002 03:22:05.161925 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d58758bb5-g5nvs"] Oct 02 03:22:05 crc kubenswrapper[4775]: I1002 03:22:05.346533 4775 generic.go:334] "Generic (PLEG): container finished" podID="218a6647-9a67-46cf-ac2e-29d9d7adaa26" containerID="d4a3591df683883562265c916e04b36c0804c69e3b09d993150e8f888bf2beed" exitCode=0 Oct 02 03:22:05 crc kubenswrapper[4775]: I1002 03:22:05.346617 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fw8sc" event={"ID":"218a6647-9a67-46cf-ac2e-29d9d7adaa26","Type":"ContainerDied","Data":"d4a3591df683883562265c916e04b36c0804c69e3b09d993150e8f888bf2beed"} Oct 02 03:22:05 crc kubenswrapper[4775]: I1002 03:22:05.346887 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d58758bb5-g5nvs" podUID="60a04c3a-7911-4229-ad11-ce2e56c7c994" containerName="horizon-log" containerID="cri-o://08daed43b5b2c3af8cd8d2ed33d2ab6c8d4215a06a0afa0b06eac6c17223cc76" gracePeriod=30 Oct 02 03:22:05 crc kubenswrapper[4775]: I1002 03:22:05.347899 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d58758bb5-g5nvs" podUID="60a04c3a-7911-4229-ad11-ce2e56c7c994" containerName="horizon" containerID="cri-o://c906ec36c1f6343283414031f74bada9b06dccd28fcf9b63b7e95d29c96e0463" gracePeriod=30 Oct 02 03:22:06 crc kubenswrapper[4775]: I1002 03:22:06.359124 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fw8sc" event={"ID":"218a6647-9a67-46cf-ac2e-29d9d7adaa26","Type":"ContainerStarted","Data":"a884200792431aaff9d9cd02fde00347607d48a61ac3850906c91dadeae7431b"} Oct 02 03:22:06 crc kubenswrapper[4775]: I1002 03:22:06.397775 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fw8sc" podStartSLOduration=2.784382993 podStartE2EDuration="7.397748508s" podCreationTimestamp="2025-10-02 03:21:59 +0000 UTC" firstStartedPulling="2025-10-02 03:22:01.278122096 +0000 UTC m=+6058.444866156" lastFinishedPulling="2025-10-02 03:22:05.891487621 +0000 UTC m=+6063.058231671" observedRunningTime="2025-10-02 03:22:06.390740406 +0000 UTC m=+6063.557484446" watchObservedRunningTime="2025-10-02 03:22:06.397748508 +0000 UTC m=+6063.564492578" Oct 02 03:22:09 crc kubenswrapper[4775]: I1002 03:22:09.411038 4775 generic.go:334] "Generic (PLEG): container finished" podID="60a04c3a-7911-4229-ad11-ce2e56c7c994" containerID="c906ec36c1f6343283414031f74bada9b06dccd28fcf9b63b7e95d29c96e0463" exitCode=0 Oct 02 03:22:09 crc kubenswrapper[4775]: I1002 03:22:09.411140 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d58758bb5-g5nvs" event={"ID":"60a04c3a-7911-4229-ad11-ce2e56c7c994","Type":"ContainerDied","Data":"c906ec36c1f6343283414031f74bada9b06dccd28fcf9b63b7e95d29c96e0463"} Oct 02 03:22:09 crc kubenswrapper[4775]: I1002 03:22:09.717446 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:22:09 crc kubenswrapper[4775]: I1002 03:22:09.717927 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:22:10 crc kubenswrapper[4775]: I1002 03:22:10.803040 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fw8sc" podUID="218a6647-9a67-46cf-ac2e-29d9d7adaa26" containerName="registry-server" probeResult="failure" output=< Oct 02 03:22:10 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 03:22:10 crc kubenswrapper[4775]: > Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.094000 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7d58758bb5-g5nvs" podUID="60a04c3a-7911-4229-ad11-ce2e56c7c994" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.112:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.112:8080: connect: connection refused" Oct 02 03:22:11 crc kubenswrapper[4775]: E1002 03:22:11.260569 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69e1b98c_f22d_44a9_a1bf_69a83ee62930.slice/crio-conmon-64326c41f019909555f9a29048ce5f52f58aa35c55b23859039ccdd6fecbaf50.scope\": RecentStats: unable to find data in memory cache]" Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.432796 4775 generic.go:334] "Generic (PLEG): container finished" podID="69e1b98c-f22d-44a9-a1bf-69a83ee62930" containerID="64326c41f019909555f9a29048ce5f52f58aa35c55b23859039ccdd6fecbaf50" exitCode=137 Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.433069 4775 generic.go:334] "Generic (PLEG): container finished" podID="69e1b98c-f22d-44a9-a1bf-69a83ee62930" containerID="093c7b5016f58ac468b44bcde981ff59684d781a1871d62c5b8003687c1c1678" exitCode=137 Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.432891 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75d677f769-kctl2" event={"ID":"69e1b98c-f22d-44a9-a1bf-69a83ee62930","Type":"ContainerDied","Data":"64326c41f019909555f9a29048ce5f52f58aa35c55b23859039ccdd6fecbaf50"} Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.433105 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75d677f769-kctl2" event={"ID":"69e1b98c-f22d-44a9-a1bf-69a83ee62930","Type":"ContainerDied","Data":"093c7b5016f58ac468b44bcde981ff59684d781a1871d62c5b8003687c1c1678"} Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.433120 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-75d677f769-kctl2" event={"ID":"69e1b98c-f22d-44a9-a1bf-69a83ee62930","Type":"ContainerDied","Data":"bcf15176b520860b10d059a19d513fdb60a91f5f0241a3919c69ebcebf87f6f3"} Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.433132 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bcf15176b520860b10d059a19d513fdb60a91f5f0241a3919c69ebcebf87f6f3" Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.466227 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.515546 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ldkd\" (UniqueName: \"kubernetes.io/projected/69e1b98c-f22d-44a9-a1bf-69a83ee62930-kube-api-access-8ldkd\") pod \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.515602 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/69e1b98c-f22d-44a9-a1bf-69a83ee62930-horizon-secret-key\") pod \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.515661 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69e1b98c-f22d-44a9-a1bf-69a83ee62930-logs\") pod \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.515782 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69e1b98c-f22d-44a9-a1bf-69a83ee62930-scripts\") pod \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.515973 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69e1b98c-f22d-44a9-a1bf-69a83ee62930-config-data\") pod \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\" (UID: \"69e1b98c-f22d-44a9-a1bf-69a83ee62930\") " Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.527540 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69e1b98c-f22d-44a9-a1bf-69a83ee62930-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "69e1b98c-f22d-44a9-a1bf-69a83ee62930" (UID: "69e1b98c-f22d-44a9-a1bf-69a83ee62930"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.529166 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69e1b98c-f22d-44a9-a1bf-69a83ee62930-kube-api-access-8ldkd" (OuterVolumeSpecName: "kube-api-access-8ldkd") pod "69e1b98c-f22d-44a9-a1bf-69a83ee62930" (UID: "69e1b98c-f22d-44a9-a1bf-69a83ee62930"). InnerVolumeSpecName "kube-api-access-8ldkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.529804 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/69e1b98c-f22d-44a9-a1bf-69a83ee62930-logs" (OuterVolumeSpecName: "logs") pod "69e1b98c-f22d-44a9-a1bf-69a83ee62930" (UID: "69e1b98c-f22d-44a9-a1bf-69a83ee62930"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.558210 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69e1b98c-f22d-44a9-a1bf-69a83ee62930-scripts" (OuterVolumeSpecName: "scripts") pod "69e1b98c-f22d-44a9-a1bf-69a83ee62930" (UID: "69e1b98c-f22d-44a9-a1bf-69a83ee62930"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.564684 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/69e1b98c-f22d-44a9-a1bf-69a83ee62930-config-data" (OuterVolumeSpecName: "config-data") pod "69e1b98c-f22d-44a9-a1bf-69a83ee62930" (UID: "69e1b98c-f22d-44a9-a1bf-69a83ee62930"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.619103 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/69e1b98c-f22d-44a9-a1bf-69a83ee62930-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.619145 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ldkd\" (UniqueName: \"kubernetes.io/projected/69e1b98c-f22d-44a9-a1bf-69a83ee62930-kube-api-access-8ldkd\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.619161 4775 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/69e1b98c-f22d-44a9-a1bf-69a83ee62930-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.619173 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/69e1b98c-f22d-44a9-a1bf-69a83ee62930-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:11 crc kubenswrapper[4775]: I1002 03:22:11.619185 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/69e1b98c-f22d-44a9-a1bf-69a83ee62930-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:12 crc kubenswrapper[4775]: I1002 03:22:12.442484 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-75d677f769-kctl2" Oct 02 03:22:12 crc kubenswrapper[4775]: I1002 03:22:12.473190 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-75d677f769-kctl2"] Oct 02 03:22:12 crc kubenswrapper[4775]: I1002 03:22:12.479651 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-75d677f769-kctl2"] Oct 02 03:22:12 crc kubenswrapper[4775]: I1002 03:22:12.975930 4775 scope.go:117] "RemoveContainer" containerID="01888fcf624bb0d784c04b9a18d9ae756e2805a8eafc6ab7ee3f1f6d4ac35205" Oct 02 03:22:13 crc kubenswrapper[4775]: I1002 03:22:13.003898 4775 scope.go:117] "RemoveContainer" containerID="114977fc7487ad08c77de37183e952cd3c0a1f3d5556a334a3ab56b30ef5e73b" Oct 02 03:22:13 crc kubenswrapper[4775]: I1002 03:22:13.053889 4775 scope.go:117] "RemoveContainer" containerID="f21a152c6724d033fb12f1a28fba8be83d85d2152ccf9e7ec09d12eb09e5ddd2" Oct 02 03:22:13 crc kubenswrapper[4775]: I1002 03:22:13.126022 4775 scope.go:117] "RemoveContainer" containerID="0f868e320aa80c1ea6a82a0589215ec5bbc223f3e0310275dc59bf61e150b317" Oct 02 03:22:13 crc kubenswrapper[4775]: I1002 03:22:13.195014 4775 scope.go:117] "RemoveContainer" containerID="9a77ab16b75e89acc319f1545b286cbeb894e8d0f4f377c5bd36b4a679a8dfe2" Oct 02 03:22:13 crc kubenswrapper[4775]: I1002 03:22:13.785136 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69e1b98c-f22d-44a9-a1bf-69a83ee62930" path="/var/lib/kubelet/pods/69e1b98c-f22d-44a9-a1bf-69a83ee62930/volumes" Oct 02 03:22:19 crc kubenswrapper[4775]: I1002 03:22:19.785083 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:22:19 crc kubenswrapper[4775]: I1002 03:22:19.846059 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:22:20 crc kubenswrapper[4775]: I1002 03:22:20.028734 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fw8sc"] Oct 02 03:22:21 crc kubenswrapper[4775]: I1002 03:22:21.094124 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7d58758bb5-g5nvs" podUID="60a04c3a-7911-4229-ad11-ce2e56c7c994" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.112:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.112:8080: connect: connection refused" Oct 02 03:22:21 crc kubenswrapper[4775]: I1002 03:22:21.552049 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fw8sc" podUID="218a6647-9a67-46cf-ac2e-29d9d7adaa26" containerName="registry-server" containerID="cri-o://a884200792431aaff9d9cd02fde00347607d48a61ac3850906c91dadeae7431b" gracePeriod=2 Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.039878 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.097770 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xxvx\" (UniqueName: \"kubernetes.io/projected/218a6647-9a67-46cf-ac2e-29d9d7adaa26-kube-api-access-6xxvx\") pod \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\" (UID: \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\") " Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.097856 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218a6647-9a67-46cf-ac2e-29d9d7adaa26-catalog-content\") pod \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\" (UID: \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\") " Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.098014 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218a6647-9a67-46cf-ac2e-29d9d7adaa26-utilities\") pod \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\" (UID: \"218a6647-9a67-46cf-ac2e-29d9d7adaa26\") " Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.100256 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/218a6647-9a67-46cf-ac2e-29d9d7adaa26-utilities" (OuterVolumeSpecName: "utilities") pod "218a6647-9a67-46cf-ac2e-29d9d7adaa26" (UID: "218a6647-9a67-46cf-ac2e-29d9d7adaa26"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.120205 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/218a6647-9a67-46cf-ac2e-29d9d7adaa26-kube-api-access-6xxvx" (OuterVolumeSpecName: "kube-api-access-6xxvx") pod "218a6647-9a67-46cf-ac2e-29d9d7adaa26" (UID: "218a6647-9a67-46cf-ac2e-29d9d7adaa26"). InnerVolumeSpecName "kube-api-access-6xxvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.199901 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xxvx\" (UniqueName: \"kubernetes.io/projected/218a6647-9a67-46cf-ac2e-29d9d7adaa26-kube-api-access-6xxvx\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.199932 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/218a6647-9a67-46cf-ac2e-29d9d7adaa26-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.204872 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/218a6647-9a67-46cf-ac2e-29d9d7adaa26-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "218a6647-9a67-46cf-ac2e-29d9d7adaa26" (UID: "218a6647-9a67-46cf-ac2e-29d9d7adaa26"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.301530 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/218a6647-9a67-46cf-ac2e-29d9d7adaa26-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.569113 4775 generic.go:334] "Generic (PLEG): container finished" podID="218a6647-9a67-46cf-ac2e-29d9d7adaa26" containerID="a884200792431aaff9d9cd02fde00347607d48a61ac3850906c91dadeae7431b" exitCode=0 Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.569204 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fw8sc" event={"ID":"218a6647-9a67-46cf-ac2e-29d9d7adaa26","Type":"ContainerDied","Data":"a884200792431aaff9d9cd02fde00347607d48a61ac3850906c91dadeae7431b"} Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.569241 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fw8sc" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.569319 4775 scope.go:117] "RemoveContainer" containerID="a884200792431aaff9d9cd02fde00347607d48a61ac3850906c91dadeae7431b" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.569296 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fw8sc" event={"ID":"218a6647-9a67-46cf-ac2e-29d9d7adaa26","Type":"ContainerDied","Data":"033c5b82ecebb9fc39bdae4fa16c322433319138f587a6f9b145e74ae7dae90c"} Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.621279 4775 scope.go:117] "RemoveContainer" containerID="d4a3591df683883562265c916e04b36c0804c69e3b09d993150e8f888bf2beed" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.631356 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fw8sc"] Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.644924 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fw8sc"] Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.672992 4775 scope.go:117] "RemoveContainer" containerID="a88f400c69ba6b4b1c7cb188da2701101bcbb79bb0eaeca6f24d93769e544c46" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.714099 4775 scope.go:117] "RemoveContainer" containerID="a884200792431aaff9d9cd02fde00347607d48a61ac3850906c91dadeae7431b" Oct 02 03:22:22 crc kubenswrapper[4775]: E1002 03:22:22.714705 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a884200792431aaff9d9cd02fde00347607d48a61ac3850906c91dadeae7431b\": container with ID starting with a884200792431aaff9d9cd02fde00347607d48a61ac3850906c91dadeae7431b not found: ID does not exist" containerID="a884200792431aaff9d9cd02fde00347607d48a61ac3850906c91dadeae7431b" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.714792 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a884200792431aaff9d9cd02fde00347607d48a61ac3850906c91dadeae7431b"} err="failed to get container status \"a884200792431aaff9d9cd02fde00347607d48a61ac3850906c91dadeae7431b\": rpc error: code = NotFound desc = could not find container \"a884200792431aaff9d9cd02fde00347607d48a61ac3850906c91dadeae7431b\": container with ID starting with a884200792431aaff9d9cd02fde00347607d48a61ac3850906c91dadeae7431b not found: ID does not exist" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.714837 4775 scope.go:117] "RemoveContainer" containerID="d4a3591df683883562265c916e04b36c0804c69e3b09d993150e8f888bf2beed" Oct 02 03:22:22 crc kubenswrapper[4775]: E1002 03:22:22.715459 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4a3591df683883562265c916e04b36c0804c69e3b09d993150e8f888bf2beed\": container with ID starting with d4a3591df683883562265c916e04b36c0804c69e3b09d993150e8f888bf2beed not found: ID does not exist" containerID="d4a3591df683883562265c916e04b36c0804c69e3b09d993150e8f888bf2beed" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.715512 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4a3591df683883562265c916e04b36c0804c69e3b09d993150e8f888bf2beed"} err="failed to get container status \"d4a3591df683883562265c916e04b36c0804c69e3b09d993150e8f888bf2beed\": rpc error: code = NotFound desc = could not find container \"d4a3591df683883562265c916e04b36c0804c69e3b09d993150e8f888bf2beed\": container with ID starting with d4a3591df683883562265c916e04b36c0804c69e3b09d993150e8f888bf2beed not found: ID does not exist" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.715552 4775 scope.go:117] "RemoveContainer" containerID="a88f400c69ba6b4b1c7cb188da2701101bcbb79bb0eaeca6f24d93769e544c46" Oct 02 03:22:22 crc kubenswrapper[4775]: E1002 03:22:22.716187 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a88f400c69ba6b4b1c7cb188da2701101bcbb79bb0eaeca6f24d93769e544c46\": container with ID starting with a88f400c69ba6b4b1c7cb188da2701101bcbb79bb0eaeca6f24d93769e544c46 not found: ID does not exist" containerID="a88f400c69ba6b4b1c7cb188da2701101bcbb79bb0eaeca6f24d93769e544c46" Oct 02 03:22:22 crc kubenswrapper[4775]: I1002 03:22:22.716235 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a88f400c69ba6b4b1c7cb188da2701101bcbb79bb0eaeca6f24d93769e544c46"} err="failed to get container status \"a88f400c69ba6b4b1c7cb188da2701101bcbb79bb0eaeca6f24d93769e544c46\": rpc error: code = NotFound desc = could not find container \"a88f400c69ba6b4b1c7cb188da2701101bcbb79bb0eaeca6f24d93769e544c46\": container with ID starting with a88f400c69ba6b4b1c7cb188da2701101bcbb79bb0eaeca6f24d93769e544c46 not found: ID does not exist" Oct 02 03:22:23 crc kubenswrapper[4775]: I1002 03:22:23.073855 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-gzf2f"] Oct 02 03:22:23 crc kubenswrapper[4775]: I1002 03:22:23.085877 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-gzf2f"] Oct 02 03:22:23 crc kubenswrapper[4775]: I1002 03:22:23.786730 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="218a6647-9a67-46cf-ac2e-29d9d7adaa26" path="/var/lib/kubelet/pods/218a6647-9a67-46cf-ac2e-29d9d7adaa26/volumes" Oct 02 03:22:23 crc kubenswrapper[4775]: I1002 03:22:23.790153 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="436596a4-5708-4e75-9e51-cf45a2600dc4" path="/var/lib/kubelet/pods/436596a4-5708-4e75-9e51-cf45a2600dc4/volumes" Oct 02 03:22:31 crc kubenswrapper[4775]: I1002 03:22:31.094723 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7d58758bb5-g5nvs" podUID="60a04c3a-7911-4229-ad11-ce2e56c7c994" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.112:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.112:8080: connect: connection refused" Oct 02 03:22:31 crc kubenswrapper[4775]: I1002 03:22:31.095587 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:22:33 crc kubenswrapper[4775]: I1002 03:22:33.034309 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-8804-account-create-vkrcs"] Oct 02 03:22:33 crc kubenswrapper[4775]: I1002 03:22:33.047670 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-8804-account-create-vkrcs"] Oct 02 03:22:33 crc kubenswrapper[4775]: I1002 03:22:33.783682 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e9ed2ad-06c2-4e53-9359-2c26438ae139" path="/var/lib/kubelet/pods/7e9ed2ad-06c2-4e53-9359-2c26438ae139/volumes" Oct 02 03:22:35 crc kubenswrapper[4775]: I1002 03:22:35.749338 4775 generic.go:334] "Generic (PLEG): container finished" podID="60a04c3a-7911-4229-ad11-ce2e56c7c994" containerID="08daed43b5b2c3af8cd8d2ed33d2ab6c8d4215a06a0afa0b06eac6c17223cc76" exitCode=137 Oct 02 03:22:35 crc kubenswrapper[4775]: I1002 03:22:35.750265 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d58758bb5-g5nvs" event={"ID":"60a04c3a-7911-4229-ad11-ce2e56c7c994","Type":"ContainerDied","Data":"08daed43b5b2c3af8cd8d2ed33d2ab6c8d4215a06a0afa0b06eac6c17223cc76"} Oct 02 03:22:35 crc kubenswrapper[4775]: I1002 03:22:35.750307 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d58758bb5-g5nvs" event={"ID":"60a04c3a-7911-4229-ad11-ce2e56c7c994","Type":"ContainerDied","Data":"99f0b00c8e106dc11c623933257d1d86c4ec44c91e1e26677ff9131c690df746"} Oct 02 03:22:35 crc kubenswrapper[4775]: I1002 03:22:35.750326 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99f0b00c8e106dc11c623933257d1d86c4ec44c91e1e26677ff9131c690df746" Oct 02 03:22:35 crc kubenswrapper[4775]: I1002 03:22:35.856643 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.037112 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60a04c3a-7911-4229-ad11-ce2e56c7c994-horizon-secret-key\") pod \"60a04c3a-7911-4229-ad11-ce2e56c7c994\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.037353 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60a04c3a-7911-4229-ad11-ce2e56c7c994-config-data\") pod \"60a04c3a-7911-4229-ad11-ce2e56c7c994\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.037392 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60a04c3a-7911-4229-ad11-ce2e56c7c994-scripts\") pod \"60a04c3a-7911-4229-ad11-ce2e56c7c994\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.037535 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60a04c3a-7911-4229-ad11-ce2e56c7c994-logs\") pod \"60a04c3a-7911-4229-ad11-ce2e56c7c994\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.037622 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h97hp\" (UniqueName: \"kubernetes.io/projected/60a04c3a-7911-4229-ad11-ce2e56c7c994-kube-api-access-h97hp\") pod \"60a04c3a-7911-4229-ad11-ce2e56c7c994\" (UID: \"60a04c3a-7911-4229-ad11-ce2e56c7c994\") " Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.038235 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60a04c3a-7911-4229-ad11-ce2e56c7c994-logs" (OuterVolumeSpecName: "logs") pod "60a04c3a-7911-4229-ad11-ce2e56c7c994" (UID: "60a04c3a-7911-4229-ad11-ce2e56c7c994"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.046113 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60a04c3a-7911-4229-ad11-ce2e56c7c994-kube-api-access-h97hp" (OuterVolumeSpecName: "kube-api-access-h97hp") pod "60a04c3a-7911-4229-ad11-ce2e56c7c994" (UID: "60a04c3a-7911-4229-ad11-ce2e56c7c994"). InnerVolumeSpecName "kube-api-access-h97hp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.055934 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60a04c3a-7911-4229-ad11-ce2e56c7c994-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "60a04c3a-7911-4229-ad11-ce2e56c7c994" (UID: "60a04c3a-7911-4229-ad11-ce2e56c7c994"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.085579 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60a04c3a-7911-4229-ad11-ce2e56c7c994-config-data" (OuterVolumeSpecName: "config-data") pod "60a04c3a-7911-4229-ad11-ce2e56c7c994" (UID: "60a04c3a-7911-4229-ad11-ce2e56c7c994"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.091390 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60a04c3a-7911-4229-ad11-ce2e56c7c994-scripts" (OuterVolumeSpecName: "scripts") pod "60a04c3a-7911-4229-ad11-ce2e56c7c994" (UID: "60a04c3a-7911-4229-ad11-ce2e56c7c994"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.139733 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/60a04c3a-7911-4229-ad11-ce2e56c7c994-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.139765 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/60a04c3a-7911-4229-ad11-ce2e56c7c994-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.139774 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60a04c3a-7911-4229-ad11-ce2e56c7c994-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.139783 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h97hp\" (UniqueName: \"kubernetes.io/projected/60a04c3a-7911-4229-ad11-ce2e56c7c994-kube-api-access-h97hp\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.139792 4775 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/60a04c3a-7911-4229-ad11-ce2e56c7c994-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.761019 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d58758bb5-g5nvs" Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.794833 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d58758bb5-g5nvs"] Oct 02 03:22:36 crc kubenswrapper[4775]: I1002 03:22:36.801289 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7d58758bb5-g5nvs"] Oct 02 03:22:37 crc kubenswrapper[4775]: I1002 03:22:37.802204 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60a04c3a-7911-4229-ad11-ce2e56c7c994" path="/var/lib/kubelet/pods/60a04c3a-7911-4229-ad11-ce2e56c7c994/volumes" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.039318 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-v4zst"] Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.050094 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-v4zst"] Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.782814 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afdda688-8e5c-49e4-83fc-1ba569359404" path="/var/lib/kubelet/pods/afdda688-8e5c-49e4-83fc-1ba569359404/volumes" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.946805 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-598f77fd5-6jh6q"] Oct 02 03:22:39 crc kubenswrapper[4775]: E1002 03:22:39.947209 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60a04c3a-7911-4229-ad11-ce2e56c7c994" containerName="horizon-log" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.947229 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="60a04c3a-7911-4229-ad11-ce2e56c7c994" containerName="horizon-log" Oct 02 03:22:39 crc kubenswrapper[4775]: E1002 03:22:39.947243 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69e1b98c-f22d-44a9-a1bf-69a83ee62930" containerName="horizon" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.947249 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="69e1b98c-f22d-44a9-a1bf-69a83ee62930" containerName="horizon" Oct 02 03:22:39 crc kubenswrapper[4775]: E1002 03:22:39.947259 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218a6647-9a67-46cf-ac2e-29d9d7adaa26" containerName="extract-content" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.947266 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="218a6647-9a67-46cf-ac2e-29d9d7adaa26" containerName="extract-content" Oct 02 03:22:39 crc kubenswrapper[4775]: E1002 03:22:39.947280 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218a6647-9a67-46cf-ac2e-29d9d7adaa26" containerName="registry-server" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.947285 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="218a6647-9a67-46cf-ac2e-29d9d7adaa26" containerName="registry-server" Oct 02 03:22:39 crc kubenswrapper[4775]: E1002 03:22:39.947302 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69e1b98c-f22d-44a9-a1bf-69a83ee62930" containerName="horizon-log" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.947308 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="69e1b98c-f22d-44a9-a1bf-69a83ee62930" containerName="horizon-log" Oct 02 03:22:39 crc kubenswrapper[4775]: E1002 03:22:39.947323 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60a04c3a-7911-4229-ad11-ce2e56c7c994" containerName="horizon" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.947330 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="60a04c3a-7911-4229-ad11-ce2e56c7c994" containerName="horizon" Oct 02 03:22:39 crc kubenswrapper[4775]: E1002 03:22:39.947349 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="218a6647-9a67-46cf-ac2e-29d9d7adaa26" containerName="extract-utilities" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.947355 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="218a6647-9a67-46cf-ac2e-29d9d7adaa26" containerName="extract-utilities" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.947530 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="69e1b98c-f22d-44a9-a1bf-69a83ee62930" containerName="horizon-log" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.947540 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="60a04c3a-7911-4229-ad11-ce2e56c7c994" containerName="horizon-log" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.947556 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="218a6647-9a67-46cf-ac2e-29d9d7adaa26" containerName="registry-server" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.947566 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="60a04c3a-7911-4229-ad11-ce2e56c7c994" containerName="horizon" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.947589 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="69e1b98c-f22d-44a9-a1bf-69a83ee62930" containerName="horizon" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.948620 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:39 crc kubenswrapper[4775]: I1002 03:22:39.970816 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-598f77fd5-6jh6q"] Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.126256 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2d29\" (UniqueName: \"kubernetes.io/projected/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-kube-api-access-r2d29\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.126331 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-horizon-secret-key\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.126440 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-scripts\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.126653 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-logs\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.126890 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-config-data\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.228839 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-config-data\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.228926 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2d29\" (UniqueName: \"kubernetes.io/projected/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-kube-api-access-r2d29\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.228982 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-horizon-secret-key\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.229060 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-scripts\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.229113 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-logs\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.229671 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-logs\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.230108 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-scripts\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.231117 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-config-data\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.237407 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-horizon-secret-key\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.245067 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2d29\" (UniqueName: \"kubernetes.io/projected/2b046c0b-b2f9-4808-84fa-5474dd3f6d19-kube-api-access-r2d29\") pod \"horizon-598f77fd5-6jh6q\" (UID: \"2b046c0b-b2f9-4808-84fa-5474dd3f6d19\") " pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.267629 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:40 crc kubenswrapper[4775]: I1002 03:22:40.827414 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-598f77fd5-6jh6q"] Oct 02 03:22:41 crc kubenswrapper[4775]: I1002 03:22:41.527118 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-ckn24"] Oct 02 03:22:41 crc kubenswrapper[4775]: I1002 03:22:41.529164 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-ckn24" Oct 02 03:22:41 crc kubenswrapper[4775]: I1002 03:22:41.536360 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-ckn24"] Oct 02 03:22:41 crc kubenswrapper[4775]: I1002 03:22:41.666578 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qqnf\" (UniqueName: \"kubernetes.io/projected/2c9ebe75-71af-415f-b74a-d24d0e677411-kube-api-access-7qqnf\") pod \"heat-db-create-ckn24\" (UID: \"2c9ebe75-71af-415f-b74a-d24d0e677411\") " pod="openstack/heat-db-create-ckn24" Oct 02 03:22:41 crc kubenswrapper[4775]: I1002 03:22:41.769297 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qqnf\" (UniqueName: \"kubernetes.io/projected/2c9ebe75-71af-415f-b74a-d24d0e677411-kube-api-access-7qqnf\") pod \"heat-db-create-ckn24\" (UID: \"2c9ebe75-71af-415f-b74a-d24d0e677411\") " pod="openstack/heat-db-create-ckn24" Oct 02 03:22:41 crc kubenswrapper[4775]: I1002 03:22:41.785529 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qqnf\" (UniqueName: \"kubernetes.io/projected/2c9ebe75-71af-415f-b74a-d24d0e677411-kube-api-access-7qqnf\") pod \"heat-db-create-ckn24\" (UID: \"2c9ebe75-71af-415f-b74a-d24d0e677411\") " pod="openstack/heat-db-create-ckn24" Oct 02 03:22:41 crc kubenswrapper[4775]: I1002 03:22:41.813521 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-598f77fd5-6jh6q" event={"ID":"2b046c0b-b2f9-4808-84fa-5474dd3f6d19","Type":"ContainerStarted","Data":"1db468bf901eba1cd08dab7db9e9738746ac1f51c50e9f691825351ab32df1f9"} Oct 02 03:22:41 crc kubenswrapper[4775]: I1002 03:22:41.813561 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-598f77fd5-6jh6q" event={"ID":"2b046c0b-b2f9-4808-84fa-5474dd3f6d19","Type":"ContainerStarted","Data":"6c85201a2e81623edd64d47668d9cd16c69f6dacebaa7894fe115aa14258d681"} Oct 02 03:22:41 crc kubenswrapper[4775]: I1002 03:22:41.813571 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-598f77fd5-6jh6q" event={"ID":"2b046c0b-b2f9-4808-84fa-5474dd3f6d19","Type":"ContainerStarted","Data":"2b5018e172d1a87e648fbac02708220c98fb88af7ad4582179327250d0550167"} Oct 02 03:22:41 crc kubenswrapper[4775]: I1002 03:22:41.846574 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-598f77fd5-6jh6q" podStartSLOduration=2.846552736 podStartE2EDuration="2.846552736s" podCreationTimestamp="2025-10-02 03:22:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:22:41.83512775 +0000 UTC m=+6099.001871820" watchObservedRunningTime="2025-10-02 03:22:41.846552736 +0000 UTC m=+6099.013296786" Oct 02 03:22:41 crc kubenswrapper[4775]: I1002 03:22:41.887439 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-ckn24" Oct 02 03:22:42 crc kubenswrapper[4775]: I1002 03:22:42.371882 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-ckn24"] Oct 02 03:22:42 crc kubenswrapper[4775]: W1002 03:22:42.376527 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c9ebe75_71af_415f_b74a_d24d0e677411.slice/crio-faa002b610b0fc0441a885d9b5e3a1d86e9be95a0be26cc0d3f9949bd021eaf8 WatchSource:0}: Error finding container faa002b610b0fc0441a885d9b5e3a1d86e9be95a0be26cc0d3f9949bd021eaf8: Status 404 returned error can't find the container with id faa002b610b0fc0441a885d9b5e3a1d86e9be95a0be26cc0d3f9949bd021eaf8 Oct 02 03:22:42 crc kubenswrapper[4775]: I1002 03:22:42.824489 4775 generic.go:334] "Generic (PLEG): container finished" podID="2c9ebe75-71af-415f-b74a-d24d0e677411" containerID="9098f3167b6e483dcd46e2cfcb81813b0005e374075fe2ffdfd2c1c1d0d5dfd9" exitCode=0 Oct 02 03:22:42 crc kubenswrapper[4775]: I1002 03:22:42.824531 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-ckn24" event={"ID":"2c9ebe75-71af-415f-b74a-d24d0e677411","Type":"ContainerDied","Data":"9098f3167b6e483dcd46e2cfcb81813b0005e374075fe2ffdfd2c1c1d0d5dfd9"} Oct 02 03:22:42 crc kubenswrapper[4775]: I1002 03:22:42.824794 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-ckn24" event={"ID":"2c9ebe75-71af-415f-b74a-d24d0e677411","Type":"ContainerStarted","Data":"faa002b610b0fc0441a885d9b5e3a1d86e9be95a0be26cc0d3f9949bd021eaf8"} Oct 02 03:22:44 crc kubenswrapper[4775]: I1002 03:22:44.297599 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-ckn24" Oct 02 03:22:44 crc kubenswrapper[4775]: I1002 03:22:44.430544 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qqnf\" (UniqueName: \"kubernetes.io/projected/2c9ebe75-71af-415f-b74a-d24d0e677411-kube-api-access-7qqnf\") pod \"2c9ebe75-71af-415f-b74a-d24d0e677411\" (UID: \"2c9ebe75-71af-415f-b74a-d24d0e677411\") " Oct 02 03:22:44 crc kubenswrapper[4775]: I1002 03:22:44.445162 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c9ebe75-71af-415f-b74a-d24d0e677411-kube-api-access-7qqnf" (OuterVolumeSpecName: "kube-api-access-7qqnf") pod "2c9ebe75-71af-415f-b74a-d24d0e677411" (UID: "2c9ebe75-71af-415f-b74a-d24d0e677411"). InnerVolumeSpecName "kube-api-access-7qqnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:22:44 crc kubenswrapper[4775]: I1002 03:22:44.533150 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qqnf\" (UniqueName: \"kubernetes.io/projected/2c9ebe75-71af-415f-b74a-d24d0e677411-kube-api-access-7qqnf\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:44 crc kubenswrapper[4775]: I1002 03:22:44.854377 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-ckn24" event={"ID":"2c9ebe75-71af-415f-b74a-d24d0e677411","Type":"ContainerDied","Data":"faa002b610b0fc0441a885d9b5e3a1d86e9be95a0be26cc0d3f9949bd021eaf8"} Oct 02 03:22:44 crc kubenswrapper[4775]: I1002 03:22:44.854435 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="faa002b610b0fc0441a885d9b5e3a1d86e9be95a0be26cc0d3f9949bd021eaf8" Oct 02 03:22:44 crc kubenswrapper[4775]: I1002 03:22:44.854520 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-ckn24" Oct 02 03:22:50 crc kubenswrapper[4775]: I1002 03:22:50.268094 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:50 crc kubenswrapper[4775]: I1002 03:22:50.268687 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:22:51 crc kubenswrapper[4775]: I1002 03:22:51.760744 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-53fd-account-create-px9tb"] Oct 02 03:22:51 crc kubenswrapper[4775]: E1002 03:22:51.761594 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c9ebe75-71af-415f-b74a-d24d0e677411" containerName="mariadb-database-create" Oct 02 03:22:51 crc kubenswrapper[4775]: I1002 03:22:51.761611 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c9ebe75-71af-415f-b74a-d24d0e677411" containerName="mariadb-database-create" Oct 02 03:22:51 crc kubenswrapper[4775]: I1002 03:22:51.761828 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c9ebe75-71af-415f-b74a-d24d0e677411" containerName="mariadb-database-create" Oct 02 03:22:51 crc kubenswrapper[4775]: I1002 03:22:51.762640 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-53fd-account-create-px9tb" Oct 02 03:22:51 crc kubenswrapper[4775]: I1002 03:22:51.766327 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Oct 02 03:22:51 crc kubenswrapper[4775]: I1002 03:22:51.794197 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-53fd-account-create-px9tb"] Oct 02 03:22:51 crc kubenswrapper[4775]: I1002 03:22:51.927319 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2jjh\" (UniqueName: \"kubernetes.io/projected/394faa8e-649e-4dc0-8b2c-ed1e9523e241-kube-api-access-p2jjh\") pod \"heat-53fd-account-create-px9tb\" (UID: \"394faa8e-649e-4dc0-8b2c-ed1e9523e241\") " pod="openstack/heat-53fd-account-create-px9tb" Oct 02 03:22:52 crc kubenswrapper[4775]: I1002 03:22:52.029727 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2jjh\" (UniqueName: \"kubernetes.io/projected/394faa8e-649e-4dc0-8b2c-ed1e9523e241-kube-api-access-p2jjh\") pod \"heat-53fd-account-create-px9tb\" (UID: \"394faa8e-649e-4dc0-8b2c-ed1e9523e241\") " pod="openstack/heat-53fd-account-create-px9tb" Oct 02 03:22:52 crc kubenswrapper[4775]: I1002 03:22:52.060640 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2jjh\" (UniqueName: \"kubernetes.io/projected/394faa8e-649e-4dc0-8b2c-ed1e9523e241-kube-api-access-p2jjh\") pod \"heat-53fd-account-create-px9tb\" (UID: \"394faa8e-649e-4dc0-8b2c-ed1e9523e241\") " pod="openstack/heat-53fd-account-create-px9tb" Oct 02 03:22:52 crc kubenswrapper[4775]: I1002 03:22:52.097468 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-53fd-account-create-px9tb" Oct 02 03:22:52 crc kubenswrapper[4775]: I1002 03:22:52.674929 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-53fd-account-create-px9tb"] Oct 02 03:22:52 crc kubenswrapper[4775]: W1002 03:22:52.680684 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod394faa8e_649e_4dc0_8b2c_ed1e9523e241.slice/crio-2dd45780fa93a1b15d51b1b873f14fcaa0f63e2d8e238c8f87e3379d9b7f8984 WatchSource:0}: Error finding container 2dd45780fa93a1b15d51b1b873f14fcaa0f63e2d8e238c8f87e3379d9b7f8984: Status 404 returned error can't find the container with id 2dd45780fa93a1b15d51b1b873f14fcaa0f63e2d8e238c8f87e3379d9b7f8984 Oct 02 03:22:52 crc kubenswrapper[4775]: I1002 03:22:52.948394 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-53fd-account-create-px9tb" event={"ID":"394faa8e-649e-4dc0-8b2c-ed1e9523e241","Type":"ContainerStarted","Data":"2dd45780fa93a1b15d51b1b873f14fcaa0f63e2d8e238c8f87e3379d9b7f8984"} Oct 02 03:22:53 crc kubenswrapper[4775]: I1002 03:22:53.964708 4775 generic.go:334] "Generic (PLEG): container finished" podID="394faa8e-649e-4dc0-8b2c-ed1e9523e241" containerID="c007d3ba64951b37481cc0a06ad17d7b0b7cf1181c9e44a84424b285a67ecdf0" exitCode=0 Oct 02 03:22:53 crc kubenswrapper[4775]: I1002 03:22:53.964798 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-53fd-account-create-px9tb" event={"ID":"394faa8e-649e-4dc0-8b2c-ed1e9523e241","Type":"ContainerDied","Data":"c007d3ba64951b37481cc0a06ad17d7b0b7cf1181c9e44a84424b285a67ecdf0"} Oct 02 03:22:55 crc kubenswrapper[4775]: I1002 03:22:55.402617 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-53fd-account-create-px9tb" Oct 02 03:22:55 crc kubenswrapper[4775]: I1002 03:22:55.524835 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2jjh\" (UniqueName: \"kubernetes.io/projected/394faa8e-649e-4dc0-8b2c-ed1e9523e241-kube-api-access-p2jjh\") pod \"394faa8e-649e-4dc0-8b2c-ed1e9523e241\" (UID: \"394faa8e-649e-4dc0-8b2c-ed1e9523e241\") " Oct 02 03:22:55 crc kubenswrapper[4775]: I1002 03:22:55.533436 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/394faa8e-649e-4dc0-8b2c-ed1e9523e241-kube-api-access-p2jjh" (OuterVolumeSpecName: "kube-api-access-p2jjh") pod "394faa8e-649e-4dc0-8b2c-ed1e9523e241" (UID: "394faa8e-649e-4dc0-8b2c-ed1e9523e241"). InnerVolumeSpecName "kube-api-access-p2jjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:22:55 crc kubenswrapper[4775]: I1002 03:22:55.627767 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2jjh\" (UniqueName: \"kubernetes.io/projected/394faa8e-649e-4dc0-8b2c-ed1e9523e241-kube-api-access-p2jjh\") on node \"crc\" DevicePath \"\"" Oct 02 03:22:55 crc kubenswrapper[4775]: I1002 03:22:55.990016 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-53fd-account-create-px9tb" event={"ID":"394faa8e-649e-4dc0-8b2c-ed1e9523e241","Type":"ContainerDied","Data":"2dd45780fa93a1b15d51b1b873f14fcaa0f63e2d8e238c8f87e3379d9b7f8984"} Oct 02 03:22:55 crc kubenswrapper[4775]: I1002 03:22:55.990073 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dd45780fa93a1b15d51b1b873f14fcaa0f63e2d8e238c8f87e3379d9b7f8984" Oct 02 03:22:55 crc kubenswrapper[4775]: I1002 03:22:55.990122 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-53fd-account-create-px9tb" Oct 02 03:22:56 crc kubenswrapper[4775]: I1002 03:22:56.820790 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-25qfl"] Oct 02 03:22:56 crc kubenswrapper[4775]: E1002 03:22:56.821823 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="394faa8e-649e-4dc0-8b2c-ed1e9523e241" containerName="mariadb-account-create" Oct 02 03:22:56 crc kubenswrapper[4775]: I1002 03:22:56.821843 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="394faa8e-649e-4dc0-8b2c-ed1e9523e241" containerName="mariadb-account-create" Oct 02 03:22:56 crc kubenswrapper[4775]: I1002 03:22:56.822170 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="394faa8e-649e-4dc0-8b2c-ed1e9523e241" containerName="mariadb-account-create" Oct 02 03:22:56 crc kubenswrapper[4775]: I1002 03:22:56.823192 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-25qfl" Oct 02 03:22:56 crc kubenswrapper[4775]: I1002 03:22:56.826374 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-4z9td" Oct 02 03:22:56 crc kubenswrapper[4775]: I1002 03:22:56.832898 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 02 03:22:56 crc kubenswrapper[4775]: I1002 03:22:56.844757 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-25qfl"] Oct 02 03:22:56 crc kubenswrapper[4775]: I1002 03:22:56.958376 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8qrt\" (UniqueName: \"kubernetes.io/projected/7bc90542-eaa4-4cb7-b041-8222fef25ceb-kube-api-access-g8qrt\") pod \"heat-db-sync-25qfl\" (UID: \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\") " pod="openstack/heat-db-sync-25qfl" Oct 02 03:22:56 crc kubenswrapper[4775]: I1002 03:22:56.958489 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc90542-eaa4-4cb7-b041-8222fef25ceb-combined-ca-bundle\") pod \"heat-db-sync-25qfl\" (UID: \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\") " pod="openstack/heat-db-sync-25qfl" Oct 02 03:22:56 crc kubenswrapper[4775]: I1002 03:22:56.958641 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc90542-eaa4-4cb7-b041-8222fef25ceb-config-data\") pod \"heat-db-sync-25qfl\" (UID: \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\") " pod="openstack/heat-db-sync-25qfl" Oct 02 03:22:57 crc kubenswrapper[4775]: I1002 03:22:57.062071 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8qrt\" (UniqueName: \"kubernetes.io/projected/7bc90542-eaa4-4cb7-b041-8222fef25ceb-kube-api-access-g8qrt\") pod \"heat-db-sync-25qfl\" (UID: \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\") " pod="openstack/heat-db-sync-25qfl" Oct 02 03:22:57 crc kubenswrapper[4775]: I1002 03:22:57.062144 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc90542-eaa4-4cb7-b041-8222fef25ceb-combined-ca-bundle\") pod \"heat-db-sync-25qfl\" (UID: \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\") " pod="openstack/heat-db-sync-25qfl" Oct 02 03:22:57 crc kubenswrapper[4775]: I1002 03:22:57.063353 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc90542-eaa4-4cb7-b041-8222fef25ceb-config-data\") pod \"heat-db-sync-25qfl\" (UID: \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\") " pod="openstack/heat-db-sync-25qfl" Oct 02 03:22:57 crc kubenswrapper[4775]: I1002 03:22:57.071864 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc90542-eaa4-4cb7-b041-8222fef25ceb-config-data\") pod \"heat-db-sync-25qfl\" (UID: \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\") " pod="openstack/heat-db-sync-25qfl" Oct 02 03:22:57 crc kubenswrapper[4775]: I1002 03:22:57.083646 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc90542-eaa4-4cb7-b041-8222fef25ceb-combined-ca-bundle\") pod \"heat-db-sync-25qfl\" (UID: \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\") " pod="openstack/heat-db-sync-25qfl" Oct 02 03:22:57 crc kubenswrapper[4775]: I1002 03:22:57.099322 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8qrt\" (UniqueName: \"kubernetes.io/projected/7bc90542-eaa4-4cb7-b041-8222fef25ceb-kube-api-access-g8qrt\") pod \"heat-db-sync-25qfl\" (UID: \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\") " pod="openstack/heat-db-sync-25qfl" Oct 02 03:22:57 crc kubenswrapper[4775]: I1002 03:22:57.143533 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-25qfl" Oct 02 03:22:57 crc kubenswrapper[4775]: I1002 03:22:57.636759 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-25qfl"] Oct 02 03:22:58 crc kubenswrapper[4775]: I1002 03:22:58.009833 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-25qfl" event={"ID":"7bc90542-eaa4-4cb7-b041-8222fef25ceb","Type":"ContainerStarted","Data":"d13377ef7732d892615b4162181a9192995e5cec7e3f73b96597168a09f9db30"} Oct 02 03:23:01 crc kubenswrapper[4775]: I1002 03:23:01.864603 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:23:03 crc kubenswrapper[4775]: I1002 03:23:03.399526 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-598f77fd5-6jh6q" Oct 02 03:23:03 crc kubenswrapper[4775]: I1002 03:23:03.484091 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-678b44477-rgx9c"] Oct 02 03:23:03 crc kubenswrapper[4775]: I1002 03:23:03.484311 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-678b44477-rgx9c" podUID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" containerName="horizon-log" containerID="cri-o://5dcf8616f73f8f2882adde2b4a02ca43c6fab0c9724bf4733d7d04335ce88e6c" gracePeriod=30 Oct 02 03:23:03 crc kubenswrapper[4775]: I1002 03:23:03.484687 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-678b44477-rgx9c" podUID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" containerName="horizon" containerID="cri-o://a9a1f6e63b4665842f12b06154913dd76656efb6b696a752765227d6f9efd426" gracePeriod=30 Oct 02 03:23:05 crc kubenswrapper[4775]: I1002 03:23:05.098445 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-25qfl" event={"ID":"7bc90542-eaa4-4cb7-b041-8222fef25ceb","Type":"ContainerStarted","Data":"48c43a6bd7ba4e609dec4579a08cf76cedf58641bac90e753d4e5bb4bb3abce8"} Oct 02 03:23:05 crc kubenswrapper[4775]: I1002 03:23:05.128759 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-25qfl" podStartSLOduration=2.908007764 podStartE2EDuration="9.128694482s" podCreationTimestamp="2025-10-02 03:22:56 +0000 UTC" firstStartedPulling="2025-10-02 03:22:57.655592555 +0000 UTC m=+6114.822336625" lastFinishedPulling="2025-10-02 03:23:03.876279303 +0000 UTC m=+6121.043023343" observedRunningTime="2025-10-02 03:23:05.122781149 +0000 UTC m=+6122.289525249" watchObservedRunningTime="2025-10-02 03:23:05.128694482 +0000 UTC m=+6122.295438582" Oct 02 03:23:07 crc kubenswrapper[4775]: I1002 03:23:07.139804 4775 generic.go:334] "Generic (PLEG): container finished" podID="7bc90542-eaa4-4cb7-b041-8222fef25ceb" containerID="48c43a6bd7ba4e609dec4579a08cf76cedf58641bac90e753d4e5bb4bb3abce8" exitCode=0 Oct 02 03:23:07 crc kubenswrapper[4775]: I1002 03:23:07.139883 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-25qfl" event={"ID":"7bc90542-eaa4-4cb7-b041-8222fef25ceb","Type":"ContainerDied","Data":"48c43a6bd7ba4e609dec4579a08cf76cedf58641bac90e753d4e5bb4bb3abce8"} Oct 02 03:23:07 crc kubenswrapper[4775]: I1002 03:23:07.147931 4775 generic.go:334] "Generic (PLEG): container finished" podID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" containerID="a9a1f6e63b4665842f12b06154913dd76656efb6b696a752765227d6f9efd426" exitCode=0 Oct 02 03:23:07 crc kubenswrapper[4775]: I1002 03:23:07.148025 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-678b44477-rgx9c" event={"ID":"fa5bfdee-6816-4e3a-b12d-e720bc834be8","Type":"ContainerDied","Data":"a9a1f6e63b4665842f12b06154913dd76656efb6b696a752765227d6f9efd426"} Oct 02 03:23:07 crc kubenswrapper[4775]: I1002 03:23:07.233615 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:23:07 crc kubenswrapper[4775]: I1002 03:23:07.233675 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:23:08 crc kubenswrapper[4775]: I1002 03:23:08.620223 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-25qfl" Oct 02 03:23:08 crc kubenswrapper[4775]: I1002 03:23:08.763586 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc90542-eaa4-4cb7-b041-8222fef25ceb-combined-ca-bundle\") pod \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\" (UID: \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\") " Oct 02 03:23:08 crc kubenswrapper[4775]: I1002 03:23:08.763730 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8qrt\" (UniqueName: \"kubernetes.io/projected/7bc90542-eaa4-4cb7-b041-8222fef25ceb-kube-api-access-g8qrt\") pod \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\" (UID: \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\") " Oct 02 03:23:08 crc kubenswrapper[4775]: I1002 03:23:08.764089 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc90542-eaa4-4cb7-b041-8222fef25ceb-config-data\") pod \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\" (UID: \"7bc90542-eaa4-4cb7-b041-8222fef25ceb\") " Oct 02 03:23:08 crc kubenswrapper[4775]: I1002 03:23:08.768857 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bc90542-eaa4-4cb7-b041-8222fef25ceb-kube-api-access-g8qrt" (OuterVolumeSpecName: "kube-api-access-g8qrt") pod "7bc90542-eaa4-4cb7-b041-8222fef25ceb" (UID: "7bc90542-eaa4-4cb7-b041-8222fef25ceb"). InnerVolumeSpecName "kube-api-access-g8qrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:23:08 crc kubenswrapper[4775]: I1002 03:23:08.796298 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc90542-eaa4-4cb7-b041-8222fef25ceb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bc90542-eaa4-4cb7-b041-8222fef25ceb" (UID: "7bc90542-eaa4-4cb7-b041-8222fef25ceb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:23:08 crc kubenswrapper[4775]: I1002 03:23:08.864275 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc90542-eaa4-4cb7-b041-8222fef25ceb-config-data" (OuterVolumeSpecName: "config-data") pod "7bc90542-eaa4-4cb7-b041-8222fef25ceb" (UID: "7bc90542-eaa4-4cb7-b041-8222fef25ceb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:23:08 crc kubenswrapper[4775]: I1002 03:23:08.867061 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bc90542-eaa4-4cb7-b041-8222fef25ceb-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:23:08 crc kubenswrapper[4775]: I1002 03:23:08.867106 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bc90542-eaa4-4cb7-b041-8222fef25ceb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:23:08 crc kubenswrapper[4775]: I1002 03:23:08.867126 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8qrt\" (UniqueName: \"kubernetes.io/projected/7bc90542-eaa4-4cb7-b041-8222fef25ceb-kube-api-access-g8qrt\") on node \"crc\" DevicePath \"\"" Oct 02 03:23:09 crc kubenswrapper[4775]: I1002 03:23:09.181320 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-25qfl" event={"ID":"7bc90542-eaa4-4cb7-b041-8222fef25ceb","Type":"ContainerDied","Data":"d13377ef7732d892615b4162181a9192995e5cec7e3f73b96597168a09f9db30"} Oct 02 03:23:09 crc kubenswrapper[4775]: I1002 03:23:09.181384 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d13377ef7732d892615b4162181a9192995e5cec7e3f73b96597168a09f9db30" Oct 02 03:23:09 crc kubenswrapper[4775]: I1002 03:23:09.181433 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-25qfl" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.740590 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-68fb455f9-b7mpg"] Oct 02 03:23:10 crc kubenswrapper[4775]: E1002 03:23:10.741209 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc90542-eaa4-4cb7-b041-8222fef25ceb" containerName="heat-db-sync" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.741223 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc90542-eaa4-4cb7-b041-8222fef25ceb" containerName="heat-db-sync" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.741456 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bc90542-eaa4-4cb7-b041-8222fef25ceb" containerName="heat-db-sync" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.742433 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.744915 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.745212 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.747633 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-4z9td" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.767612 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-68fb455f9-b7mpg"] Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.789013 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-d7c784464-67wzm"] Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.790518 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.793618 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.793795 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-d7c784464-67wzm"] Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.823228 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-9fcb5b9c4-kpksr"] Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.826631 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.828742 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.850546 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-9fcb5b9c4-kpksr"] Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.917283 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/654406cb-d5cc-46d0-9319-cd678ba45b72-combined-ca-bundle\") pod \"heat-cfnapi-9fcb5b9c4-kpksr\" (UID: \"654406cb-d5cc-46d0-9319-cd678ba45b72\") " pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.917364 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/654406cb-d5cc-46d0-9319-cd678ba45b72-config-data\") pod \"heat-cfnapi-9fcb5b9c4-kpksr\" (UID: \"654406cb-d5cc-46d0-9319-cd678ba45b72\") " pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.917392 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d843cd4a-9895-4385-90ca-6d9f91954adc-config-data-custom\") pod \"heat-engine-68fb455f9-b7mpg\" (UID: \"d843cd4a-9895-4385-90ca-6d9f91954adc\") " pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.917548 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cmpg\" (UniqueName: \"kubernetes.io/projected/654406cb-d5cc-46d0-9319-cd678ba45b72-kube-api-access-4cmpg\") pod \"heat-cfnapi-9fcb5b9c4-kpksr\" (UID: \"654406cb-d5cc-46d0-9319-cd678ba45b72\") " pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.917641 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62bxp\" (UniqueName: \"kubernetes.io/projected/37f06ca5-4d12-4f82-8403-3fc12c9bb1f0-kube-api-access-62bxp\") pod \"heat-api-d7c784464-67wzm\" (UID: \"37f06ca5-4d12-4f82-8403-3fc12c9bb1f0\") " pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.917698 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d843cd4a-9895-4385-90ca-6d9f91954adc-combined-ca-bundle\") pod \"heat-engine-68fb455f9-b7mpg\" (UID: \"d843cd4a-9895-4385-90ca-6d9f91954adc\") " pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.917743 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37f06ca5-4d12-4f82-8403-3fc12c9bb1f0-config-data\") pod \"heat-api-d7c784464-67wzm\" (UID: \"37f06ca5-4d12-4f82-8403-3fc12c9bb1f0\") " pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.917772 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d843cd4a-9895-4385-90ca-6d9f91954adc-config-data\") pod \"heat-engine-68fb455f9-b7mpg\" (UID: \"d843cd4a-9895-4385-90ca-6d9f91954adc\") " pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.917844 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37f06ca5-4d12-4f82-8403-3fc12c9bb1f0-config-data-custom\") pod \"heat-api-d7c784464-67wzm\" (UID: \"37f06ca5-4d12-4f82-8403-3fc12c9bb1f0\") " pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.917875 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/654406cb-d5cc-46d0-9319-cd678ba45b72-config-data-custom\") pod \"heat-cfnapi-9fcb5b9c4-kpksr\" (UID: \"654406cb-d5cc-46d0-9319-cd678ba45b72\") " pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.918087 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89bgb\" (UniqueName: \"kubernetes.io/projected/d843cd4a-9895-4385-90ca-6d9f91954adc-kube-api-access-89bgb\") pod \"heat-engine-68fb455f9-b7mpg\" (UID: \"d843cd4a-9895-4385-90ca-6d9f91954adc\") " pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:10 crc kubenswrapper[4775]: I1002 03:23:10.918187 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37f06ca5-4d12-4f82-8403-3fc12c9bb1f0-combined-ca-bundle\") pod \"heat-api-d7c784464-67wzm\" (UID: \"37f06ca5-4d12-4f82-8403-3fc12c9bb1f0\") " pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.019727 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/654406cb-d5cc-46d0-9319-cd678ba45b72-combined-ca-bundle\") pod \"heat-cfnapi-9fcb5b9c4-kpksr\" (UID: \"654406cb-d5cc-46d0-9319-cd678ba45b72\") " pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.019838 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/654406cb-d5cc-46d0-9319-cd678ba45b72-config-data\") pod \"heat-cfnapi-9fcb5b9c4-kpksr\" (UID: \"654406cb-d5cc-46d0-9319-cd678ba45b72\") " pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.019871 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d843cd4a-9895-4385-90ca-6d9f91954adc-config-data-custom\") pod \"heat-engine-68fb455f9-b7mpg\" (UID: \"d843cd4a-9895-4385-90ca-6d9f91954adc\") " pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.019897 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cmpg\" (UniqueName: \"kubernetes.io/projected/654406cb-d5cc-46d0-9319-cd678ba45b72-kube-api-access-4cmpg\") pod \"heat-cfnapi-9fcb5b9c4-kpksr\" (UID: \"654406cb-d5cc-46d0-9319-cd678ba45b72\") " pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.019933 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62bxp\" (UniqueName: \"kubernetes.io/projected/37f06ca5-4d12-4f82-8403-3fc12c9bb1f0-kube-api-access-62bxp\") pod \"heat-api-d7c784464-67wzm\" (UID: \"37f06ca5-4d12-4f82-8403-3fc12c9bb1f0\") " pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.019983 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d843cd4a-9895-4385-90ca-6d9f91954adc-combined-ca-bundle\") pod \"heat-engine-68fb455f9-b7mpg\" (UID: \"d843cd4a-9895-4385-90ca-6d9f91954adc\") " pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.020770 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37f06ca5-4d12-4f82-8403-3fc12c9bb1f0-config-data\") pod \"heat-api-d7c784464-67wzm\" (UID: \"37f06ca5-4d12-4f82-8403-3fc12c9bb1f0\") " pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.020813 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d843cd4a-9895-4385-90ca-6d9f91954adc-config-data\") pod \"heat-engine-68fb455f9-b7mpg\" (UID: \"d843cd4a-9895-4385-90ca-6d9f91954adc\") " pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.020854 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37f06ca5-4d12-4f82-8403-3fc12c9bb1f0-config-data-custom\") pod \"heat-api-d7c784464-67wzm\" (UID: \"37f06ca5-4d12-4f82-8403-3fc12c9bb1f0\") " pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.020879 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/654406cb-d5cc-46d0-9319-cd678ba45b72-config-data-custom\") pod \"heat-cfnapi-9fcb5b9c4-kpksr\" (UID: \"654406cb-d5cc-46d0-9319-cd678ba45b72\") " pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.020991 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89bgb\" (UniqueName: \"kubernetes.io/projected/d843cd4a-9895-4385-90ca-6d9f91954adc-kube-api-access-89bgb\") pod \"heat-engine-68fb455f9-b7mpg\" (UID: \"d843cd4a-9895-4385-90ca-6d9f91954adc\") " pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.021040 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37f06ca5-4d12-4f82-8403-3fc12c9bb1f0-combined-ca-bundle\") pod \"heat-api-d7c784464-67wzm\" (UID: \"37f06ca5-4d12-4f82-8403-3fc12c9bb1f0\") " pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.025650 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d843cd4a-9895-4385-90ca-6d9f91954adc-combined-ca-bundle\") pod \"heat-engine-68fb455f9-b7mpg\" (UID: \"d843cd4a-9895-4385-90ca-6d9f91954adc\") " pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.026635 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/654406cb-d5cc-46d0-9319-cd678ba45b72-config-data\") pod \"heat-cfnapi-9fcb5b9c4-kpksr\" (UID: \"654406cb-d5cc-46d0-9319-cd678ba45b72\") " pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.027055 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37f06ca5-4d12-4f82-8403-3fc12c9bb1f0-combined-ca-bundle\") pod \"heat-api-d7c784464-67wzm\" (UID: \"37f06ca5-4d12-4f82-8403-3fc12c9bb1f0\") " pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.027668 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d843cd4a-9895-4385-90ca-6d9f91954adc-config-data\") pod \"heat-engine-68fb455f9-b7mpg\" (UID: \"d843cd4a-9895-4385-90ca-6d9f91954adc\") " pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.027911 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/654406cb-d5cc-46d0-9319-cd678ba45b72-combined-ca-bundle\") pod \"heat-cfnapi-9fcb5b9c4-kpksr\" (UID: \"654406cb-d5cc-46d0-9319-cd678ba45b72\") " pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.030610 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/654406cb-d5cc-46d0-9319-cd678ba45b72-config-data-custom\") pod \"heat-cfnapi-9fcb5b9c4-kpksr\" (UID: \"654406cb-d5cc-46d0-9319-cd678ba45b72\") " pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.032262 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37f06ca5-4d12-4f82-8403-3fc12c9bb1f0-config-data\") pod \"heat-api-d7c784464-67wzm\" (UID: \"37f06ca5-4d12-4f82-8403-3fc12c9bb1f0\") " pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.037409 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d843cd4a-9895-4385-90ca-6d9f91954adc-config-data-custom\") pod \"heat-engine-68fb455f9-b7mpg\" (UID: \"d843cd4a-9895-4385-90ca-6d9f91954adc\") " pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.039174 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cmpg\" (UniqueName: \"kubernetes.io/projected/654406cb-d5cc-46d0-9319-cd678ba45b72-kube-api-access-4cmpg\") pod \"heat-cfnapi-9fcb5b9c4-kpksr\" (UID: \"654406cb-d5cc-46d0-9319-cd678ba45b72\") " pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.040869 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/37f06ca5-4d12-4f82-8403-3fc12c9bb1f0-config-data-custom\") pod \"heat-api-d7c784464-67wzm\" (UID: \"37f06ca5-4d12-4f82-8403-3fc12c9bb1f0\") " pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.042187 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62bxp\" (UniqueName: \"kubernetes.io/projected/37f06ca5-4d12-4f82-8403-3fc12c9bb1f0-kube-api-access-62bxp\") pod \"heat-api-d7c784464-67wzm\" (UID: \"37f06ca5-4d12-4f82-8403-3fc12c9bb1f0\") " pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.042531 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89bgb\" (UniqueName: \"kubernetes.io/projected/d843cd4a-9895-4385-90ca-6d9f91954adc-kube-api-access-89bgb\") pod \"heat-engine-68fb455f9-b7mpg\" (UID: \"d843cd4a-9895-4385-90ca-6d9f91954adc\") " pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.076852 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.128741 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.154261 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.637490 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-d7c784464-67wzm"] Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.649222 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-68fb455f9-b7mpg"] Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.744242 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-9fcb5b9c4-kpksr"] Oct 02 03:23:11 crc kubenswrapper[4775]: W1002 03:23:11.755233 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod654406cb_d5cc_46d0_9319_cd678ba45b72.slice/crio-b14861a5123ae47a717cfef07241e2037a535bcb88846d49a3e53e86ff036392 WatchSource:0}: Error finding container b14861a5123ae47a717cfef07241e2037a535bcb88846d49a3e53e86ff036392: Status 404 returned error can't find the container with id b14861a5123ae47a717cfef07241e2037a535bcb88846d49a3e53e86ff036392 Oct 02 03:23:11 crc kubenswrapper[4775]: I1002 03:23:11.835287 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-678b44477-rgx9c" podUID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 02 03:23:12 crc kubenswrapper[4775]: I1002 03:23:12.226900 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" event={"ID":"654406cb-d5cc-46d0-9319-cd678ba45b72","Type":"ContainerStarted","Data":"b14861a5123ae47a717cfef07241e2037a535bcb88846d49a3e53e86ff036392"} Oct 02 03:23:12 crc kubenswrapper[4775]: I1002 03:23:12.228146 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-d7c784464-67wzm" event={"ID":"37f06ca5-4d12-4f82-8403-3fc12c9bb1f0","Type":"ContainerStarted","Data":"65075c41a888021fbbfd948542815cd4771fcd4280b8305ad532fe0bd88f66ab"} Oct 02 03:23:12 crc kubenswrapper[4775]: I1002 03:23:12.230544 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-68fb455f9-b7mpg" event={"ID":"d843cd4a-9895-4385-90ca-6d9f91954adc","Type":"ContainerStarted","Data":"bb4f338d6ef1cc7c6c30ca5c37a4375aa3c289a13bdf480d65545d72da357268"} Oct 02 03:23:12 crc kubenswrapper[4775]: I1002 03:23:12.230603 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-68fb455f9-b7mpg" event={"ID":"d843cd4a-9895-4385-90ca-6d9f91954adc","Type":"ContainerStarted","Data":"87c064a417d51190f0a4af72d8283a600e3c0ffdb2ed2ac85ed3973d4de6970e"} Oct 02 03:23:12 crc kubenswrapper[4775]: I1002 03:23:12.230697 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:12 crc kubenswrapper[4775]: I1002 03:23:12.245528 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-68fb455f9-b7mpg" podStartSLOduration=2.245491847 podStartE2EDuration="2.245491847s" podCreationTimestamp="2025-10-02 03:23:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:23:12.244566694 +0000 UTC m=+6129.411310734" watchObservedRunningTime="2025-10-02 03:23:12.245491847 +0000 UTC m=+6129.412235887" Oct 02 03:23:13 crc kubenswrapper[4775]: I1002 03:23:13.383859 4775 scope.go:117] "RemoveContainer" containerID="d2c7bfa8a50cd48145fbcb8de04dd2aafa3e5db017948be1c67e31d1957bc6c3" Oct 02 03:23:13 crc kubenswrapper[4775]: I1002 03:23:13.457272 4775 scope.go:117] "RemoveContainer" containerID="8437f8b75a0f92054fb112385564c062db0176d5340decff9fefa4b465aeb915" Oct 02 03:23:13 crc kubenswrapper[4775]: I1002 03:23:13.510811 4775 scope.go:117] "RemoveContainer" containerID="bdf0be3d019ed7273b9673dd8adf2a9d0a9e8dda92f31132581db793bcad9217" Oct 02 03:23:14 crc kubenswrapper[4775]: I1002 03:23:14.268880 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-d7c784464-67wzm" event={"ID":"37f06ca5-4d12-4f82-8403-3fc12c9bb1f0","Type":"ContainerStarted","Data":"60183d280ed9848d89ad5319847280dcd640898e77b60739b30570ff3764dcf4"} Oct 02 03:23:14 crc kubenswrapper[4775]: I1002 03:23:14.269006 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:14 crc kubenswrapper[4775]: I1002 03:23:14.270385 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" event={"ID":"654406cb-d5cc-46d0-9319-cd678ba45b72","Type":"ContainerStarted","Data":"3a67d2248998779c9ce894d130d73b206fc9d24163975bd96e1bbda92308392c"} Oct 02 03:23:14 crc kubenswrapper[4775]: I1002 03:23:14.270486 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:14 crc kubenswrapper[4775]: I1002 03:23:14.290728 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-d7c784464-67wzm" podStartSLOduration=2.706329597 podStartE2EDuration="4.290711366s" podCreationTimestamp="2025-10-02 03:23:10 +0000 UTC" firstStartedPulling="2025-10-02 03:23:11.651278762 +0000 UTC m=+6128.818022792" lastFinishedPulling="2025-10-02 03:23:13.235660511 +0000 UTC m=+6130.402404561" observedRunningTime="2025-10-02 03:23:14.28586331 +0000 UTC m=+6131.452607350" watchObservedRunningTime="2025-10-02 03:23:14.290711366 +0000 UTC m=+6131.457455406" Oct 02 03:23:14 crc kubenswrapper[4775]: I1002 03:23:14.309589 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" podStartSLOduration=2.8265002299999997 podStartE2EDuration="4.309574744s" podCreationTimestamp="2025-10-02 03:23:10 +0000 UTC" firstStartedPulling="2025-10-02 03:23:11.75772859 +0000 UTC m=+6128.924472630" lastFinishedPulling="2025-10-02 03:23:13.240803104 +0000 UTC m=+6130.407547144" observedRunningTime="2025-10-02 03:23:14.30437743 +0000 UTC m=+6131.471121480" watchObservedRunningTime="2025-10-02 03:23:14.309574744 +0000 UTC m=+6131.476318784" Oct 02 03:23:21 crc kubenswrapper[4775]: I1002 03:23:21.133511 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-68fb455f9-b7mpg" Oct 02 03:23:21 crc kubenswrapper[4775]: I1002 03:23:21.835753 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-678b44477-rgx9c" podUID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 02 03:23:22 crc kubenswrapper[4775]: I1002 03:23:22.338051 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-9fcb5b9c4-kpksr" Oct 02 03:23:22 crc kubenswrapper[4775]: I1002 03:23:22.419770 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-d7c784464-67wzm" Oct 02 03:23:31 crc kubenswrapper[4775]: I1002 03:23:31.835219 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-678b44477-rgx9c" podUID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.114:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.114:8080: connect: connection refused" Oct 02 03:23:31 crc kubenswrapper[4775]: I1002 03:23:31.835712 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:23:33 crc kubenswrapper[4775]: E1002 03:23:33.712483 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa5bfdee_6816_4e3a_b12d_e720bc834be8.slice/crio-5dcf8616f73f8f2882adde2b4a02ca43c6fab0c9724bf4733d7d04335ce88e6c.scope\": RecentStats: unable to find data in memory cache]" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.076660 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.130991 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa5bfdee-6816-4e3a-b12d-e720bc834be8-logs\") pod \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.131481 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa5bfdee-6816-4e3a-b12d-e720bc834be8-scripts\") pod \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.131598 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa5bfdee-6816-4e3a-b12d-e720bc834be8-config-data\") pod \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.131631 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98rfd\" (UniqueName: \"kubernetes.io/projected/fa5bfdee-6816-4e3a-b12d-e720bc834be8-kube-api-access-98rfd\") pod \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.131707 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fa5bfdee-6816-4e3a-b12d-e720bc834be8-horizon-secret-key\") pod \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\" (UID: \"fa5bfdee-6816-4e3a-b12d-e720bc834be8\") " Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.132060 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa5bfdee-6816-4e3a-b12d-e720bc834be8-logs" (OuterVolumeSpecName: "logs") pod "fa5bfdee-6816-4e3a-b12d-e720bc834be8" (UID: "fa5bfdee-6816-4e3a-b12d-e720bc834be8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.132593 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fa5bfdee-6816-4e3a-b12d-e720bc834be8-logs\") on node \"crc\" DevicePath \"\"" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.150143 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa5bfdee-6816-4e3a-b12d-e720bc834be8-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "fa5bfdee-6816-4e3a-b12d-e720bc834be8" (UID: "fa5bfdee-6816-4e3a-b12d-e720bc834be8"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.151433 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa5bfdee-6816-4e3a-b12d-e720bc834be8-kube-api-access-98rfd" (OuterVolumeSpecName: "kube-api-access-98rfd") pod "fa5bfdee-6816-4e3a-b12d-e720bc834be8" (UID: "fa5bfdee-6816-4e3a-b12d-e720bc834be8"). InnerVolumeSpecName "kube-api-access-98rfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.160191 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa5bfdee-6816-4e3a-b12d-e720bc834be8-scripts" (OuterVolumeSpecName: "scripts") pod "fa5bfdee-6816-4e3a-b12d-e720bc834be8" (UID: "fa5bfdee-6816-4e3a-b12d-e720bc834be8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.166224 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa5bfdee-6816-4e3a-b12d-e720bc834be8-config-data" (OuterVolumeSpecName: "config-data") pod "fa5bfdee-6816-4e3a-b12d-e720bc834be8" (UID: "fa5bfdee-6816-4e3a-b12d-e720bc834be8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.234547 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fa5bfdee-6816-4e3a-b12d-e720bc834be8-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.234574 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fa5bfdee-6816-4e3a-b12d-e720bc834be8-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.234585 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98rfd\" (UniqueName: \"kubernetes.io/projected/fa5bfdee-6816-4e3a-b12d-e720bc834be8-kube-api-access-98rfd\") on node \"crc\" DevicePath \"\"" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.234594 4775 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/fa5bfdee-6816-4e3a-b12d-e720bc834be8-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.506813 4775 generic.go:334] "Generic (PLEG): container finished" podID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" containerID="5dcf8616f73f8f2882adde2b4a02ca43c6fab0c9724bf4733d7d04335ce88e6c" exitCode=137 Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.506861 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-678b44477-rgx9c" event={"ID":"fa5bfdee-6816-4e3a-b12d-e720bc834be8","Type":"ContainerDied","Data":"5dcf8616f73f8f2882adde2b4a02ca43c6fab0c9724bf4733d7d04335ce88e6c"} Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.506900 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-678b44477-rgx9c" event={"ID":"fa5bfdee-6816-4e3a-b12d-e720bc834be8","Type":"ContainerDied","Data":"6d1790dd50c313380fcfbf2748cb8a7f86cad990fbd85bca6d3e049d188aa807"} Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.506893 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-678b44477-rgx9c" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.506920 4775 scope.go:117] "RemoveContainer" containerID="a9a1f6e63b4665842f12b06154913dd76656efb6b696a752765227d6f9efd426" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.545323 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-678b44477-rgx9c"] Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.552523 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-678b44477-rgx9c"] Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.718340 4775 scope.go:117] "RemoveContainer" containerID="5dcf8616f73f8f2882adde2b4a02ca43c6fab0c9724bf4733d7d04335ce88e6c" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.738336 4775 scope.go:117] "RemoveContainer" containerID="a9a1f6e63b4665842f12b06154913dd76656efb6b696a752765227d6f9efd426" Oct 02 03:23:34 crc kubenswrapper[4775]: E1002 03:23:34.738696 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9a1f6e63b4665842f12b06154913dd76656efb6b696a752765227d6f9efd426\": container with ID starting with a9a1f6e63b4665842f12b06154913dd76656efb6b696a752765227d6f9efd426 not found: ID does not exist" containerID="a9a1f6e63b4665842f12b06154913dd76656efb6b696a752765227d6f9efd426" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.738753 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9a1f6e63b4665842f12b06154913dd76656efb6b696a752765227d6f9efd426"} err="failed to get container status \"a9a1f6e63b4665842f12b06154913dd76656efb6b696a752765227d6f9efd426\": rpc error: code = NotFound desc = could not find container \"a9a1f6e63b4665842f12b06154913dd76656efb6b696a752765227d6f9efd426\": container with ID starting with a9a1f6e63b4665842f12b06154913dd76656efb6b696a752765227d6f9efd426 not found: ID does not exist" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.738784 4775 scope.go:117] "RemoveContainer" containerID="5dcf8616f73f8f2882adde2b4a02ca43c6fab0c9724bf4733d7d04335ce88e6c" Oct 02 03:23:34 crc kubenswrapper[4775]: E1002 03:23:34.739248 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dcf8616f73f8f2882adde2b4a02ca43c6fab0c9724bf4733d7d04335ce88e6c\": container with ID starting with 5dcf8616f73f8f2882adde2b4a02ca43c6fab0c9724bf4733d7d04335ce88e6c not found: ID does not exist" containerID="5dcf8616f73f8f2882adde2b4a02ca43c6fab0c9724bf4733d7d04335ce88e6c" Oct 02 03:23:34 crc kubenswrapper[4775]: I1002 03:23:34.739279 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dcf8616f73f8f2882adde2b4a02ca43c6fab0c9724bf4733d7d04335ce88e6c"} err="failed to get container status \"5dcf8616f73f8f2882adde2b4a02ca43c6fab0c9724bf4733d7d04335ce88e6c\": rpc error: code = NotFound desc = could not find container \"5dcf8616f73f8f2882adde2b4a02ca43c6fab0c9724bf4733d7d04335ce88e6c\": container with ID starting with 5dcf8616f73f8f2882adde2b4a02ca43c6fab0c9724bf4733d7d04335ce88e6c not found: ID does not exist" Oct 02 03:23:35 crc kubenswrapper[4775]: I1002 03:23:35.781351 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" path="/var/lib/kubelet/pods/fa5bfdee-6816-4e3a-b12d-e720bc834be8/volumes" Oct 02 03:23:37 crc kubenswrapper[4775]: I1002 03:23:37.044376 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-srnpv"] Oct 02 03:23:37 crc kubenswrapper[4775]: I1002 03:23:37.051559 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-kzdtw"] Oct 02 03:23:37 crc kubenswrapper[4775]: I1002 03:23:37.057834 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-srnpv"] Oct 02 03:23:37 crc kubenswrapper[4775]: I1002 03:23:37.064694 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-kzdtw"] Oct 02 03:23:37 crc kubenswrapper[4775]: I1002 03:23:37.234397 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:23:37 crc kubenswrapper[4775]: I1002 03:23:37.234448 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:23:37 crc kubenswrapper[4775]: I1002 03:23:37.784205 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6223610a-fc09-44d4-9306-2b20384affeb" path="/var/lib/kubelet/pods/6223610a-fc09-44d4-9306-2b20384affeb/volumes" Oct 02 03:23:37 crc kubenswrapper[4775]: I1002 03:23:37.786404 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b563b0be-d1f0-431b-9e23-39b5a55ffc12" path="/var/lib/kubelet/pods/b563b0be-d1f0-431b-9e23-39b5a55ffc12/volumes" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.033013 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-4zbzp"] Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.050880 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-4zbzp"] Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.517243 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm"] Oct 02 03:23:38 crc kubenswrapper[4775]: E1002 03:23:38.517730 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" containerName="horizon-log" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.517751 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" containerName="horizon-log" Oct 02 03:23:38 crc kubenswrapper[4775]: E1002 03:23:38.517793 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" containerName="horizon" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.517803 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" containerName="horizon" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.518074 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" containerName="horizon-log" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.518105 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa5bfdee-6816-4e3a-b12d-e720bc834be8" containerName="horizon" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.519817 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.522467 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.539215 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frd7h\" (UniqueName: \"kubernetes.io/projected/4dbd62e7-a723-4c3b-945a-9014be98cb02-kube-api-access-frd7h\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm\" (UID: \"4dbd62e7-a723-4c3b-945a-9014be98cb02\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.539481 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dbd62e7-a723-4c3b-945a-9014be98cb02-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm\" (UID: \"4dbd62e7-a723-4c3b-945a-9014be98cb02\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.539678 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dbd62e7-a723-4c3b-945a-9014be98cb02-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm\" (UID: \"4dbd62e7-a723-4c3b-945a-9014be98cb02\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.541736 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm"] Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.642046 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dbd62e7-a723-4c3b-945a-9014be98cb02-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm\" (UID: \"4dbd62e7-a723-4c3b-945a-9014be98cb02\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.642250 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dbd62e7-a723-4c3b-945a-9014be98cb02-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm\" (UID: \"4dbd62e7-a723-4c3b-945a-9014be98cb02\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.642353 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frd7h\" (UniqueName: \"kubernetes.io/projected/4dbd62e7-a723-4c3b-945a-9014be98cb02-kube-api-access-frd7h\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm\" (UID: \"4dbd62e7-a723-4c3b-945a-9014be98cb02\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.642649 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dbd62e7-a723-4c3b-945a-9014be98cb02-bundle\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm\" (UID: \"4dbd62e7-a723-4c3b-945a-9014be98cb02\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.642990 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dbd62e7-a723-4c3b-945a-9014be98cb02-util\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm\" (UID: \"4dbd62e7-a723-4c3b-945a-9014be98cb02\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.667571 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frd7h\" (UniqueName: \"kubernetes.io/projected/4dbd62e7-a723-4c3b-945a-9014be98cb02-kube-api-access-frd7h\") pod \"a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm\" (UID: \"4dbd62e7-a723-4c3b-945a-9014be98cb02\") " pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" Oct 02 03:23:38 crc kubenswrapper[4775]: I1002 03:23:38.843474 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" Oct 02 03:23:39 crc kubenswrapper[4775]: I1002 03:23:39.210412 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm"] Oct 02 03:23:39 crc kubenswrapper[4775]: W1002 03:23:39.220934 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4dbd62e7_a723_4c3b_945a_9014be98cb02.slice/crio-9222824f6c29e4b12fb3984ed7aedb9fa96a2f5a24698cc0e19471304d1f286f WatchSource:0}: Error finding container 9222824f6c29e4b12fb3984ed7aedb9fa96a2f5a24698cc0e19471304d1f286f: Status 404 returned error can't find the container with id 9222824f6c29e4b12fb3984ed7aedb9fa96a2f5a24698cc0e19471304d1f286f Oct 02 03:23:39 crc kubenswrapper[4775]: I1002 03:23:39.577620 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" event={"ID":"4dbd62e7-a723-4c3b-945a-9014be98cb02","Type":"ContainerStarted","Data":"281ce3e21736aed641f0ae873dc806447d74e1ae84f09aa8a5f40c033c306c85"} Oct 02 03:23:39 crc kubenswrapper[4775]: I1002 03:23:39.577693 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" event={"ID":"4dbd62e7-a723-4c3b-945a-9014be98cb02","Type":"ContainerStarted","Data":"9222824f6c29e4b12fb3984ed7aedb9fa96a2f5a24698cc0e19471304d1f286f"} Oct 02 03:23:39 crc kubenswrapper[4775]: I1002 03:23:39.787851 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6960b657-431b-433a-b1b1-175d8fa098a7" path="/var/lib/kubelet/pods/6960b657-431b-433a-b1b1-175d8fa098a7/volumes" Oct 02 03:23:40 crc kubenswrapper[4775]: I1002 03:23:40.593801 4775 generic.go:334] "Generic (PLEG): container finished" podID="4dbd62e7-a723-4c3b-945a-9014be98cb02" containerID="281ce3e21736aed641f0ae873dc806447d74e1ae84f09aa8a5f40c033c306c85" exitCode=0 Oct 02 03:23:40 crc kubenswrapper[4775]: I1002 03:23:40.595532 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" event={"ID":"4dbd62e7-a723-4c3b-945a-9014be98cb02","Type":"ContainerDied","Data":"281ce3e21736aed641f0ae873dc806447d74e1ae84f09aa8a5f40c033c306c85"} Oct 02 03:23:42 crc kubenswrapper[4775]: I1002 03:23:42.622447 4775 generic.go:334] "Generic (PLEG): container finished" podID="4dbd62e7-a723-4c3b-945a-9014be98cb02" containerID="58fcce93c17bad30b46fe39f2c185c2dc0dc2bb7bbba68ace9cbcfa63d1b57f0" exitCode=0 Oct 02 03:23:42 crc kubenswrapper[4775]: I1002 03:23:42.622579 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" event={"ID":"4dbd62e7-a723-4c3b-945a-9014be98cb02","Type":"ContainerDied","Data":"58fcce93c17bad30b46fe39f2c185c2dc0dc2bb7bbba68ace9cbcfa63d1b57f0"} Oct 02 03:23:43 crc kubenswrapper[4775]: I1002 03:23:43.638991 4775 generic.go:334] "Generic (PLEG): container finished" podID="4dbd62e7-a723-4c3b-945a-9014be98cb02" containerID="fef1da59f5cc4c18785ba44a6129672723e5ba7b4a441e824e8df5f61076d22b" exitCode=0 Oct 02 03:23:43 crc kubenswrapper[4775]: I1002 03:23:43.639085 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" event={"ID":"4dbd62e7-a723-4c3b-945a-9014be98cb02","Type":"ContainerDied","Data":"fef1da59f5cc4c18785ba44a6129672723e5ba7b4a441e824e8df5f61076d22b"} Oct 02 03:23:45 crc kubenswrapper[4775]: I1002 03:23:45.130673 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" Oct 02 03:23:45 crc kubenswrapper[4775]: I1002 03:23:45.330586 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dbd62e7-a723-4c3b-945a-9014be98cb02-util\") pod \"4dbd62e7-a723-4c3b-945a-9014be98cb02\" (UID: \"4dbd62e7-a723-4c3b-945a-9014be98cb02\") " Oct 02 03:23:45 crc kubenswrapper[4775]: I1002 03:23:45.331116 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frd7h\" (UniqueName: \"kubernetes.io/projected/4dbd62e7-a723-4c3b-945a-9014be98cb02-kube-api-access-frd7h\") pod \"4dbd62e7-a723-4c3b-945a-9014be98cb02\" (UID: \"4dbd62e7-a723-4c3b-945a-9014be98cb02\") " Oct 02 03:23:45 crc kubenswrapper[4775]: I1002 03:23:45.331277 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dbd62e7-a723-4c3b-945a-9014be98cb02-bundle\") pod \"4dbd62e7-a723-4c3b-945a-9014be98cb02\" (UID: \"4dbd62e7-a723-4c3b-945a-9014be98cb02\") " Oct 02 03:23:45 crc kubenswrapper[4775]: I1002 03:23:45.335153 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dbd62e7-a723-4c3b-945a-9014be98cb02-bundle" (OuterVolumeSpecName: "bundle") pod "4dbd62e7-a723-4c3b-945a-9014be98cb02" (UID: "4dbd62e7-a723-4c3b-945a-9014be98cb02"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:23:45 crc kubenswrapper[4775]: I1002 03:23:45.339101 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dbd62e7-a723-4c3b-945a-9014be98cb02-kube-api-access-frd7h" (OuterVolumeSpecName: "kube-api-access-frd7h") pod "4dbd62e7-a723-4c3b-945a-9014be98cb02" (UID: "4dbd62e7-a723-4c3b-945a-9014be98cb02"). InnerVolumeSpecName "kube-api-access-frd7h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:23:45 crc kubenswrapper[4775]: I1002 03:23:45.352594 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dbd62e7-a723-4c3b-945a-9014be98cb02-util" (OuterVolumeSpecName: "util") pod "4dbd62e7-a723-4c3b-945a-9014be98cb02" (UID: "4dbd62e7-a723-4c3b-945a-9014be98cb02"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:23:45 crc kubenswrapper[4775]: I1002 03:23:45.434646 4775 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dbd62e7-a723-4c3b-945a-9014be98cb02-util\") on node \"crc\" DevicePath \"\"" Oct 02 03:23:45 crc kubenswrapper[4775]: I1002 03:23:45.434691 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frd7h\" (UniqueName: \"kubernetes.io/projected/4dbd62e7-a723-4c3b-945a-9014be98cb02-kube-api-access-frd7h\") on node \"crc\" DevicePath \"\"" Oct 02 03:23:45 crc kubenswrapper[4775]: I1002 03:23:45.434714 4775 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dbd62e7-a723-4c3b-945a-9014be98cb02-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:23:45 crc kubenswrapper[4775]: I1002 03:23:45.666490 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" event={"ID":"4dbd62e7-a723-4c3b-945a-9014be98cb02","Type":"ContainerDied","Data":"9222824f6c29e4b12fb3984ed7aedb9fa96a2f5a24698cc0e19471304d1f286f"} Oct 02 03:23:45 crc kubenswrapper[4775]: I1002 03:23:45.666548 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9222824f6c29e4b12fb3984ed7aedb9fa96a2f5a24698cc0e19471304d1f286f" Oct 02 03:23:45 crc kubenswrapper[4775]: I1002 03:23:45.666625 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm" Oct 02 03:23:47 crc kubenswrapper[4775]: I1002 03:23:47.038934 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-f7fd-account-create-29988"] Oct 02 03:23:47 crc kubenswrapper[4775]: I1002 03:23:47.047070 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-f7fd-account-create-29988"] Oct 02 03:23:47 crc kubenswrapper[4775]: I1002 03:23:47.786537 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f38e9c6-c014-4516-947b-08247cdc35c7" path="/var/lib/kubelet/pods/7f38e9c6-c014-4516-947b-08247cdc35c7/volumes" Oct 02 03:23:48 crc kubenswrapper[4775]: I1002 03:23:48.050793 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-08e5-account-create-2rsjm"] Oct 02 03:23:48 crc kubenswrapper[4775]: I1002 03:23:48.068158 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-7669-account-create-lm5z5"] Oct 02 03:23:48 crc kubenswrapper[4775]: I1002 03:23:48.077447 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-08e5-account-create-2rsjm"] Oct 02 03:23:48 crc kubenswrapper[4775]: I1002 03:23:48.087002 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-7669-account-create-lm5z5"] Oct 02 03:23:49 crc kubenswrapper[4775]: I1002 03:23:49.787793 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63300bd0-8471-47ca-b748-ee6526e02957" path="/var/lib/kubelet/pods/63300bd0-8471-47ca-b748-ee6526e02957/volumes" Oct 02 03:23:49 crc kubenswrapper[4775]: I1002 03:23:49.788554 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d122ae48-f8d4-46c3-bb0e-a890010f250a" path="/var/lib/kubelet/pods/d122ae48-f8d4-46c3-bb0e-a890010f250a/volumes" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.531504 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-twdtn"] Oct 02 03:23:55 crc kubenswrapper[4775]: E1002 03:23:55.532319 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dbd62e7-a723-4c3b-945a-9014be98cb02" containerName="util" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.532331 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dbd62e7-a723-4c3b-945a-9014be98cb02" containerName="util" Oct 02 03:23:55 crc kubenswrapper[4775]: E1002 03:23:55.532358 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dbd62e7-a723-4c3b-945a-9014be98cb02" containerName="pull" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.532364 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dbd62e7-a723-4c3b-945a-9014be98cb02" containerName="pull" Oct 02 03:23:55 crc kubenswrapper[4775]: E1002 03:23:55.532375 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dbd62e7-a723-4c3b-945a-9014be98cb02" containerName="extract" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.532381 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dbd62e7-a723-4c3b-945a-9014be98cb02" containerName="extract" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.532574 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dbd62e7-a723-4c3b-945a-9014be98cb02" containerName="extract" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.533217 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-twdtn" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.535141 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-txqmm" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.535836 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.541024 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.559307 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-twdtn"] Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.647067 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq"] Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.671035 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxck6\" (UniqueName: \"kubernetes.io/projected/09b1503e-76b6-40c6-9494-b69347e2ba6b-kube-api-access-vxck6\") pod \"obo-prometheus-operator-7c8cf85677-twdtn\" (UID: \"09b1503e-76b6-40c6-9494-b69347e2ba6b\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-twdtn" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.678565 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt"] Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.678762 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.680669 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.687710 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-njblx" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.687906 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.700400 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt"] Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.713035 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq"] Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.772572 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxck6\" (UniqueName: \"kubernetes.io/projected/09b1503e-76b6-40c6-9494-b69347e2ba6b-kube-api-access-vxck6\") pod \"obo-prometheus-operator-7c8cf85677-twdtn\" (UID: \"09b1503e-76b6-40c6-9494-b69347e2ba6b\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-twdtn" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.795363 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxck6\" (UniqueName: \"kubernetes.io/projected/09b1503e-76b6-40c6-9494-b69347e2ba6b-kube-api-access-vxck6\") pod \"obo-prometheus-operator-7c8cf85677-twdtn\" (UID: \"09b1503e-76b6-40c6-9494-b69347e2ba6b\") " pod="openshift-operators/obo-prometheus-operator-7c8cf85677-twdtn" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.851203 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-twdtn" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.862126 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-2gm7h"] Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.863316 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-2gm7h" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.865731 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-mmhz5" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.865849 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.874357 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq\" (UID: \"b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.874451 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq\" (UID: \"b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.874626 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/902f36d9-43f8-451f-b83a-382bcc50e46c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt\" (UID: \"902f36d9-43f8-451f-b83a-382bcc50e46c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.874748 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/902f36d9-43f8-451f-b83a-382bcc50e46c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt\" (UID: \"902f36d9-43f8-451f-b83a-382bcc50e46c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.878654 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-2gm7h"] Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.976161 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/902f36d9-43f8-451f-b83a-382bcc50e46c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt\" (UID: \"902f36d9-43f8-451f-b83a-382bcc50e46c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.976252 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/902f36d9-43f8-451f-b83a-382bcc50e46c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt\" (UID: \"902f36d9-43f8-451f-b83a-382bcc50e46c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.976289 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-844lt\" (UniqueName: \"kubernetes.io/projected/6705f52c-cdc8-4b21-a92d-88eadcdf8754-kube-api-access-844lt\") pod \"observability-operator-cc5f78dfc-2gm7h\" (UID: \"6705f52c-cdc8-4b21-a92d-88eadcdf8754\") " pod="openshift-operators/observability-operator-cc5f78dfc-2gm7h" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.976309 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/6705f52c-cdc8-4b21-a92d-88eadcdf8754-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-2gm7h\" (UID: \"6705f52c-cdc8-4b21-a92d-88eadcdf8754\") " pod="openshift-operators/observability-operator-cc5f78dfc-2gm7h" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.976348 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq\" (UID: \"b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.976379 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq\" (UID: \"b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.980432 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq\" (UID: \"b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.980432 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/902f36d9-43f8-451f-b83a-382bcc50e46c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt\" (UID: \"902f36d9-43f8-451f-b83a-382bcc50e46c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.981494 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq\" (UID: \"b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq" Oct 02 03:23:55 crc kubenswrapper[4775]: I1002 03:23:55.981539 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/902f36d9-43f8-451f-b83a-382bcc50e46c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt\" (UID: \"902f36d9-43f8-451f-b83a-382bcc50e46c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.002181 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.013197 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.083851 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-844lt\" (UniqueName: \"kubernetes.io/projected/6705f52c-cdc8-4b21-a92d-88eadcdf8754-kube-api-access-844lt\") pod \"observability-operator-cc5f78dfc-2gm7h\" (UID: \"6705f52c-cdc8-4b21-a92d-88eadcdf8754\") " pod="openshift-operators/observability-operator-cc5f78dfc-2gm7h" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.084202 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/6705f52c-cdc8-4b21-a92d-88eadcdf8754-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-2gm7h\" (UID: \"6705f52c-cdc8-4b21-a92d-88eadcdf8754\") " pod="openshift-operators/observability-operator-cc5f78dfc-2gm7h" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.092200 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-lzh9v"] Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.093893 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-lzh9v" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.096507 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-vh2xx" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.099071 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/6705f52c-cdc8-4b21-a92d-88eadcdf8754-observability-operator-tls\") pod \"observability-operator-cc5f78dfc-2gm7h\" (UID: \"6705f52c-cdc8-4b21-a92d-88eadcdf8754\") " pod="openshift-operators/observability-operator-cc5f78dfc-2gm7h" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.126993 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-844lt\" (UniqueName: \"kubernetes.io/projected/6705f52c-cdc8-4b21-a92d-88eadcdf8754-kube-api-access-844lt\") pod \"observability-operator-cc5f78dfc-2gm7h\" (UID: \"6705f52c-cdc8-4b21-a92d-88eadcdf8754\") " pod="openshift-operators/observability-operator-cc5f78dfc-2gm7h" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.127061 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-lzh9v"] Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.186204 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/4dac8c52-da57-4264-a2c1-43641d9a4429-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-lzh9v\" (UID: \"4dac8c52-da57-4264-a2c1-43641d9a4429\") " pod="openshift-operators/perses-operator-54bc95c9fb-lzh9v" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.186571 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv24q\" (UniqueName: \"kubernetes.io/projected/4dac8c52-da57-4264-a2c1-43641d9a4429-kube-api-access-sv24q\") pod \"perses-operator-54bc95c9fb-lzh9v\" (UID: \"4dac8c52-da57-4264-a2c1-43641d9a4429\") " pod="openshift-operators/perses-operator-54bc95c9fb-lzh9v" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.278429 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-cc5f78dfc-2gm7h" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.288848 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv24q\" (UniqueName: \"kubernetes.io/projected/4dac8c52-da57-4264-a2c1-43641d9a4429-kube-api-access-sv24q\") pod \"perses-operator-54bc95c9fb-lzh9v\" (UID: \"4dac8c52-da57-4264-a2c1-43641d9a4429\") " pod="openshift-operators/perses-operator-54bc95c9fb-lzh9v" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.288944 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/4dac8c52-da57-4264-a2c1-43641d9a4429-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-lzh9v\" (UID: \"4dac8c52-da57-4264-a2c1-43641d9a4429\") " pod="openshift-operators/perses-operator-54bc95c9fb-lzh9v" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.289753 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/4dac8c52-da57-4264-a2c1-43641d9a4429-openshift-service-ca\") pod \"perses-operator-54bc95c9fb-lzh9v\" (UID: \"4dac8c52-da57-4264-a2c1-43641d9a4429\") " pod="openshift-operators/perses-operator-54bc95c9fb-lzh9v" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.305544 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv24q\" (UniqueName: \"kubernetes.io/projected/4dac8c52-da57-4264-a2c1-43641d9a4429-kube-api-access-sv24q\") pod \"perses-operator-54bc95c9fb-lzh9v\" (UID: \"4dac8c52-da57-4264-a2c1-43641d9a4429\") " pod="openshift-operators/perses-operator-54bc95c9fb-lzh9v" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.441344 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-54bc95c9fb-lzh9v" Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.549483 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-7c8cf85677-twdtn"] Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.627261 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq"] Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.732931 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt"] Oct 02 03:23:56 crc kubenswrapper[4775]: W1002 03:23:56.741585 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod902f36d9_43f8_451f_b83a_382bcc50e46c.slice/crio-373d065305ba1e114151d23d3d8427825ce99921dbaf5d8e937cfb9f4c7f60bb WatchSource:0}: Error finding container 373d065305ba1e114151d23d3d8427825ce99921dbaf5d8e937cfb9f4c7f60bb: Status 404 returned error can't find the container with id 373d065305ba1e114151d23d3d8427825ce99921dbaf5d8e937cfb9f4c7f60bb Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.813984 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq" event={"ID":"b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07","Type":"ContainerStarted","Data":"a6f88b694f13b381394ec3c13853019b1b9c1609d1e17371ff930e832808e804"} Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.815226 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-twdtn" event={"ID":"09b1503e-76b6-40c6-9494-b69347e2ba6b","Type":"ContainerStarted","Data":"8ce0ef56aa1e11f0cc477766a3468991d6247e4f0a139eae5a258bac2630dbcd"} Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.816064 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt" event={"ID":"902f36d9-43f8-451f-b83a-382bcc50e46c","Type":"ContainerStarted","Data":"373d065305ba1e114151d23d3d8427825ce99921dbaf5d8e937cfb9f4c7f60bb"} Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.819359 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-cc5f78dfc-2gm7h"] Oct 02 03:23:56 crc kubenswrapper[4775]: I1002 03:23:56.984584 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-54bc95c9fb-lzh9v"] Oct 02 03:23:56 crc kubenswrapper[4775]: W1002 03:23:56.988871 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4dac8c52_da57_4264_a2c1_43641d9a4429.slice/crio-b7f0da0662f7fc889bcc3963f7ecc300302a0503b54143851bfd96b112156753 WatchSource:0}: Error finding container b7f0da0662f7fc889bcc3963f7ecc300302a0503b54143851bfd96b112156753: Status 404 returned error can't find the container with id b7f0da0662f7fc889bcc3963f7ecc300302a0503b54143851bfd96b112156753 Oct 02 03:23:57 crc kubenswrapper[4775]: I1002 03:23:57.033313 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tfpzn"] Oct 02 03:23:57 crc kubenswrapper[4775]: I1002 03:23:57.041366 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-tfpzn"] Oct 02 03:23:57 crc kubenswrapper[4775]: I1002 03:23:57.783966 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ecc60ed-4386-41a8-8827-421edf691ef5" path="/var/lib/kubelet/pods/7ecc60ed-4386-41a8-8827-421edf691ef5/volumes" Oct 02 03:23:57 crc kubenswrapper[4775]: I1002 03:23:57.834224 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-lzh9v" event={"ID":"4dac8c52-da57-4264-a2c1-43641d9a4429","Type":"ContainerStarted","Data":"b7f0da0662f7fc889bcc3963f7ecc300302a0503b54143851bfd96b112156753"} Oct 02 03:23:57 crc kubenswrapper[4775]: I1002 03:23:57.836900 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-2gm7h" event={"ID":"6705f52c-cdc8-4b21-a92d-88eadcdf8754","Type":"ContainerStarted","Data":"4dd4de9c4327c7c800daa9e960d8e688f06c87b70ae34a283ded2ef5d44de875"} Oct 02 03:24:02 crc kubenswrapper[4775]: I1002 03:24:02.889096 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt" event={"ID":"902f36d9-43f8-451f-b83a-382bcc50e46c","Type":"ContainerStarted","Data":"c1f5977d97ab573f741afb1e679ea1612d6ac3c65d13b15abd26caaff2e5339a"} Oct 02 03:24:02 crc kubenswrapper[4775]: I1002 03:24:02.907623 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt" podStartSLOduration=3.051935362 podStartE2EDuration="7.907585125s" podCreationTimestamp="2025-10-02 03:23:55 +0000 UTC" firstStartedPulling="2025-10-02 03:23:56.745323289 +0000 UTC m=+6173.912067329" lastFinishedPulling="2025-10-02 03:24:01.600973052 +0000 UTC m=+6178.767717092" observedRunningTime="2025-10-02 03:24:02.904111815 +0000 UTC m=+6180.070855865" watchObservedRunningTime="2025-10-02 03:24:02.907585125 +0000 UTC m=+6180.074329165" Oct 02 03:24:04 crc kubenswrapper[4775]: I1002 03:24:04.924368 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-54bc95c9fb-lzh9v" event={"ID":"4dac8c52-da57-4264-a2c1-43641d9a4429","Type":"ContainerStarted","Data":"fa08a66bff24e987c7ae018be6023426acc32d0eef25fc0a7f39b9c03744a59b"} Oct 02 03:24:04 crc kubenswrapper[4775]: I1002 03:24:04.924851 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-54bc95c9fb-lzh9v" Oct 02 03:24:04 crc kubenswrapper[4775]: I1002 03:24:04.926133 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-cc5f78dfc-2gm7h" event={"ID":"6705f52c-cdc8-4b21-a92d-88eadcdf8754","Type":"ContainerStarted","Data":"965b0ba9a2ed0d6bf2dfb4219bdf22fb06e63af8819a202569274ac1cf94f2aa"} Oct 02 03:24:04 crc kubenswrapper[4775]: I1002 03:24:04.927078 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-cc5f78dfc-2gm7h" Oct 02 03:24:04 crc kubenswrapper[4775]: I1002 03:24:04.928860 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq" event={"ID":"b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07","Type":"ContainerStarted","Data":"7292417e7484a01292bde25cbf2e00e01a8c30367d53c02efe80c6edb9cf20fe"} Oct 02 03:24:04 crc kubenswrapper[4775]: I1002 03:24:04.930723 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-twdtn" event={"ID":"09b1503e-76b6-40c6-9494-b69347e2ba6b","Type":"ContainerStarted","Data":"8892ceeab91cf2d882c17fbe996d2d060c9efa43a88098d490c39936a329d40d"} Oct 02 03:24:04 crc kubenswrapper[4775]: I1002 03:24:04.950026 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-cc5f78dfc-2gm7h" Oct 02 03:24:04 crc kubenswrapper[4775]: I1002 03:24:04.964303 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-54bc95c9fb-lzh9v" podStartSLOduration=4.2982295 podStartE2EDuration="8.96428361s" podCreationTimestamp="2025-10-02 03:23:56 +0000 UTC" firstStartedPulling="2025-10-02 03:23:56.991818216 +0000 UTC m=+6174.158562256" lastFinishedPulling="2025-10-02 03:24:01.657872326 +0000 UTC m=+6178.824616366" observedRunningTime="2025-10-02 03:24:04.943561813 +0000 UTC m=+6182.110305853" watchObservedRunningTime="2025-10-02 03:24:04.96428361 +0000 UTC m=+6182.131027650" Oct 02 03:24:04 crc kubenswrapper[4775]: I1002 03:24:04.964499 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq" podStartSLOduration=5.000090255 podStartE2EDuration="9.964495466s" podCreationTimestamp="2025-10-02 03:23:55 +0000 UTC" firstStartedPulling="2025-10-02 03:23:56.635638267 +0000 UTC m=+6173.802382307" lastFinishedPulling="2025-10-02 03:24:01.600043478 +0000 UTC m=+6178.766787518" observedRunningTime="2025-10-02 03:24:04.960328918 +0000 UTC m=+6182.127072958" watchObservedRunningTime="2025-10-02 03:24:04.964495466 +0000 UTC m=+6182.131239506" Oct 02 03:24:04 crc kubenswrapper[4775]: I1002 03:24:04.981872 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-cc5f78dfc-2gm7h" podStartSLOduration=2.5349577070000002 podStartE2EDuration="9.981779743s" podCreationTimestamp="2025-10-02 03:23:55 +0000 UTC" firstStartedPulling="2025-10-02 03:23:56.822007976 +0000 UTC m=+6173.988752026" lastFinishedPulling="2025-10-02 03:24:04.268830022 +0000 UTC m=+6181.435574062" observedRunningTime="2025-10-02 03:24:04.977840641 +0000 UTC m=+6182.144584681" watchObservedRunningTime="2025-10-02 03:24:04.981779743 +0000 UTC m=+6182.148523783" Oct 02 03:24:04 crc kubenswrapper[4775]: I1002 03:24:04.999047 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-7c8cf85677-twdtn" podStartSLOduration=4.90382153 podStartE2EDuration="9.99903227s" podCreationTimestamp="2025-10-02 03:23:55 +0000 UTC" firstStartedPulling="2025-10-02 03:23:56.563872698 +0000 UTC m=+6173.730616738" lastFinishedPulling="2025-10-02 03:24:01.659083438 +0000 UTC m=+6178.825827478" observedRunningTime="2025-10-02 03:24:04.998125397 +0000 UTC m=+6182.164869437" watchObservedRunningTime="2025-10-02 03:24:04.99903227 +0000 UTC m=+6182.165776300" Oct 02 03:24:07 crc kubenswrapper[4775]: I1002 03:24:07.233865 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:24:07 crc kubenswrapper[4775]: I1002 03:24:07.234216 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:24:07 crc kubenswrapper[4775]: I1002 03:24:07.234269 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 03:24:07 crc kubenswrapper[4775]: I1002 03:24:07.235033 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c4930148cbbbfb59b99fb94d29b145a85b1cfbffe3a70db63ae0fd48271fb062"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 03:24:07 crc kubenswrapper[4775]: I1002 03:24:07.235083 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://c4930148cbbbfb59b99fb94d29b145a85b1cfbffe3a70db63ae0fd48271fb062" gracePeriod=600 Oct 02 03:24:07 crc kubenswrapper[4775]: I1002 03:24:07.957176 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="c4930148cbbbfb59b99fb94d29b145a85b1cfbffe3a70db63ae0fd48271fb062" exitCode=0 Oct 02 03:24:07 crc kubenswrapper[4775]: I1002 03:24:07.957240 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"c4930148cbbbfb59b99fb94d29b145a85b1cfbffe3a70db63ae0fd48271fb062"} Oct 02 03:24:07 crc kubenswrapper[4775]: I1002 03:24:07.957864 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769"} Oct 02 03:24:07 crc kubenswrapper[4775]: I1002 03:24:07.957900 4775 scope.go:117] "RemoveContainer" containerID="db8af7c8ca14f618348b65b685b769f9f4472d56444529cc140a47b697e24ab6" Oct 02 03:24:13 crc kubenswrapper[4775]: I1002 03:24:13.611743 4775 scope.go:117] "RemoveContainer" containerID="ca8a1b9c60e3cb73943079a5feebcddd6e13a38d4652d5a7dd75f6d2f56800dc" Oct 02 03:24:13 crc kubenswrapper[4775]: I1002 03:24:13.647034 4775 scope.go:117] "RemoveContainer" containerID="12e278188355751276f4519517fc8b44496a13bafdcb735332297fcc0afb18f3" Oct 02 03:24:13 crc kubenswrapper[4775]: I1002 03:24:13.725313 4775 scope.go:117] "RemoveContainer" containerID="5f18572a1e2a6bf4b7db01b5cf5f77d2ded1434fe39bfaa5aca4a12897b7b110" Oct 02 03:24:13 crc kubenswrapper[4775]: I1002 03:24:13.780745 4775 scope.go:117] "RemoveContainer" containerID="f79193914f690f7de1e103b446483f0daf8e40b21386f110c8f0d4824bff35c3" Oct 02 03:24:13 crc kubenswrapper[4775]: I1002 03:24:13.827147 4775 scope.go:117] "RemoveContainer" containerID="4d6f5f14508cc10948f3ef5a7a4dbd1ff3ed6e25ed871567115099e112d9ceab" Oct 02 03:24:13 crc kubenswrapper[4775]: I1002 03:24:13.907402 4775 scope.go:117] "RemoveContainer" containerID="ce1ed864de6e669bebfce7170ddb56905622149fb28c1c5d1adf537e0f06c83c" Oct 02 03:24:13 crc kubenswrapper[4775]: I1002 03:24:13.931764 4775 scope.go:117] "RemoveContainer" containerID="fbe8b70ea6913b6eabce06048c0b6162a6cd2cd0250406c9ab4709c54ad75d8c" Oct 02 03:24:15 crc kubenswrapper[4775]: I1002 03:24:15.034221 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l7nrv"] Oct 02 03:24:15 crc kubenswrapper[4775]: I1002 03:24:15.043087 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-l7nrv"] Oct 02 03:24:15 crc kubenswrapper[4775]: I1002 03:24:15.777808 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2543f41-75cd-42d5-9f28-9b7c5f73756c" path="/var/lib/kubelet/pods/c2543f41-75cd-42d5-9f28-9b7c5f73756c/volumes" Oct 02 03:24:16 crc kubenswrapper[4775]: I1002 03:24:16.045423 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-ft765"] Oct 02 03:24:16 crc kubenswrapper[4775]: I1002 03:24:16.052715 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-ft765"] Oct 02 03:24:16 crc kubenswrapper[4775]: I1002 03:24:16.443872 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-54bc95c9fb-lzh9v" Oct 02 03:24:17 crc kubenswrapper[4775]: I1002 03:24:17.783700 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a0025f0-7fa5-4798-b9ec-09a3cec748d7" path="/var/lib/kubelet/pods/9a0025f0-7fa5-4798-b9ec-09a3cec748d7/volumes" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.098244 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.098746 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="bcc7bd03-693b-4733-9f7f-dbd7fd15e490" containerName="openstackclient" containerID="cri-o://d89c6c8b2bc6e48a2f1b06ce8dc58577b83fcdb8ccb8e824108f02adb5655b1d" gracePeriod=2 Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.110809 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.145797 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 02 03:24:19 crc kubenswrapper[4775]: E1002 03:24:19.146222 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcc7bd03-693b-4733-9f7f-dbd7fd15e490" containerName="openstackclient" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.146240 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcc7bd03-693b-4733-9f7f-dbd7fd15e490" containerName="openstackclient" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.146469 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcc7bd03-693b-4733-9f7f-dbd7fd15e490" containerName="openstackclient" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.147179 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.152241 4775 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="bcc7bd03-693b-4733-9f7f-dbd7fd15e490" podUID="d77ec247-8a1e-4882-9017-755408b8dc34" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.154889 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.322337 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.323513 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.324626 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d77ec247-8a1e-4882-9017-755408b8dc34-openstack-config-secret\") pod \"openstackclient\" (UID: \"d77ec247-8a1e-4882-9017-755408b8dc34\") " pod="openstack/openstackclient" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.324703 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d77ec247-8a1e-4882-9017-755408b8dc34-openstack-config\") pod \"openstackclient\" (UID: \"d77ec247-8a1e-4882-9017-755408b8dc34\") " pod="openstack/openstackclient" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.325046 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvffz\" (UniqueName: \"kubernetes.io/projected/d77ec247-8a1e-4882-9017-755408b8dc34-kube-api-access-kvffz\") pod \"openstackclient\" (UID: \"d77ec247-8a1e-4882-9017-755408b8dc34\") " pod="openstack/openstackclient" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.327836 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-nnfmg" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.341846 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.427591 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvffz\" (UniqueName: \"kubernetes.io/projected/d77ec247-8a1e-4882-9017-755408b8dc34-kube-api-access-kvffz\") pod \"openstackclient\" (UID: \"d77ec247-8a1e-4882-9017-755408b8dc34\") " pod="openstack/openstackclient" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.427728 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d77ec247-8a1e-4882-9017-755408b8dc34-openstack-config-secret\") pod \"openstackclient\" (UID: \"d77ec247-8a1e-4882-9017-755408b8dc34\") " pod="openstack/openstackclient" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.427763 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d77ec247-8a1e-4882-9017-755408b8dc34-openstack-config\") pod \"openstackclient\" (UID: \"d77ec247-8a1e-4882-9017-755408b8dc34\") " pod="openstack/openstackclient" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.427832 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxg4p\" (UniqueName: \"kubernetes.io/projected/3fa0fce1-781d-41a9-8aaf-8c0589df9cf8-kube-api-access-wxg4p\") pod \"kube-state-metrics-0\" (UID: \"3fa0fce1-781d-41a9-8aaf-8c0589df9cf8\") " pod="openstack/kube-state-metrics-0" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.429090 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d77ec247-8a1e-4882-9017-755408b8dc34-openstack-config\") pod \"openstackclient\" (UID: \"d77ec247-8a1e-4882-9017-755408b8dc34\") " pod="openstack/openstackclient" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.445511 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d77ec247-8a1e-4882-9017-755408b8dc34-openstack-config-secret\") pod \"openstackclient\" (UID: \"d77ec247-8a1e-4882-9017-755408b8dc34\") " pod="openstack/openstackclient" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.462142 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvffz\" (UniqueName: \"kubernetes.io/projected/d77ec247-8a1e-4882-9017-755408b8dc34-kube-api-access-kvffz\") pod \"openstackclient\" (UID: \"d77ec247-8a1e-4882-9017-755408b8dc34\") " pod="openstack/openstackclient" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.471409 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.529144 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxg4p\" (UniqueName: \"kubernetes.io/projected/3fa0fce1-781d-41a9-8aaf-8c0589df9cf8-kube-api-access-wxg4p\") pod \"kube-state-metrics-0\" (UID: \"3fa0fce1-781d-41a9-8aaf-8c0589df9cf8\") " pod="openstack/kube-state-metrics-0" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.559124 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxg4p\" (UniqueName: \"kubernetes.io/projected/3fa0fce1-781d-41a9-8aaf-8c0589df9cf8-kube-api-access-wxg4p\") pod \"kube-state-metrics-0\" (UID: \"3fa0fce1-781d-41a9-8aaf-8c0589df9cf8\") " pod="openstack/kube-state-metrics-0" Oct 02 03:24:19 crc kubenswrapper[4775]: I1002 03:24:19.641239 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.041110 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.045409 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.054064 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.054306 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.054442 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-d6s2h" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.054518 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.070759 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.157640 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.157713 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbbsf\" (UniqueName: \"kubernetes.io/projected/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-kube-api-access-fbbsf\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.158210 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.158245 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.158306 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.158335 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.259503 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.259579 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.259667 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.259716 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbbsf\" (UniqueName: \"kubernetes.io/projected/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-kube-api-access-fbbsf\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.259778 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.259795 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.260591 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.265148 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.265937 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.267023 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.269672 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.290528 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbbsf\" (UniqueName: \"kubernetes.io/projected/2e0082f9-da1f-4b8d-8177-aebe807d9a1f-kube-api-access-fbbsf\") pod \"alertmanager-metric-storage-0\" (UID: \"2e0082f9-da1f-4b8d-8177-aebe807d9a1f\") " pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.358563 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.376300 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.377942 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: W1002 03:24:20.381707 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3fa0fce1_781d_41a9_8aaf_8c0589df9cf8.slice/crio-53a67656ff015c416a7d9567636dcfecf5e30ab21bb74153f72dc2b0de4bdf32 WatchSource:0}: Error finding container 53a67656ff015c416a7d9567636dcfecf5e30ab21bb74153f72dc2b0de4bdf32: Status 404 returned error can't find the container with id 53a67656ff015c416a7d9567636dcfecf5e30ab21bb74153f72dc2b0de4bdf32 Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.594356 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.604653 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.612097 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.619102 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.619155 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.619232 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.619352 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-8znlr" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.630974 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.645659 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.804122 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/73818738-ba3d-43f6-b0d1-9c96fd768406-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.804468 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/73818738-ba3d-43f6-b0d1-9c96fd768406-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.804496 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/73818738-ba3d-43f6-b0d1-9c96fd768406-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.804516 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/73818738-ba3d-43f6-b0d1-9c96fd768406-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.804536 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/73818738-ba3d-43f6-b0d1-9c96fd768406-config\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.804565 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/73818738-ba3d-43f6-b0d1-9c96fd768406-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.804590 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzk4p\" (UniqueName: \"kubernetes.io/projected/73818738-ba3d-43f6-b0d1-9c96fd768406-kube-api-access-tzk4p\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.804617 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-2bacc16e-1414-4a2c-89e5-1bd07db56426\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bacc16e-1414-4a2c-89e5-1bd07db56426\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.906835 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/73818738-ba3d-43f6-b0d1-9c96fd768406-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.906931 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/73818738-ba3d-43f6-b0d1-9c96fd768406-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.906970 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/73818738-ba3d-43f6-b0d1-9c96fd768406-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.906989 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/73818738-ba3d-43f6-b0d1-9c96fd768406-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.907008 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/73818738-ba3d-43f6-b0d1-9c96fd768406-config\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.907037 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/73818738-ba3d-43f6-b0d1-9c96fd768406-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.907077 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzk4p\" (UniqueName: \"kubernetes.io/projected/73818738-ba3d-43f6-b0d1-9c96fd768406-kube-api-access-tzk4p\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.907105 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-2bacc16e-1414-4a2c-89e5-1bd07db56426\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bacc16e-1414-4a2c-89e5-1bd07db56426\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.909877 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/73818738-ba3d-43f6-b0d1-9c96fd768406-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.940092 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/73818738-ba3d-43f6-b0d1-9c96fd768406-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.946574 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/73818738-ba3d-43f6-b0d1-9c96fd768406-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.947652 4775 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.947713 4775 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-2bacc16e-1414-4a2c-89e5-1bd07db56426\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bacc16e-1414-4a2c-89e5-1bd07db56426\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a9ef6a06bf21ded4c511f36c1b7007e3c76bea9ddaa41988a6475785b1fda122/globalmount\"" pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.974281 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/73818738-ba3d-43f6-b0d1-9c96fd768406-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.975499 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/73818738-ba3d-43f6-b0d1-9c96fd768406-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:20 crc kubenswrapper[4775]: I1002 03:24:20.995684 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzk4p\" (UniqueName: \"kubernetes.io/projected/73818738-ba3d-43f6-b0d1-9c96fd768406-kube-api-access-tzk4p\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.013574 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/73818738-ba3d-43f6-b0d1-9c96fd768406-config\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.161291 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3fa0fce1-781d-41a9-8aaf-8c0589df9cf8","Type":"ContainerStarted","Data":"53a67656ff015c416a7d9567636dcfecf5e30ab21bb74153f72dc2b0de4bdf32"} Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.187842 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"d77ec247-8a1e-4882-9017-755408b8dc34","Type":"ContainerStarted","Data":"ba83fd96e54db76926e963b9924befb374bd6da927ccdcb393cc5c7282c3720f"} Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.351614 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-2bacc16e-1414-4a2c-89e5-1bd07db56426\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-2bacc16e-1414-4a2c-89e5-1bd07db56426\") pod \"prometheus-metric-storage-0\" (UID: \"73818738-ba3d-43f6-b0d1-9c96fd768406\") " pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.361078 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.535214 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.723887 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.854009 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp6qw\" (UniqueName: \"kubernetes.io/projected/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-kube-api-access-zp6qw\") pod \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\" (UID: \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\") " Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.854479 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-openstack-config-secret\") pod \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\" (UID: \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\") " Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.854604 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-openstack-config\") pod \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\" (UID: \"bcc7bd03-693b-4733-9f7f-dbd7fd15e490\") " Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.882908 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-kube-api-access-zp6qw" (OuterVolumeSpecName: "kube-api-access-zp6qw") pod "bcc7bd03-693b-4733-9f7f-dbd7fd15e490" (UID: "bcc7bd03-693b-4733-9f7f-dbd7fd15e490"). InnerVolumeSpecName "kube-api-access-zp6qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.897188 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "bcc7bd03-693b-4733-9f7f-dbd7fd15e490" (UID: "bcc7bd03-693b-4733-9f7f-dbd7fd15e490"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.944500 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "bcc7bd03-693b-4733-9f7f-dbd7fd15e490" (UID: "bcc7bd03-693b-4733-9f7f-dbd7fd15e490"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.957189 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.957210 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:24:21 crc kubenswrapper[4775]: I1002 03:24:21.957220 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp6qw\" (UniqueName: \"kubernetes.io/projected/bcc7bd03-693b-4733-9f7f-dbd7fd15e490-kube-api-access-zp6qw\") on node \"crc\" DevicePath \"\"" Oct 02 03:24:22 crc kubenswrapper[4775]: W1002 03:24:22.130406 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73818738_ba3d_43f6_b0d1_9c96fd768406.slice/crio-2fa4b7972d43013a1fda1fac9af78801addf058b006f758d9a2b11a9fcb810f4 WatchSource:0}: Error finding container 2fa4b7972d43013a1fda1fac9af78801addf058b006f758d9a2b11a9fcb810f4: Status 404 returned error can't find the container with id 2fa4b7972d43013a1fda1fac9af78801addf058b006f758d9a2b11a9fcb810f4 Oct 02 03:24:22 crc kubenswrapper[4775]: I1002 03:24:22.131813 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Oct 02 03:24:22 crc kubenswrapper[4775]: I1002 03:24:22.203482 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"2e0082f9-da1f-4b8d-8177-aebe807d9a1f","Type":"ContainerStarted","Data":"639b854871bd8648a07443a2c3e109fabe366a25d893bdb480fd00bcc6ca0a45"} Oct 02 03:24:22 crc kubenswrapper[4775]: I1002 03:24:22.205764 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"d77ec247-8a1e-4882-9017-755408b8dc34","Type":"ContainerStarted","Data":"cc71990c66bb69d70a9290cb4548156894d2de416a21075c3149e4612ced3c64"} Oct 02 03:24:22 crc kubenswrapper[4775]: I1002 03:24:22.208598 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"73818738-ba3d-43f6-b0d1-9c96fd768406","Type":"ContainerStarted","Data":"2fa4b7972d43013a1fda1fac9af78801addf058b006f758d9a2b11a9fcb810f4"} Oct 02 03:24:22 crc kubenswrapper[4775]: I1002 03:24:22.209993 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3fa0fce1-781d-41a9-8aaf-8c0589df9cf8","Type":"ContainerStarted","Data":"26c748cfc3b11fcf89f4b6744b84c7ebe86be909363ae8ec6aa0272984365a86"} Oct 02 03:24:22 crc kubenswrapper[4775]: I1002 03:24:22.210187 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 02 03:24:22 crc kubenswrapper[4775]: I1002 03:24:22.211602 4775 generic.go:334] "Generic (PLEG): container finished" podID="bcc7bd03-693b-4733-9f7f-dbd7fd15e490" containerID="d89c6c8b2bc6e48a2f1b06ce8dc58577b83fcdb8ccb8e824108f02adb5655b1d" exitCode=137 Oct 02 03:24:22 crc kubenswrapper[4775]: I1002 03:24:22.211667 4775 scope.go:117] "RemoveContainer" containerID="d89c6c8b2bc6e48a2f1b06ce8dc58577b83fcdb8ccb8e824108f02adb5655b1d" Oct 02 03:24:22 crc kubenswrapper[4775]: I1002 03:24:22.211773 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 02 03:24:22 crc kubenswrapper[4775]: I1002 03:24:22.253926 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.253902218 podStartE2EDuration="3.253902218s" podCreationTimestamp="2025-10-02 03:24:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:24:22.218773948 +0000 UTC m=+6199.385517988" watchObservedRunningTime="2025-10-02 03:24:22.253902218 +0000 UTC m=+6199.420646258" Oct 02 03:24:22 crc kubenswrapper[4775]: I1002 03:24:22.258440 4775 scope.go:117] "RemoveContainer" containerID="d89c6c8b2bc6e48a2f1b06ce8dc58577b83fcdb8ccb8e824108f02adb5655b1d" Oct 02 03:24:22 crc kubenswrapper[4775]: E1002 03:24:22.258742 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d89c6c8b2bc6e48a2f1b06ce8dc58577b83fcdb8ccb8e824108f02adb5655b1d\": container with ID starting with d89c6c8b2bc6e48a2f1b06ce8dc58577b83fcdb8ccb8e824108f02adb5655b1d not found: ID does not exist" containerID="d89c6c8b2bc6e48a2f1b06ce8dc58577b83fcdb8ccb8e824108f02adb5655b1d" Oct 02 03:24:22 crc kubenswrapper[4775]: I1002 03:24:22.258772 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d89c6c8b2bc6e48a2f1b06ce8dc58577b83fcdb8ccb8e824108f02adb5655b1d"} err="failed to get container status \"d89c6c8b2bc6e48a2f1b06ce8dc58577b83fcdb8ccb8e824108f02adb5655b1d\": rpc error: code = NotFound desc = could not find container \"d89c6c8b2bc6e48a2f1b06ce8dc58577b83fcdb8ccb8e824108f02adb5655b1d\": container with ID starting with d89c6c8b2bc6e48a2f1b06ce8dc58577b83fcdb8ccb8e824108f02adb5655b1d not found: ID does not exist" Oct 02 03:24:22 crc kubenswrapper[4775]: I1002 03:24:22.261341 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.589828042 podStartE2EDuration="3.26132629s" podCreationTimestamp="2025-10-02 03:24:19 +0000 UTC" firstStartedPulling="2025-10-02 03:24:20.401320521 +0000 UTC m=+6197.568064561" lastFinishedPulling="2025-10-02 03:24:21.072818769 +0000 UTC m=+6198.239562809" observedRunningTime="2025-10-02 03:24:22.233765256 +0000 UTC m=+6199.400509296" watchObservedRunningTime="2025-10-02 03:24:22.26132629 +0000 UTC m=+6199.428070330" Oct 02 03:24:23 crc kubenswrapper[4775]: I1002 03:24:23.780482 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcc7bd03-693b-4733-9f7f-dbd7fd15e490" path="/var/lib/kubelet/pods/bcc7bd03-693b-4733-9f7f-dbd7fd15e490/volumes" Oct 02 03:24:28 crc kubenswrapper[4775]: I1002 03:24:28.056151 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-99l8t"] Oct 02 03:24:28 crc kubenswrapper[4775]: I1002 03:24:28.065494 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-99l8t"] Oct 02 03:24:28 crc kubenswrapper[4775]: I1002 03:24:28.313100 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"2e0082f9-da1f-4b8d-8177-aebe807d9a1f","Type":"ContainerStarted","Data":"59711fcb1be08bca9c58d1456f2c9778f07092662de38d86306afdafd3a19176"} Oct 02 03:24:29 crc kubenswrapper[4775]: I1002 03:24:29.334437 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"73818738-ba3d-43f6-b0d1-9c96fd768406","Type":"ContainerStarted","Data":"76277e6cda5495b6d6304264309808ed2f2ef3be6f6c716c1213fc675947ea3c"} Oct 02 03:24:29 crc kubenswrapper[4775]: I1002 03:24:29.650769 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 02 03:24:29 crc kubenswrapper[4775]: I1002 03:24:29.777140 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a133e8d0-41e4-46fe-b32c-81d5d6fe0888" path="/var/lib/kubelet/pods/a133e8d0-41e4-46fe-b32c-81d5d6fe0888/volumes" Oct 02 03:24:38 crc kubenswrapper[4775]: I1002 03:24:38.474068 4775 generic.go:334] "Generic (PLEG): container finished" podID="73818738-ba3d-43f6-b0d1-9c96fd768406" containerID="76277e6cda5495b6d6304264309808ed2f2ef3be6f6c716c1213fc675947ea3c" exitCode=0 Oct 02 03:24:38 crc kubenswrapper[4775]: I1002 03:24:38.475068 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"73818738-ba3d-43f6-b0d1-9c96fd768406","Type":"ContainerDied","Data":"76277e6cda5495b6d6304264309808ed2f2ef3be6f6c716c1213fc675947ea3c"} Oct 02 03:24:40 crc kubenswrapper[4775]: I1002 03:24:40.517540 4775 generic.go:334] "Generic (PLEG): container finished" podID="2e0082f9-da1f-4b8d-8177-aebe807d9a1f" containerID="59711fcb1be08bca9c58d1456f2c9778f07092662de38d86306afdafd3a19176" exitCode=0 Oct 02 03:24:40 crc kubenswrapper[4775]: I1002 03:24:40.517618 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"2e0082f9-da1f-4b8d-8177-aebe807d9a1f","Type":"ContainerDied","Data":"59711fcb1be08bca9c58d1456f2c9778f07092662de38d86306afdafd3a19176"} Oct 02 03:24:44 crc kubenswrapper[4775]: I1002 03:24:44.565085 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"73818738-ba3d-43f6-b0d1-9c96fd768406","Type":"ContainerStarted","Data":"1188dc295b111e9d07df5b63404a64c334b9bb78f72be6e24943a5d60981ac52"} Oct 02 03:24:44 crc kubenswrapper[4775]: I1002 03:24:44.567208 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"2e0082f9-da1f-4b8d-8177-aebe807d9a1f","Type":"ContainerStarted","Data":"3c30ff84cecc2ba88c99e2dbed51b85d9a5a7a7fe0219c384e6fb6ae99f14be9"} Oct 02 03:24:48 crc kubenswrapper[4775]: I1002 03:24:48.638828 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"73818738-ba3d-43f6-b0d1-9c96fd768406","Type":"ContainerStarted","Data":"0031d878851dc5a8cea2fe636203a067709047966f63c01e877c13e7a179f95b"} Oct 02 03:24:49 crc kubenswrapper[4775]: I1002 03:24:49.664146 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"2e0082f9-da1f-4b8d-8177-aebe807d9a1f","Type":"ContainerStarted","Data":"caf3afe5b3830fe6f222a33e86cd775c1b89b55d0f7628fb27b38e550aac4b19"} Oct 02 03:24:49 crc kubenswrapper[4775]: I1002 03:24:49.664227 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:49 crc kubenswrapper[4775]: I1002 03:24:49.671409 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Oct 02 03:24:49 crc kubenswrapper[4775]: I1002 03:24:49.725718 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=6.914341744 podStartE2EDuration="29.725696021s" podCreationTimestamp="2025-10-02 03:24:20 +0000 UTC" firstStartedPulling="2025-10-02 03:24:21.350918093 +0000 UTC m=+6198.517662133" lastFinishedPulling="2025-10-02 03:24:44.16227236 +0000 UTC m=+6221.329016410" observedRunningTime="2025-10-02 03:24:49.697037558 +0000 UTC m=+6226.863781638" watchObservedRunningTime="2025-10-02 03:24:49.725696021 +0000 UTC m=+6226.892440071" Oct 02 03:24:51 crc kubenswrapper[4775]: I1002 03:24:51.687416 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"73818738-ba3d-43f6-b0d1-9c96fd768406","Type":"ContainerStarted","Data":"bbb69073f07c7f7f05f065cd05ac4250caae736c8d5651cb47e5d9756806e2e4"} Oct 02 03:24:51 crc kubenswrapper[4775]: I1002 03:24:51.732751 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.5439487830000003 podStartE2EDuration="32.732733169s" podCreationTimestamp="2025-10-02 03:24:19 +0000 UTC" firstStartedPulling="2025-10-02 03:24:22.138870358 +0000 UTC m=+6199.305614398" lastFinishedPulling="2025-10-02 03:24:51.327654744 +0000 UTC m=+6228.494398784" observedRunningTime="2025-10-02 03:24:51.718679555 +0000 UTC m=+6228.885423585" watchObservedRunningTime="2025-10-02 03:24:51.732733169 +0000 UTC m=+6228.899477209" Oct 02 03:24:56 crc kubenswrapper[4775]: I1002 03:24:56.535486 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Oct 02 03:24:56 crc kubenswrapper[4775]: I1002 03:24:56.922664 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:24:56 crc kubenswrapper[4775]: I1002 03:24:56.925334 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 03:24:56 crc kubenswrapper[4775]: I1002 03:24:56.927467 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 03:24:56 crc kubenswrapper[4775]: I1002 03:24:56.927518 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 03:24:56 crc kubenswrapper[4775]: I1002 03:24:56.937808 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.081698 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.081764 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-scripts\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.081860 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.081889 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e77c3fe-d4ee-4b70-813e-087f3803a65b-run-httpd\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.081905 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnlmg\" (UniqueName: \"kubernetes.io/projected/2e77c3fe-d4ee-4b70-813e-087f3803a65b-kube-api-access-pnlmg\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.082044 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e77c3fe-d4ee-4b70-813e-087f3803a65b-log-httpd\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.082153 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-config-data\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.183520 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.183577 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-scripts\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.183651 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.183677 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e77c3fe-d4ee-4b70-813e-087f3803a65b-run-httpd\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.183692 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnlmg\" (UniqueName: \"kubernetes.io/projected/2e77c3fe-d4ee-4b70-813e-087f3803a65b-kube-api-access-pnlmg\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.183724 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e77c3fe-d4ee-4b70-813e-087f3803a65b-log-httpd\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.183756 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-config-data\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.184384 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e77c3fe-d4ee-4b70-813e-087f3803a65b-run-httpd\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.184640 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e77c3fe-d4ee-4b70-813e-087f3803a65b-log-httpd\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.191660 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.196182 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-scripts\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.199139 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnlmg\" (UniqueName: \"kubernetes.io/projected/2e77c3fe-d4ee-4b70-813e-087f3803a65b-kube-api-access-pnlmg\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.200801 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-config-data\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.203539 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.245547 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 03:24:57 crc kubenswrapper[4775]: I1002 03:24:57.754975 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:24:58 crc kubenswrapper[4775]: I1002 03:24:58.772890 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e77c3fe-d4ee-4b70-813e-087f3803a65b","Type":"ContainerStarted","Data":"bfdef500eaece7f7830bc337caa87ec969ec4d7e93370e120fe226d7a0008671"} Oct 02 03:24:58 crc kubenswrapper[4775]: I1002 03:24:58.774550 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e77c3fe-d4ee-4b70-813e-087f3803a65b","Type":"ContainerStarted","Data":"7f2301bf621615c887e521dc87621a84cb172ad23838620fae2c318ac607743b"} Oct 02 03:24:59 crc kubenswrapper[4775]: I1002 03:24:59.783911 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e77c3fe-d4ee-4b70-813e-087f3803a65b","Type":"ContainerStarted","Data":"60ed7d313741ef68eb33311bc5902260e621602438bbc25f94495da1a6cff306"} Oct 02 03:25:00 crc kubenswrapper[4775]: I1002 03:25:00.796578 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e77c3fe-d4ee-4b70-813e-087f3803a65b","Type":"ContainerStarted","Data":"cde04bcfe6831402ac5d46a91be7dc401194428a7456aecf4f9f541e995a4683"} Oct 02 03:25:05 crc kubenswrapper[4775]: I1002 03:25:05.869395 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e77c3fe-d4ee-4b70-813e-087f3803a65b","Type":"ContainerStarted","Data":"1b768b2912fc328d8ebe6baf56a330f9e136b8ca0216da0b519b91ac5ee113b4"} Oct 02 03:25:05 crc kubenswrapper[4775]: I1002 03:25:05.870361 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 03:25:06 crc kubenswrapper[4775]: I1002 03:25:06.536454 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Oct 02 03:25:06 crc kubenswrapper[4775]: I1002 03:25:06.538759 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Oct 02 03:25:06 crc kubenswrapper[4775]: I1002 03:25:06.570241 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.4729404280000002 podStartE2EDuration="10.570224427s" podCreationTimestamp="2025-10-02 03:24:56 +0000 UTC" firstStartedPulling="2025-10-02 03:24:57.762151453 +0000 UTC m=+6234.928895503" lastFinishedPulling="2025-10-02 03:25:04.859435472 +0000 UTC m=+6242.026179502" observedRunningTime="2025-10-02 03:25:05.923404708 +0000 UTC m=+6243.090148758" watchObservedRunningTime="2025-10-02 03:25:06.570224427 +0000 UTC m=+6243.736968467" Oct 02 03:25:06 crc kubenswrapper[4775]: I1002 03:25:06.879208 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Oct 02 03:25:09 crc kubenswrapper[4775]: I1002 03:25:09.248689 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-tl697"] Oct 02 03:25:09 crc kubenswrapper[4775]: I1002 03:25:09.251393 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-tl697" Oct 02 03:25:09 crc kubenswrapper[4775]: I1002 03:25:09.265072 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-tl697"] Oct 02 03:25:09 crc kubenswrapper[4775]: I1002 03:25:09.395564 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scwsf\" (UniqueName: \"kubernetes.io/projected/42bb3fa6-b8c1-478f-9b7e-7986a17501c3-kube-api-access-scwsf\") pod \"aodh-db-create-tl697\" (UID: \"42bb3fa6-b8c1-478f-9b7e-7986a17501c3\") " pod="openstack/aodh-db-create-tl697" Oct 02 03:25:09 crc kubenswrapper[4775]: I1002 03:25:09.497427 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scwsf\" (UniqueName: \"kubernetes.io/projected/42bb3fa6-b8c1-478f-9b7e-7986a17501c3-kube-api-access-scwsf\") pod \"aodh-db-create-tl697\" (UID: \"42bb3fa6-b8c1-478f-9b7e-7986a17501c3\") " pod="openstack/aodh-db-create-tl697" Oct 02 03:25:09 crc kubenswrapper[4775]: I1002 03:25:09.538779 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scwsf\" (UniqueName: \"kubernetes.io/projected/42bb3fa6-b8c1-478f-9b7e-7986a17501c3-kube-api-access-scwsf\") pod \"aodh-db-create-tl697\" (UID: \"42bb3fa6-b8c1-478f-9b7e-7986a17501c3\") " pod="openstack/aodh-db-create-tl697" Oct 02 03:25:09 crc kubenswrapper[4775]: I1002 03:25:09.588815 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-tl697" Oct 02 03:25:10 crc kubenswrapper[4775]: I1002 03:25:10.068615 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-tl697"] Oct 02 03:25:10 crc kubenswrapper[4775]: I1002 03:25:10.929335 4775 generic.go:334] "Generic (PLEG): container finished" podID="42bb3fa6-b8c1-478f-9b7e-7986a17501c3" containerID="52680f1826704472e47edf9cbbb4bf827264949fa8bce2bc4fea413c313edc5c" exitCode=0 Oct 02 03:25:10 crc kubenswrapper[4775]: I1002 03:25:10.929415 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-tl697" event={"ID":"42bb3fa6-b8c1-478f-9b7e-7986a17501c3","Type":"ContainerDied","Data":"52680f1826704472e47edf9cbbb4bf827264949fa8bce2bc4fea413c313edc5c"} Oct 02 03:25:10 crc kubenswrapper[4775]: I1002 03:25:10.929726 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-tl697" event={"ID":"42bb3fa6-b8c1-478f-9b7e-7986a17501c3","Type":"ContainerStarted","Data":"753ac4beb92a82fee476e92574ca7092fd8bc5c6dc65dfc86c73d39789fec395"} Oct 02 03:25:12 crc kubenswrapper[4775]: I1002 03:25:12.437183 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-tl697" Oct 02 03:25:12 crc kubenswrapper[4775]: I1002 03:25:12.569464 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scwsf\" (UniqueName: \"kubernetes.io/projected/42bb3fa6-b8c1-478f-9b7e-7986a17501c3-kube-api-access-scwsf\") pod \"42bb3fa6-b8c1-478f-9b7e-7986a17501c3\" (UID: \"42bb3fa6-b8c1-478f-9b7e-7986a17501c3\") " Oct 02 03:25:12 crc kubenswrapper[4775]: I1002 03:25:12.580500 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42bb3fa6-b8c1-478f-9b7e-7986a17501c3-kube-api-access-scwsf" (OuterVolumeSpecName: "kube-api-access-scwsf") pod "42bb3fa6-b8c1-478f-9b7e-7986a17501c3" (UID: "42bb3fa6-b8c1-478f-9b7e-7986a17501c3"). InnerVolumeSpecName "kube-api-access-scwsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:25:12 crc kubenswrapper[4775]: I1002 03:25:12.674599 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scwsf\" (UniqueName: \"kubernetes.io/projected/42bb3fa6-b8c1-478f-9b7e-7986a17501c3-kube-api-access-scwsf\") on node \"crc\" DevicePath \"\"" Oct 02 03:25:12 crc kubenswrapper[4775]: I1002 03:25:12.961857 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-tl697" Oct 02 03:25:12 crc kubenswrapper[4775]: I1002 03:25:12.961778 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-tl697" event={"ID":"42bb3fa6-b8c1-478f-9b7e-7986a17501c3","Type":"ContainerDied","Data":"753ac4beb92a82fee476e92574ca7092fd8bc5c6dc65dfc86c73d39789fec395"} Oct 02 03:25:12 crc kubenswrapper[4775]: I1002 03:25:12.961991 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="753ac4beb92a82fee476e92574ca7092fd8bc5c6dc65dfc86c73d39789fec395" Oct 02 03:25:14 crc kubenswrapper[4775]: I1002 03:25:14.096216 4775 scope.go:117] "RemoveContainer" containerID="ef652e1657b9f30f838a21acdcb8147d59deae3ab9a424a851ee7dffcd018736" Oct 02 03:25:14 crc kubenswrapper[4775]: I1002 03:25:14.179475 4775 scope.go:117] "RemoveContainer" containerID="bb9d8d149247f5f8edf1694299e0391c3bcfa5804378d2ea7f3f39e6692b3aba" Oct 02 03:25:14 crc kubenswrapper[4775]: I1002 03:25:14.223010 4775 scope.go:117] "RemoveContainer" containerID="a3bd73887099309942bafcb6324a4078b20d9d9f62ba786f3e1506b9ac2d3ee4" Oct 02 03:25:14 crc kubenswrapper[4775]: I1002 03:25:14.294710 4775 scope.go:117] "RemoveContainer" containerID="bee34c6b58d6da6c70e62be6a6cd4b4aa98a51b42fbc8c8a91501d8fa0c59e65" Oct 02 03:25:15 crc kubenswrapper[4775]: I1002 03:25:15.043733 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-cpx6g"] Oct 02 03:25:15 crc kubenswrapper[4775]: I1002 03:25:15.056368 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-cpx6g"] Oct 02 03:25:15 crc kubenswrapper[4775]: I1002 03:25:15.812237 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84ad728b-c2c6-44a3-8fb9-23a29a82ef65" path="/var/lib/kubelet/pods/84ad728b-c2c6-44a3-8fb9-23a29a82ef65/volumes" Oct 02 03:25:19 crc kubenswrapper[4775]: I1002 03:25:19.409900 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0319-account-create-fcjrh"] Oct 02 03:25:19 crc kubenswrapper[4775]: E1002 03:25:19.411288 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42bb3fa6-b8c1-478f-9b7e-7986a17501c3" containerName="mariadb-database-create" Oct 02 03:25:19 crc kubenswrapper[4775]: I1002 03:25:19.411333 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="42bb3fa6-b8c1-478f-9b7e-7986a17501c3" containerName="mariadb-database-create" Oct 02 03:25:19 crc kubenswrapper[4775]: I1002 03:25:19.411713 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="42bb3fa6-b8c1-478f-9b7e-7986a17501c3" containerName="mariadb-database-create" Oct 02 03:25:19 crc kubenswrapper[4775]: I1002 03:25:19.413081 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0319-account-create-fcjrh" Oct 02 03:25:19 crc kubenswrapper[4775]: I1002 03:25:19.423664 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Oct 02 03:25:19 crc kubenswrapper[4775]: I1002 03:25:19.428063 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0319-account-create-fcjrh"] Oct 02 03:25:19 crc kubenswrapper[4775]: I1002 03:25:19.488760 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fh2c\" (UniqueName: \"kubernetes.io/projected/2006decc-1f50-44fe-ad50-b3dd82beb654-kube-api-access-7fh2c\") pod \"aodh-0319-account-create-fcjrh\" (UID: \"2006decc-1f50-44fe-ad50-b3dd82beb654\") " pod="openstack/aodh-0319-account-create-fcjrh" Oct 02 03:25:19 crc kubenswrapper[4775]: I1002 03:25:19.591482 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fh2c\" (UniqueName: \"kubernetes.io/projected/2006decc-1f50-44fe-ad50-b3dd82beb654-kube-api-access-7fh2c\") pod \"aodh-0319-account-create-fcjrh\" (UID: \"2006decc-1f50-44fe-ad50-b3dd82beb654\") " pod="openstack/aodh-0319-account-create-fcjrh" Oct 02 03:25:19 crc kubenswrapper[4775]: I1002 03:25:19.625728 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fh2c\" (UniqueName: \"kubernetes.io/projected/2006decc-1f50-44fe-ad50-b3dd82beb654-kube-api-access-7fh2c\") pod \"aodh-0319-account-create-fcjrh\" (UID: \"2006decc-1f50-44fe-ad50-b3dd82beb654\") " pod="openstack/aodh-0319-account-create-fcjrh" Oct 02 03:25:19 crc kubenswrapper[4775]: I1002 03:25:19.757072 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0319-account-create-fcjrh" Oct 02 03:25:20 crc kubenswrapper[4775]: I1002 03:25:20.296824 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0319-account-create-fcjrh"] Oct 02 03:25:21 crc kubenswrapper[4775]: I1002 03:25:21.062513 4775 generic.go:334] "Generic (PLEG): container finished" podID="2006decc-1f50-44fe-ad50-b3dd82beb654" containerID="e4b66c8cb8a64de4a941e5d2473508aa6b43c17777fe3600d699c1bcbafd5880" exitCode=0 Oct 02 03:25:21 crc kubenswrapper[4775]: I1002 03:25:21.062564 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0319-account-create-fcjrh" event={"ID":"2006decc-1f50-44fe-ad50-b3dd82beb654","Type":"ContainerDied","Data":"e4b66c8cb8a64de4a941e5d2473508aa6b43c17777fe3600d699c1bcbafd5880"} Oct 02 03:25:21 crc kubenswrapper[4775]: I1002 03:25:21.062780 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0319-account-create-fcjrh" event={"ID":"2006decc-1f50-44fe-ad50-b3dd82beb654","Type":"ContainerStarted","Data":"716f12259cea7b9af00f6d2014fff917ab8c2d8f1db581595a0c9b78f908ab67"} Oct 02 03:25:22 crc kubenswrapper[4775]: I1002 03:25:22.581500 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0319-account-create-fcjrh" Oct 02 03:25:22 crc kubenswrapper[4775]: I1002 03:25:22.765086 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fh2c\" (UniqueName: \"kubernetes.io/projected/2006decc-1f50-44fe-ad50-b3dd82beb654-kube-api-access-7fh2c\") pod \"2006decc-1f50-44fe-ad50-b3dd82beb654\" (UID: \"2006decc-1f50-44fe-ad50-b3dd82beb654\") " Oct 02 03:25:22 crc kubenswrapper[4775]: I1002 03:25:22.780369 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2006decc-1f50-44fe-ad50-b3dd82beb654-kube-api-access-7fh2c" (OuterVolumeSpecName: "kube-api-access-7fh2c") pod "2006decc-1f50-44fe-ad50-b3dd82beb654" (UID: "2006decc-1f50-44fe-ad50-b3dd82beb654"). InnerVolumeSpecName "kube-api-access-7fh2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:25:22 crc kubenswrapper[4775]: I1002 03:25:22.871490 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fh2c\" (UniqueName: \"kubernetes.io/projected/2006decc-1f50-44fe-ad50-b3dd82beb654-kube-api-access-7fh2c\") on node \"crc\" DevicePath \"\"" Oct 02 03:25:23 crc kubenswrapper[4775]: I1002 03:25:23.099369 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0319-account-create-fcjrh" event={"ID":"2006decc-1f50-44fe-ad50-b3dd82beb654","Type":"ContainerDied","Data":"716f12259cea7b9af00f6d2014fff917ab8c2d8f1db581595a0c9b78f908ab67"} Oct 02 03:25:23 crc kubenswrapper[4775]: I1002 03:25:23.099427 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="716f12259cea7b9af00f6d2014fff917ab8c2d8f1db581595a0c9b78f908ab67" Oct 02 03:25:23 crc kubenswrapper[4775]: I1002 03:25:23.099453 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0319-account-create-fcjrh" Oct 02 03:25:24 crc kubenswrapper[4775]: I1002 03:25:24.934532 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-j97tn"] Oct 02 03:25:24 crc kubenswrapper[4775]: E1002 03:25:24.935268 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2006decc-1f50-44fe-ad50-b3dd82beb654" containerName="mariadb-account-create" Oct 02 03:25:24 crc kubenswrapper[4775]: I1002 03:25:24.935284 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2006decc-1f50-44fe-ad50-b3dd82beb654" containerName="mariadb-account-create" Oct 02 03:25:24 crc kubenswrapper[4775]: I1002 03:25:24.935561 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2006decc-1f50-44fe-ad50-b3dd82beb654" containerName="mariadb-account-create" Oct 02 03:25:24 crc kubenswrapper[4775]: I1002 03:25:24.936438 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:24 crc kubenswrapper[4775]: I1002 03:25:24.940335 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 02 03:25:24 crc kubenswrapper[4775]: I1002 03:25:24.940576 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 02 03:25:24 crc kubenswrapper[4775]: I1002 03:25:24.941025 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-44ptd" Oct 02 03:25:24 crc kubenswrapper[4775]: I1002 03:25:24.945781 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-j97tn"] Oct 02 03:25:25 crc kubenswrapper[4775]: I1002 03:25:25.020087 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-config-data\") pod \"aodh-db-sync-j97tn\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:25 crc kubenswrapper[4775]: I1002 03:25:25.020147 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh49k\" (UniqueName: \"kubernetes.io/projected/903ed25a-6e5c-46c9-8aac-04c071cc2067-kube-api-access-dh49k\") pod \"aodh-db-sync-j97tn\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:25 crc kubenswrapper[4775]: I1002 03:25:25.020238 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-combined-ca-bundle\") pod \"aodh-db-sync-j97tn\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:25 crc kubenswrapper[4775]: I1002 03:25:25.020491 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-scripts\") pod \"aodh-db-sync-j97tn\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:25 crc kubenswrapper[4775]: I1002 03:25:25.122274 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-combined-ca-bundle\") pod \"aodh-db-sync-j97tn\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:25 crc kubenswrapper[4775]: I1002 03:25:25.122411 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-scripts\") pod \"aodh-db-sync-j97tn\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:25 crc kubenswrapper[4775]: I1002 03:25:25.122622 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-config-data\") pod \"aodh-db-sync-j97tn\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:25 crc kubenswrapper[4775]: I1002 03:25:25.122662 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh49k\" (UniqueName: \"kubernetes.io/projected/903ed25a-6e5c-46c9-8aac-04c071cc2067-kube-api-access-dh49k\") pod \"aodh-db-sync-j97tn\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:25 crc kubenswrapper[4775]: I1002 03:25:25.130822 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-combined-ca-bundle\") pod \"aodh-db-sync-j97tn\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:25 crc kubenswrapper[4775]: I1002 03:25:25.131326 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-scripts\") pod \"aodh-db-sync-j97tn\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:25 crc kubenswrapper[4775]: I1002 03:25:25.137154 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-config-data\") pod \"aodh-db-sync-j97tn\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:25 crc kubenswrapper[4775]: I1002 03:25:25.144030 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh49k\" (UniqueName: \"kubernetes.io/projected/903ed25a-6e5c-46c9-8aac-04c071cc2067-kube-api-access-dh49k\") pod \"aodh-db-sync-j97tn\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:25 crc kubenswrapper[4775]: I1002 03:25:25.271221 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:25 crc kubenswrapper[4775]: I1002 03:25:25.784821 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-j97tn"] Oct 02 03:25:26 crc kubenswrapper[4775]: I1002 03:25:26.035700 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-f5a2-account-create-bwcrk"] Oct 02 03:25:26 crc kubenswrapper[4775]: I1002 03:25:26.042916 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-f5a2-account-create-bwcrk"] Oct 02 03:25:26 crc kubenswrapper[4775]: I1002 03:25:26.155615 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-j97tn" event={"ID":"903ed25a-6e5c-46c9-8aac-04c071cc2067","Type":"ContainerStarted","Data":"5d685d8ac910546448bf01360ff61b57d5109c7f97b6824e5757e801d9e75d61"} Oct 02 03:25:27 crc kubenswrapper[4775]: I1002 03:25:27.259007 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 03:25:27 crc kubenswrapper[4775]: I1002 03:25:27.784449 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="013cf425-7401-4e5d-979f-d63a79d9db31" path="/var/lib/kubelet/pods/013cf425-7401-4e5d-979f-d63a79d9db31/volumes" Oct 02 03:25:31 crc kubenswrapper[4775]: I1002 03:25:31.228660 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-j97tn" event={"ID":"903ed25a-6e5c-46c9-8aac-04c071cc2067","Type":"ContainerStarted","Data":"85d10a99025652d3b33c7723f4114227c286c946ea888f1de64ff5f5218c2717"} Oct 02 03:25:31 crc kubenswrapper[4775]: I1002 03:25:31.254173 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-j97tn" podStartSLOduration=2.274559615 podStartE2EDuration="7.254149329s" podCreationTimestamp="2025-10-02 03:25:24 +0000 UTC" firstStartedPulling="2025-10-02 03:25:25.788653606 +0000 UTC m=+6262.955397646" lastFinishedPulling="2025-10-02 03:25:30.76824332 +0000 UTC m=+6267.934987360" observedRunningTime="2025-10-02 03:25:31.249318394 +0000 UTC m=+6268.416062434" watchObservedRunningTime="2025-10-02 03:25:31.254149329 +0000 UTC m=+6268.420893369" Oct 02 03:25:32 crc kubenswrapper[4775]: I1002 03:25:32.050595 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-fklbq"] Oct 02 03:25:32 crc kubenswrapper[4775]: I1002 03:25:32.061178 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-fklbq"] Oct 02 03:25:33 crc kubenswrapper[4775]: I1002 03:25:33.789475 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af1edb92-72cb-46db-9ad8-6e22d9206bc5" path="/var/lib/kubelet/pods/af1edb92-72cb-46db-9ad8-6e22d9206bc5/volumes" Oct 02 03:25:34 crc kubenswrapper[4775]: I1002 03:25:34.283764 4775 generic.go:334] "Generic (PLEG): container finished" podID="903ed25a-6e5c-46c9-8aac-04c071cc2067" containerID="85d10a99025652d3b33c7723f4114227c286c946ea888f1de64ff5f5218c2717" exitCode=0 Oct 02 03:25:34 crc kubenswrapper[4775]: I1002 03:25:34.283826 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-j97tn" event={"ID":"903ed25a-6e5c-46c9-8aac-04c071cc2067","Type":"ContainerDied","Data":"85d10a99025652d3b33c7723f4114227c286c946ea888f1de64ff5f5218c2717"} Oct 02 03:25:35 crc kubenswrapper[4775]: I1002 03:25:35.890146 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.038109 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-scripts\") pod \"903ed25a-6e5c-46c9-8aac-04c071cc2067\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.038492 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-config-data\") pod \"903ed25a-6e5c-46c9-8aac-04c071cc2067\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.038717 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dh49k\" (UniqueName: \"kubernetes.io/projected/903ed25a-6e5c-46c9-8aac-04c071cc2067-kube-api-access-dh49k\") pod \"903ed25a-6e5c-46c9-8aac-04c071cc2067\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.038896 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-combined-ca-bundle\") pod \"903ed25a-6e5c-46c9-8aac-04c071cc2067\" (UID: \"903ed25a-6e5c-46c9-8aac-04c071cc2067\") " Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.043717 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/903ed25a-6e5c-46c9-8aac-04c071cc2067-kube-api-access-dh49k" (OuterVolumeSpecName: "kube-api-access-dh49k") pod "903ed25a-6e5c-46c9-8aac-04c071cc2067" (UID: "903ed25a-6e5c-46c9-8aac-04c071cc2067"). InnerVolumeSpecName "kube-api-access-dh49k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.051219 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-scripts" (OuterVolumeSpecName: "scripts") pod "903ed25a-6e5c-46c9-8aac-04c071cc2067" (UID: "903ed25a-6e5c-46c9-8aac-04c071cc2067"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.077333 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-config-data" (OuterVolumeSpecName: "config-data") pod "903ed25a-6e5c-46c9-8aac-04c071cc2067" (UID: "903ed25a-6e5c-46c9-8aac-04c071cc2067"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.078609 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "903ed25a-6e5c-46c9-8aac-04c071cc2067" (UID: "903ed25a-6e5c-46c9-8aac-04c071cc2067"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.141024 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.141057 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.141069 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/903ed25a-6e5c-46c9-8aac-04c071cc2067-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.141081 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dh49k\" (UniqueName: \"kubernetes.io/projected/903ed25a-6e5c-46c9-8aac-04c071cc2067-kube-api-access-dh49k\") on node \"crc\" DevicePath \"\"" Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.315360 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-j97tn" event={"ID":"903ed25a-6e5c-46c9-8aac-04c071cc2067","Type":"ContainerDied","Data":"5d685d8ac910546448bf01360ff61b57d5109c7f97b6824e5757e801d9e75d61"} Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.315418 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d685d8ac910546448bf01360ff61b57d5109c7f97b6824e5757e801d9e75d61" Oct 02 03:25:36 crc kubenswrapper[4775]: I1002 03:25:36.315456 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-j97tn" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.066042 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Oct 02 03:25:40 crc kubenswrapper[4775]: E1002 03:25:40.067072 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="903ed25a-6e5c-46c9-8aac-04c071cc2067" containerName="aodh-db-sync" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.067089 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="903ed25a-6e5c-46c9-8aac-04c071cc2067" containerName="aodh-db-sync" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.067336 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="903ed25a-6e5c-46c9-8aac-04c071cc2067" containerName="aodh-db-sync" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.069616 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.074874 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.075153 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-44ptd" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.075266 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.102245 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.246191 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nz9h\" (UniqueName: \"kubernetes.io/projected/7bb83c5f-2641-421a-ac8d-35e94398b856-kube-api-access-8nz9h\") pod \"aodh-0\" (UID: \"7bb83c5f-2641-421a-ac8d-35e94398b856\") " pod="openstack/aodh-0" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.246282 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb83c5f-2641-421a-ac8d-35e94398b856-config-data\") pod \"aodh-0\" (UID: \"7bb83c5f-2641-421a-ac8d-35e94398b856\") " pod="openstack/aodh-0" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.246408 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb83c5f-2641-421a-ac8d-35e94398b856-combined-ca-bundle\") pod \"aodh-0\" (UID: \"7bb83c5f-2641-421a-ac8d-35e94398b856\") " pod="openstack/aodh-0" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.246435 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bb83c5f-2641-421a-ac8d-35e94398b856-scripts\") pod \"aodh-0\" (UID: \"7bb83c5f-2641-421a-ac8d-35e94398b856\") " pod="openstack/aodh-0" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.348172 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb83c5f-2641-421a-ac8d-35e94398b856-combined-ca-bundle\") pod \"aodh-0\" (UID: \"7bb83c5f-2641-421a-ac8d-35e94398b856\") " pod="openstack/aodh-0" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.348213 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bb83c5f-2641-421a-ac8d-35e94398b856-scripts\") pod \"aodh-0\" (UID: \"7bb83c5f-2641-421a-ac8d-35e94398b856\") " pod="openstack/aodh-0" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.348298 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nz9h\" (UniqueName: \"kubernetes.io/projected/7bb83c5f-2641-421a-ac8d-35e94398b856-kube-api-access-8nz9h\") pod \"aodh-0\" (UID: \"7bb83c5f-2641-421a-ac8d-35e94398b856\") " pod="openstack/aodh-0" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.348341 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb83c5f-2641-421a-ac8d-35e94398b856-config-data\") pod \"aodh-0\" (UID: \"7bb83c5f-2641-421a-ac8d-35e94398b856\") " pod="openstack/aodh-0" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.354610 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bb83c5f-2641-421a-ac8d-35e94398b856-combined-ca-bundle\") pod \"aodh-0\" (UID: \"7bb83c5f-2641-421a-ac8d-35e94398b856\") " pod="openstack/aodh-0" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.355023 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bb83c5f-2641-421a-ac8d-35e94398b856-scripts\") pod \"aodh-0\" (UID: \"7bb83c5f-2641-421a-ac8d-35e94398b856\") " pod="openstack/aodh-0" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.355583 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bb83c5f-2641-421a-ac8d-35e94398b856-config-data\") pod \"aodh-0\" (UID: \"7bb83c5f-2641-421a-ac8d-35e94398b856\") " pod="openstack/aodh-0" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.378899 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nz9h\" (UniqueName: \"kubernetes.io/projected/7bb83c5f-2641-421a-ac8d-35e94398b856-kube-api-access-8nz9h\") pod \"aodh-0\" (UID: \"7bb83c5f-2641-421a-ac8d-35e94398b856\") " pod="openstack/aodh-0" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.393508 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Oct 02 03:25:40 crc kubenswrapper[4775]: I1002 03:25:40.908731 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Oct 02 03:25:41 crc kubenswrapper[4775]: I1002 03:25:41.417036 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7bb83c5f-2641-421a-ac8d-35e94398b856","Type":"ContainerStarted","Data":"a70ee6fd3fe6baf5b9728567e40c98f86baa36eb10dccc46b91d8b33436db3e5"} Oct 02 03:25:42 crc kubenswrapper[4775]: I1002 03:25:42.315566 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:25:42 crc kubenswrapper[4775]: I1002 03:25:42.316337 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="ceilometer-central-agent" containerID="cri-o://bfdef500eaece7f7830bc337caa87ec969ec4d7e93370e120fe226d7a0008671" gracePeriod=30 Oct 02 03:25:42 crc kubenswrapper[4775]: I1002 03:25:42.316378 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="proxy-httpd" containerID="cri-o://1b768b2912fc328d8ebe6baf56a330f9e136b8ca0216da0b519b91ac5ee113b4" gracePeriod=30 Oct 02 03:25:42 crc kubenswrapper[4775]: I1002 03:25:42.316458 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="ceilometer-notification-agent" containerID="cri-o://60ed7d313741ef68eb33311bc5902260e621602438bbc25f94495da1a6cff306" gracePeriod=30 Oct 02 03:25:42 crc kubenswrapper[4775]: I1002 03:25:42.316460 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="sg-core" containerID="cri-o://cde04bcfe6831402ac5d46a91be7dc401194428a7456aecf4f9f541e995a4683" gracePeriod=30 Oct 02 03:25:42 crc kubenswrapper[4775]: I1002 03:25:42.429441 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7bb83c5f-2641-421a-ac8d-35e94398b856","Type":"ContainerStarted","Data":"b4526256c3c4db5c053d4377193faa542ed825a4dccc8b433f0d6155e5078742"} Oct 02 03:25:43 crc kubenswrapper[4775]: I1002 03:25:43.446909 4775 generic.go:334] "Generic (PLEG): container finished" podID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerID="1b768b2912fc328d8ebe6baf56a330f9e136b8ca0216da0b519b91ac5ee113b4" exitCode=0 Oct 02 03:25:43 crc kubenswrapper[4775]: I1002 03:25:43.447651 4775 generic.go:334] "Generic (PLEG): container finished" podID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerID="cde04bcfe6831402ac5d46a91be7dc401194428a7456aecf4f9f541e995a4683" exitCode=2 Oct 02 03:25:43 crc kubenswrapper[4775]: I1002 03:25:43.447666 4775 generic.go:334] "Generic (PLEG): container finished" podID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerID="bfdef500eaece7f7830bc337caa87ec969ec4d7e93370e120fe226d7a0008671" exitCode=0 Oct 02 03:25:43 crc kubenswrapper[4775]: I1002 03:25:43.447001 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e77c3fe-d4ee-4b70-813e-087f3803a65b","Type":"ContainerDied","Data":"1b768b2912fc328d8ebe6baf56a330f9e136b8ca0216da0b519b91ac5ee113b4"} Oct 02 03:25:43 crc kubenswrapper[4775]: I1002 03:25:43.447758 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e77c3fe-d4ee-4b70-813e-087f3803a65b","Type":"ContainerDied","Data":"cde04bcfe6831402ac5d46a91be7dc401194428a7456aecf4f9f541e995a4683"} Oct 02 03:25:43 crc kubenswrapper[4775]: I1002 03:25:43.447785 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e77c3fe-d4ee-4b70-813e-087f3803a65b","Type":"ContainerDied","Data":"bfdef500eaece7f7830bc337caa87ec969ec4d7e93370e120fe226d7a0008671"} Oct 02 03:25:43 crc kubenswrapper[4775]: I1002 03:25:43.451364 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7bb83c5f-2641-421a-ac8d-35e94398b856","Type":"ContainerStarted","Data":"4c3f37ba02f127c86e816b95434bbe1605e084e2c7b0dc74dbbc49b1cfcab181"} Oct 02 03:25:45 crc kubenswrapper[4775]: I1002 03:25:45.478689 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7bb83c5f-2641-421a-ac8d-35e94398b856","Type":"ContainerStarted","Data":"ae1b68b8da0b0febcecd0c29a64c75c24e6ac45cbf333c38b82bfe2e08f36af8"} Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.514175 4775 generic.go:334] "Generic (PLEG): container finished" podID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerID="60ed7d313741ef68eb33311bc5902260e621602438bbc25f94495da1a6cff306" exitCode=0 Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.514272 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e77c3fe-d4ee-4b70-813e-087f3803a65b","Type":"ContainerDied","Data":"60ed7d313741ef68eb33311bc5902260e621602438bbc25f94495da1a6cff306"} Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.516929 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"7bb83c5f-2641-421a-ac8d-35e94398b856","Type":"ContainerStarted","Data":"d58042750fb04b4975c0775ec1791c45676fea4b73fc1199bbef7a26e3c65779"} Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.692130 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.712800 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=1.705493887 podStartE2EDuration="6.712777419s" podCreationTimestamp="2025-10-02 03:25:40 +0000 UTC" firstStartedPulling="2025-10-02 03:25:40.910751557 +0000 UTC m=+6278.077495597" lastFinishedPulling="2025-10-02 03:25:45.918035049 +0000 UTC m=+6283.084779129" observedRunningTime="2025-10-02 03:25:46.542945899 +0000 UTC m=+6283.709689969" watchObservedRunningTime="2025-10-02 03:25:46.712777419 +0000 UTC m=+6283.879521469" Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.827770 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-scripts\") pod \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.827920 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-config-data\") pod \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.828025 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-combined-ca-bundle\") pod \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.828046 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-sg-core-conf-yaml\") pod \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.828072 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnlmg\" (UniqueName: \"kubernetes.io/projected/2e77c3fe-d4ee-4b70-813e-087f3803a65b-kube-api-access-pnlmg\") pod \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.828127 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e77c3fe-d4ee-4b70-813e-087f3803a65b-run-httpd\") pod \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.828252 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e77c3fe-d4ee-4b70-813e-087f3803a65b-log-httpd\") pod \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\" (UID: \"2e77c3fe-d4ee-4b70-813e-087f3803a65b\") " Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.830937 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e77c3fe-d4ee-4b70-813e-087f3803a65b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2e77c3fe-d4ee-4b70-813e-087f3803a65b" (UID: "2e77c3fe-d4ee-4b70-813e-087f3803a65b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.831223 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e77c3fe-d4ee-4b70-813e-087f3803a65b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2e77c3fe-d4ee-4b70-813e-087f3803a65b" (UID: "2e77c3fe-d4ee-4b70-813e-087f3803a65b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.834687 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e77c3fe-d4ee-4b70-813e-087f3803a65b-kube-api-access-pnlmg" (OuterVolumeSpecName: "kube-api-access-pnlmg") pod "2e77c3fe-d4ee-4b70-813e-087f3803a65b" (UID: "2e77c3fe-d4ee-4b70-813e-087f3803a65b"). InnerVolumeSpecName "kube-api-access-pnlmg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.835473 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-scripts" (OuterVolumeSpecName: "scripts") pod "2e77c3fe-d4ee-4b70-813e-087f3803a65b" (UID: "2e77c3fe-d4ee-4b70-813e-087f3803a65b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.881587 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2e77c3fe-d4ee-4b70-813e-087f3803a65b" (UID: "2e77c3fe-d4ee-4b70-813e-087f3803a65b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.905347 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e77c3fe-d4ee-4b70-813e-087f3803a65b" (UID: "2e77c3fe-d4ee-4b70-813e-087f3803a65b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.944853 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e77c3fe-d4ee-4b70-813e-087f3803a65b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.944884 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e77c3fe-d4ee-4b70-813e-087f3803a65b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.944893 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.944901 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.944911 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.944919 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnlmg\" (UniqueName: \"kubernetes.io/projected/2e77c3fe-d4ee-4b70-813e-087f3803a65b-kube-api-access-pnlmg\") on node \"crc\" DevicePath \"\"" Oct 02 03:25:46 crc kubenswrapper[4775]: I1002 03:25:46.967288 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-config-data" (OuterVolumeSpecName: "config-data") pod "2e77c3fe-d4ee-4b70-813e-087f3803a65b" (UID: "2e77c3fe-d4ee-4b70-813e-087f3803a65b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.047312 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e77c3fe-d4ee-4b70-813e-087f3803a65b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.530591 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.530585 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e77c3fe-d4ee-4b70-813e-087f3803a65b","Type":"ContainerDied","Data":"7f2301bf621615c887e521dc87621a84cb172ad23838620fae2c318ac607743b"} Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.531061 4775 scope.go:117] "RemoveContainer" containerID="1b768b2912fc328d8ebe6baf56a330f9e136b8ca0216da0b519b91ac5ee113b4" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.564999 4775 scope.go:117] "RemoveContainer" containerID="cde04bcfe6831402ac5d46a91be7dc401194428a7456aecf4f9f541e995a4683" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.572439 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.602777 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.605332 4775 scope.go:117] "RemoveContainer" containerID="60ed7d313741ef68eb33311bc5902260e621602438bbc25f94495da1a6cff306" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.611463 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:25:47 crc kubenswrapper[4775]: E1002 03:25:47.612093 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="sg-core" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.612111 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="sg-core" Oct 02 03:25:47 crc kubenswrapper[4775]: E1002 03:25:47.612135 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="proxy-httpd" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.612143 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="proxy-httpd" Oct 02 03:25:47 crc kubenswrapper[4775]: E1002 03:25:47.612173 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="ceilometer-notification-agent" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.612181 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="ceilometer-notification-agent" Oct 02 03:25:47 crc kubenswrapper[4775]: E1002 03:25:47.612196 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="ceilometer-central-agent" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.612204 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="ceilometer-central-agent" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.612485 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="sg-core" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.612503 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="proxy-httpd" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.612528 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="ceilometer-central-agent" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.612537 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" containerName="ceilometer-notification-agent" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.615073 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.618574 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.631356 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.631718 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.666128 4775 scope.go:117] "RemoveContainer" containerID="bfdef500eaece7f7830bc337caa87ec969ec4d7e93370e120fe226d7a0008671" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.779428 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e77c3fe-d4ee-4b70-813e-087f3803a65b" path="/var/lib/kubelet/pods/2e77c3fe-d4ee-4b70-813e-087f3803a65b/volumes" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.787540 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgqzj\" (UniqueName: \"kubernetes.io/projected/3788020a-9d4d-4abd-8752-33df16542a43-kube-api-access-bgqzj\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.787590 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-config-data\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.787629 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-scripts\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.787670 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.787743 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3788020a-9d4d-4abd-8752-33df16542a43-log-httpd\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.787829 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3788020a-9d4d-4abd-8752-33df16542a43-run-httpd\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.787859 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.891414 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgqzj\" (UniqueName: \"kubernetes.io/projected/3788020a-9d4d-4abd-8752-33df16542a43-kube-api-access-bgqzj\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.891762 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-config-data\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.892579 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-scripts\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.892644 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.892730 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3788020a-9d4d-4abd-8752-33df16542a43-log-httpd\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.892840 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3788020a-9d4d-4abd-8752-33df16542a43-run-httpd\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.892881 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.895141 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3788020a-9d4d-4abd-8752-33df16542a43-log-httpd\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.901318 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3788020a-9d4d-4abd-8752-33df16542a43-run-httpd\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.901738 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-config-data\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.901943 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.903691 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-scripts\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.907689 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.908505 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgqzj\" (UniqueName: \"kubernetes.io/projected/3788020a-9d4d-4abd-8752-33df16542a43-kube-api-access-bgqzj\") pod \"ceilometer-0\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " pod="openstack/ceilometer-0" Oct 02 03:25:47 crc kubenswrapper[4775]: I1002 03:25:47.946200 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 03:25:48 crc kubenswrapper[4775]: I1002 03:25:48.570050 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:25:48 crc kubenswrapper[4775]: W1002 03:25:48.580813 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3788020a_9d4d_4abd_8752_33df16542a43.slice/crio-2a3d00738a89b38d77d27d1be7e2e78f33ad8fb2315e7bb5925f69fe95521bb2 WatchSource:0}: Error finding container 2a3d00738a89b38d77d27d1be7e2e78f33ad8fb2315e7bb5925f69fe95521bb2: Status 404 returned error can't find the container with id 2a3d00738a89b38d77d27d1be7e2e78f33ad8fb2315e7bb5925f69fe95521bb2 Oct 02 03:25:49 crc kubenswrapper[4775]: I1002 03:25:49.554918 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3788020a-9d4d-4abd-8752-33df16542a43","Type":"ContainerStarted","Data":"c2ec351b591ea86e68ce0416c75a9ded5c704d3d92a039b0d9822914929d0f35"} Oct 02 03:25:49 crc kubenswrapper[4775]: I1002 03:25:49.555583 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3788020a-9d4d-4abd-8752-33df16542a43","Type":"ContainerStarted","Data":"2a3d00738a89b38d77d27d1be7e2e78f33ad8fb2315e7bb5925f69fe95521bb2"} Oct 02 03:25:50 crc kubenswrapper[4775]: I1002 03:25:50.566131 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3788020a-9d4d-4abd-8752-33df16542a43","Type":"ContainerStarted","Data":"ae6c7109b74cbd862d807864adc0a904b82b8c818e01605e3d1ee73d2c502ebe"} Oct 02 03:25:51 crc kubenswrapper[4775]: I1002 03:25:51.589247 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3788020a-9d4d-4abd-8752-33df16542a43","Type":"ContainerStarted","Data":"39016a334edcad3d93505de3d6893d73b2f2a4823b34e336905dcb42a0710450"} Oct 02 03:25:51 crc kubenswrapper[4775]: I1002 03:25:51.813472 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-fjmh2"] Oct 02 03:25:51 crc kubenswrapper[4775]: I1002 03:25:51.814651 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-fjmh2" Oct 02 03:25:51 crc kubenswrapper[4775]: I1002 03:25:51.827471 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-fjmh2"] Oct 02 03:25:51 crc kubenswrapper[4775]: I1002 03:25:51.990563 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpm7q\" (UniqueName: \"kubernetes.io/projected/5e7ec537-ac4e-4f99-b9d0-e5758f145445-kube-api-access-cpm7q\") pod \"manila-db-create-fjmh2\" (UID: \"5e7ec537-ac4e-4f99-b9d0-e5758f145445\") " pod="openstack/manila-db-create-fjmh2" Oct 02 03:25:52 crc kubenswrapper[4775]: I1002 03:25:52.094429 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpm7q\" (UniqueName: \"kubernetes.io/projected/5e7ec537-ac4e-4f99-b9d0-e5758f145445-kube-api-access-cpm7q\") pod \"manila-db-create-fjmh2\" (UID: \"5e7ec537-ac4e-4f99-b9d0-e5758f145445\") " pod="openstack/manila-db-create-fjmh2" Oct 02 03:25:52 crc kubenswrapper[4775]: I1002 03:25:52.115819 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpm7q\" (UniqueName: \"kubernetes.io/projected/5e7ec537-ac4e-4f99-b9d0-e5758f145445-kube-api-access-cpm7q\") pod \"manila-db-create-fjmh2\" (UID: \"5e7ec537-ac4e-4f99-b9d0-e5758f145445\") " pod="openstack/manila-db-create-fjmh2" Oct 02 03:25:52 crc kubenswrapper[4775]: I1002 03:25:52.131735 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-fjmh2" Oct 02 03:25:52 crc kubenswrapper[4775]: I1002 03:25:52.713415 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-fjmh2"] Oct 02 03:25:53 crc kubenswrapper[4775]: I1002 03:25:53.621885 4775 generic.go:334] "Generic (PLEG): container finished" podID="5e7ec537-ac4e-4f99-b9d0-e5758f145445" containerID="e4cd09bc36b03426ae6bbcca6fc8d9574b4c0345d2367a3b930032ff24f4c3b8" exitCode=0 Oct 02 03:25:53 crc kubenswrapper[4775]: I1002 03:25:53.622350 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-fjmh2" event={"ID":"5e7ec537-ac4e-4f99-b9d0-e5758f145445","Type":"ContainerDied","Data":"e4cd09bc36b03426ae6bbcca6fc8d9574b4c0345d2367a3b930032ff24f4c3b8"} Oct 02 03:25:53 crc kubenswrapper[4775]: I1002 03:25:53.622377 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-fjmh2" event={"ID":"5e7ec537-ac4e-4f99-b9d0-e5758f145445","Type":"ContainerStarted","Data":"ccdcfa9ccdbfa4df1d9cf680cb8bb28ccbea4d5e45b0712f4865b0e33ea327fc"} Oct 02 03:25:53 crc kubenswrapper[4775]: I1002 03:25:53.626081 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3788020a-9d4d-4abd-8752-33df16542a43","Type":"ContainerStarted","Data":"cb37ad9fd671f088d7dd365c1700814f6901137af232efe560e1ef4b92ba3a58"} Oct 02 03:25:53 crc kubenswrapper[4775]: I1002 03:25:53.626393 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 03:25:53 crc kubenswrapper[4775]: I1002 03:25:53.677183 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.741042098 podStartE2EDuration="6.677156166s" podCreationTimestamp="2025-10-02 03:25:47 +0000 UTC" firstStartedPulling="2025-10-02 03:25:48.583039325 +0000 UTC m=+6285.749783365" lastFinishedPulling="2025-10-02 03:25:52.519153403 +0000 UTC m=+6289.685897433" observedRunningTime="2025-10-02 03:25:53.663850851 +0000 UTC m=+6290.830594931" watchObservedRunningTime="2025-10-02 03:25:53.677156166 +0000 UTC m=+6290.843900236" Oct 02 03:25:55 crc kubenswrapper[4775]: I1002 03:25:55.190788 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-fjmh2" Oct 02 03:25:55 crc kubenswrapper[4775]: I1002 03:25:55.293765 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpm7q\" (UniqueName: \"kubernetes.io/projected/5e7ec537-ac4e-4f99-b9d0-e5758f145445-kube-api-access-cpm7q\") pod \"5e7ec537-ac4e-4f99-b9d0-e5758f145445\" (UID: \"5e7ec537-ac4e-4f99-b9d0-e5758f145445\") " Oct 02 03:25:55 crc kubenswrapper[4775]: I1002 03:25:55.303283 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e7ec537-ac4e-4f99-b9d0-e5758f145445-kube-api-access-cpm7q" (OuterVolumeSpecName: "kube-api-access-cpm7q") pod "5e7ec537-ac4e-4f99-b9d0-e5758f145445" (UID: "5e7ec537-ac4e-4f99-b9d0-e5758f145445"). InnerVolumeSpecName "kube-api-access-cpm7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:25:55 crc kubenswrapper[4775]: I1002 03:25:55.397652 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpm7q\" (UniqueName: \"kubernetes.io/projected/5e7ec537-ac4e-4f99-b9d0-e5758f145445-kube-api-access-cpm7q\") on node \"crc\" DevicePath \"\"" Oct 02 03:25:55 crc kubenswrapper[4775]: I1002 03:25:55.661272 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-fjmh2" event={"ID":"5e7ec537-ac4e-4f99-b9d0-e5758f145445","Type":"ContainerDied","Data":"ccdcfa9ccdbfa4df1d9cf680cb8bb28ccbea4d5e45b0712f4865b0e33ea327fc"} Oct 02 03:25:55 crc kubenswrapper[4775]: I1002 03:25:55.661318 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccdcfa9ccdbfa4df1d9cf680cb8bb28ccbea4d5e45b0712f4865b0e33ea327fc" Oct 02 03:25:55 crc kubenswrapper[4775]: I1002 03:25:55.661333 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-fjmh2" Oct 02 03:26:01 crc kubenswrapper[4775]: I1002 03:26:01.975477 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-1375-account-create-hln6z"] Oct 02 03:26:01 crc kubenswrapper[4775]: E1002 03:26:01.976516 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7ec537-ac4e-4f99-b9d0-e5758f145445" containerName="mariadb-database-create" Oct 02 03:26:01 crc kubenswrapper[4775]: I1002 03:26:01.976532 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7ec537-ac4e-4f99-b9d0-e5758f145445" containerName="mariadb-database-create" Oct 02 03:26:01 crc kubenswrapper[4775]: I1002 03:26:01.976762 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e7ec537-ac4e-4f99-b9d0-e5758f145445" containerName="mariadb-database-create" Oct 02 03:26:01 crc kubenswrapper[4775]: I1002 03:26:01.977705 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-1375-account-create-hln6z" Oct 02 03:26:01 crc kubenswrapper[4775]: I1002 03:26:01.980778 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 02 03:26:01 crc kubenswrapper[4775]: I1002 03:26:01.992054 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-1375-account-create-hln6z"] Oct 02 03:26:02 crc kubenswrapper[4775]: I1002 03:26:02.090946 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkb2w\" (UniqueName: \"kubernetes.io/projected/e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf-kube-api-access-hkb2w\") pod \"manila-1375-account-create-hln6z\" (UID: \"e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf\") " pod="openstack/manila-1375-account-create-hln6z" Oct 02 03:26:02 crc kubenswrapper[4775]: I1002 03:26:02.193801 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkb2w\" (UniqueName: \"kubernetes.io/projected/e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf-kube-api-access-hkb2w\") pod \"manila-1375-account-create-hln6z\" (UID: \"e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf\") " pod="openstack/manila-1375-account-create-hln6z" Oct 02 03:26:02 crc kubenswrapper[4775]: I1002 03:26:02.224234 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkb2w\" (UniqueName: \"kubernetes.io/projected/e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf-kube-api-access-hkb2w\") pod \"manila-1375-account-create-hln6z\" (UID: \"e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf\") " pod="openstack/manila-1375-account-create-hln6z" Oct 02 03:26:02 crc kubenswrapper[4775]: I1002 03:26:02.323521 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-1375-account-create-hln6z" Oct 02 03:26:02 crc kubenswrapper[4775]: I1002 03:26:02.842680 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-1375-account-create-hln6z"] Oct 02 03:26:03 crc kubenswrapper[4775]: I1002 03:26:03.783879 4775 generic.go:334] "Generic (PLEG): container finished" podID="e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf" containerID="903bbca96494b81e50dbce334a55e19d2a4174f186fd33c486b7b166908a930f" exitCode=0 Oct 02 03:26:03 crc kubenswrapper[4775]: I1002 03:26:03.789243 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-1375-account-create-hln6z" event={"ID":"e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf","Type":"ContainerDied","Data":"903bbca96494b81e50dbce334a55e19d2a4174f186fd33c486b7b166908a930f"} Oct 02 03:26:03 crc kubenswrapper[4775]: I1002 03:26:03.789285 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-1375-account-create-hln6z" event={"ID":"e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf","Type":"ContainerStarted","Data":"5f754af99d7914e345a27f64ef5e32b9d1c40d7eb1cf867bee5ee3b3aae8fcd9"} Oct 02 03:26:05 crc kubenswrapper[4775]: I1002 03:26:05.260359 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-1375-account-create-hln6z" Oct 02 03:26:05 crc kubenswrapper[4775]: I1002 03:26:05.395786 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkb2w\" (UniqueName: \"kubernetes.io/projected/e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf-kube-api-access-hkb2w\") pod \"e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf\" (UID: \"e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf\") " Oct 02 03:26:05 crc kubenswrapper[4775]: I1002 03:26:05.405870 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf-kube-api-access-hkb2w" (OuterVolumeSpecName: "kube-api-access-hkb2w") pod "e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf" (UID: "e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf"). InnerVolumeSpecName "kube-api-access-hkb2w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:26:05 crc kubenswrapper[4775]: I1002 03:26:05.499391 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkb2w\" (UniqueName: \"kubernetes.io/projected/e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf-kube-api-access-hkb2w\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:05 crc kubenswrapper[4775]: I1002 03:26:05.815675 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-1375-account-create-hln6z" event={"ID":"e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf","Type":"ContainerDied","Data":"5f754af99d7914e345a27f64ef5e32b9d1c40d7eb1cf867bee5ee3b3aae8fcd9"} Oct 02 03:26:05 crc kubenswrapper[4775]: I1002 03:26:05.816028 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f754af99d7914e345a27f64ef5e32b9d1c40d7eb1cf867bee5ee3b3aae8fcd9" Oct 02 03:26:05 crc kubenswrapper[4775]: I1002 03:26:05.815743 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-1375-account-create-hln6z" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.246658 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.247004 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.329026 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-xlbkk"] Oct 02 03:26:07 crc kubenswrapper[4775]: E1002 03:26:07.329632 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf" containerName="mariadb-account-create" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.329658 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf" containerName="mariadb-account-create" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.329939 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf" containerName="mariadb-account-create" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.330940 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.333009 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-98rqf" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.333406 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.347219 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-xlbkk"] Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.446443 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj2ld\" (UniqueName: \"kubernetes.io/projected/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-kube-api-access-tj2ld\") pod \"manila-db-sync-xlbkk\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.446585 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-job-config-data\") pod \"manila-db-sync-xlbkk\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.446619 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-config-data\") pod \"manila-db-sync-xlbkk\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.446680 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-combined-ca-bundle\") pod \"manila-db-sync-xlbkk\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.548908 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj2ld\" (UniqueName: \"kubernetes.io/projected/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-kube-api-access-tj2ld\") pod \"manila-db-sync-xlbkk\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.549028 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-job-config-data\") pod \"manila-db-sync-xlbkk\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.549068 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-config-data\") pod \"manila-db-sync-xlbkk\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.549136 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-combined-ca-bundle\") pod \"manila-db-sync-xlbkk\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.554305 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-config-data\") pod \"manila-db-sync-xlbkk\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.556428 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-combined-ca-bundle\") pod \"manila-db-sync-xlbkk\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.563811 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-job-config-data\") pod \"manila-db-sync-xlbkk\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.563978 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj2ld\" (UniqueName: \"kubernetes.io/projected/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-kube-api-access-tj2ld\") pod \"manila-db-sync-xlbkk\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:07 crc kubenswrapper[4775]: I1002 03:26:07.650502 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:08 crc kubenswrapper[4775]: W1002 03:26:08.510650 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0987ddf_aa6c_4f61_9fcd_42370ad9a697.slice/crio-879b9c3ae604eb1b5b0abf4f36515d3d0edc349d53a9617c402b0922f59c987e WatchSource:0}: Error finding container 879b9c3ae604eb1b5b0abf4f36515d3d0edc349d53a9617c402b0922f59c987e: Status 404 returned error can't find the container with id 879b9c3ae604eb1b5b0abf4f36515d3d0edc349d53a9617c402b0922f59c987e Oct 02 03:26:08 crc kubenswrapper[4775]: I1002 03:26:08.533231 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-xlbkk"] Oct 02 03:26:08 crc kubenswrapper[4775]: I1002 03:26:08.852058 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-xlbkk" event={"ID":"b0987ddf-aa6c-4f61-9fcd-42370ad9a697","Type":"ContainerStarted","Data":"879b9c3ae604eb1b5b0abf4f36515d3d0edc349d53a9617c402b0922f59c987e"} Oct 02 03:26:14 crc kubenswrapper[4775]: I1002 03:26:14.485640 4775 scope.go:117] "RemoveContainer" containerID="13f0c894d70e585132b7ce1ad0445019d51afe10558d20a630b589d1b45956be" Oct 02 03:26:14 crc kubenswrapper[4775]: I1002 03:26:14.511582 4775 scope.go:117] "RemoveContainer" containerID="b9c22cb65f85af7868611432cb1a21b0ca9aeaef6bf00ce5287f0651041d5055" Oct 02 03:26:14 crc kubenswrapper[4775]: I1002 03:26:14.589611 4775 scope.go:117] "RemoveContainer" containerID="4b52b5567f3c3250afa5f90b477c6ddb3b3c6791df1fc3a566eb950d56971c4e" Oct 02 03:26:14 crc kubenswrapper[4775]: I1002 03:26:14.926529 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-xlbkk" event={"ID":"b0987ddf-aa6c-4f61-9fcd-42370ad9a697","Type":"ContainerStarted","Data":"3e49c5e9b5524d8baf15da9a9f6ef660fb6a545a205679265737adee43c43f76"} Oct 02 03:26:15 crc kubenswrapper[4775]: I1002 03:26:15.956493 4775 generic.go:334] "Generic (PLEG): container finished" podID="b0987ddf-aa6c-4f61-9fcd-42370ad9a697" containerID="3e49c5e9b5524d8baf15da9a9f6ef660fb6a545a205679265737adee43c43f76" exitCode=0 Oct 02 03:26:15 crc kubenswrapper[4775]: I1002 03:26:15.956830 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-xlbkk" event={"ID":"b0987ddf-aa6c-4f61-9fcd-42370ad9a697","Type":"ContainerDied","Data":"3e49c5e9b5524d8baf15da9a9f6ef660fb6a545a205679265737adee43c43f76"} Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.573352 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.604165 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-job-config-data\") pod \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.604231 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tj2ld\" (UniqueName: \"kubernetes.io/projected/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-kube-api-access-tj2ld\") pod \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.604427 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-combined-ca-bundle\") pod \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.604650 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-config-data\") pod \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\" (UID: \"b0987ddf-aa6c-4f61-9fcd-42370ad9a697\") " Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.613289 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-kube-api-access-tj2ld" (OuterVolumeSpecName: "kube-api-access-tj2ld") pod "b0987ddf-aa6c-4f61-9fcd-42370ad9a697" (UID: "b0987ddf-aa6c-4f61-9fcd-42370ad9a697"). InnerVolumeSpecName "kube-api-access-tj2ld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.614725 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-config-data" (OuterVolumeSpecName: "config-data") pod "b0987ddf-aa6c-4f61-9fcd-42370ad9a697" (UID: "b0987ddf-aa6c-4f61-9fcd-42370ad9a697"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.615582 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "b0987ddf-aa6c-4f61-9fcd-42370ad9a697" (UID: "b0987ddf-aa6c-4f61-9fcd-42370ad9a697"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.658623 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b0987ddf-aa6c-4f61-9fcd-42370ad9a697" (UID: "b0987ddf-aa6c-4f61-9fcd-42370ad9a697"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.706787 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.706820 4775 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.706831 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tj2ld\" (UniqueName: \"kubernetes.io/projected/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-kube-api-access-tj2ld\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.706841 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0987ddf-aa6c-4f61-9fcd-42370ad9a697-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.981717 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-xlbkk" event={"ID":"b0987ddf-aa6c-4f61-9fcd-42370ad9a697","Type":"ContainerDied","Data":"879b9c3ae604eb1b5b0abf4f36515d3d0edc349d53a9617c402b0922f59c987e"} Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.981757 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="879b9c3ae604eb1b5b0abf4f36515d3d0edc349d53a9617c402b0922f59c987e" Oct 02 03:26:17 crc kubenswrapper[4775]: I1002 03:26:17.981778 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-xlbkk" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.046525 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.257501 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 02 03:26:18 crc kubenswrapper[4775]: E1002 03:26:18.259649 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0987ddf-aa6c-4f61-9fcd-42370ad9a697" containerName="manila-db-sync" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.259676 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0987ddf-aa6c-4f61-9fcd-42370ad9a697" containerName="manila-db-sync" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.259987 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0987ddf-aa6c-4f61-9fcd-42370ad9a697" containerName="manila-db-sync" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.261373 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.264104 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-98rqf" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.264106 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.265563 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.269504 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.279123 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.330437 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bmzr\" (UniqueName: \"kubernetes.io/projected/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-kube-api-access-2bmzr\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.330517 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.330551 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.330575 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.330664 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-scripts\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.330697 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-config-data\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.345871 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.349874 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.357272 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.364912 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.426807 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-867c9c4db9-qs7q5"] Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.428586 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.431175 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-ovsdbserver-sb\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.431205 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.431226 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-ovsdbserver-nb\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.431249 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.431272 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.431300 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68pjf\" (UniqueName: \"kubernetes.io/projected/35407ca5-0ace-441b-8d6f-e01db17cfa26-kube-api-access-68pjf\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.431337 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dn6b\" (UniqueName: \"kubernetes.io/projected/6355bb29-85fc-474d-8370-af3339ff22c8-kube-api-access-2dn6b\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.431363 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6355bb29-85fc-474d-8370-af3339ff22c8-scripts\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.431384 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-dns-svc\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.431400 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-config\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.431432 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-scripts\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.431468 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6355bb29-85fc-474d-8370-af3339ff22c8-config-data\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.431485 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/6355bb29-85fc-474d-8370-af3339ff22c8-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.431559 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.432214 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6355bb29-85fc-474d-8370-af3339ff22c8-ceph\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.432337 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-config-data\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.432442 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bmzr\" (UniqueName: \"kubernetes.io/projected/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-kube-api-access-2bmzr\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.432470 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6355bb29-85fc-474d-8370-af3339ff22c8-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.432488 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6355bb29-85fc-474d-8370-af3339ff22c8-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.432562 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6355bb29-85fc-474d-8370-af3339ff22c8-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.435964 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-config-data\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.436082 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-scripts\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.439030 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-867c9c4db9-qs7q5"] Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.444514 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.459187 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bmzr\" (UniqueName: \"kubernetes.io/projected/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-kube-api-access-2bmzr\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.473066 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2faaa667-8d44-47aa-ab2f-27b1d49e6d70-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"2faaa667-8d44-47aa-ab2f-27b1d49e6d70\") " pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.535970 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68pjf\" (UniqueName: \"kubernetes.io/projected/35407ca5-0ace-441b-8d6f-e01db17cfa26-kube-api-access-68pjf\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.536031 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dn6b\" (UniqueName: \"kubernetes.io/projected/6355bb29-85fc-474d-8370-af3339ff22c8-kube-api-access-2dn6b\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.536061 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6355bb29-85fc-474d-8370-af3339ff22c8-scripts\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.536082 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-dns-svc\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.536100 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-config\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.536137 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6355bb29-85fc-474d-8370-af3339ff22c8-config-data\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.536153 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/6355bb29-85fc-474d-8370-af3339ff22c8-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.536173 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6355bb29-85fc-474d-8370-af3339ff22c8-ceph\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.537137 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-dns-svc\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.538292 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6355bb29-85fc-474d-8370-af3339ff22c8-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.538340 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6355bb29-85fc-474d-8370-af3339ff22c8-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.538427 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6355bb29-85fc-474d-8370-af3339ff22c8-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.538489 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-ovsdbserver-sb\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.538526 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-ovsdbserver-nb\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.539607 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-ovsdbserver-nb\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.541357 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6355bb29-85fc-474d-8370-af3339ff22c8-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.541480 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/6355bb29-85fc-474d-8370-af3339ff22c8-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.541900 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-config\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.545900 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-ovsdbserver-sb\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.545932 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6355bb29-85fc-474d-8370-af3339ff22c8-config-data\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.548481 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6355bb29-85fc-474d-8370-af3339ff22c8-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.548631 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6355bb29-85fc-474d-8370-af3339ff22c8-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.555334 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6355bb29-85fc-474d-8370-af3339ff22c8-scripts\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.555439 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6355bb29-85fc-474d-8370-af3339ff22c8-ceph\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.572186 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dn6b\" (UniqueName: \"kubernetes.io/projected/6355bb29-85fc-474d-8370-af3339ff22c8-kube-api-access-2dn6b\") pod \"manila-share-share1-0\" (UID: \"6355bb29-85fc-474d-8370-af3339ff22c8\") " pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.573683 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68pjf\" (UniqueName: \"kubernetes.io/projected/35407ca5-0ace-441b-8d6f-e01db17cfa26-kube-api-access-68pjf\") pod \"dnsmasq-dns-867c9c4db9-qs7q5\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.581838 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.619272 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.621152 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.623276 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.637213 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.646854 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e902e4c2-791d-489a-b2b7-f2e2ed851356-etc-machine-id\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.646899 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82nkd\" (UniqueName: \"kubernetes.io/projected/e902e4c2-791d-489a-b2b7-f2e2ed851356-kube-api-access-82nkd\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.646948 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e902e4c2-791d-489a-b2b7-f2e2ed851356-config-data\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.647016 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e902e4c2-791d-489a-b2b7-f2e2ed851356-config-data-custom\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.647103 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e902e4c2-791d-489a-b2b7-f2e2ed851356-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.647282 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e902e4c2-791d-489a-b2b7-f2e2ed851356-scripts\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.647332 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e902e4c2-791d-489a-b2b7-f2e2ed851356-logs\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.688694 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.736080 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.748768 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e902e4c2-791d-489a-b2b7-f2e2ed851356-scripts\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.748819 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e902e4c2-791d-489a-b2b7-f2e2ed851356-logs\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.748925 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e902e4c2-791d-489a-b2b7-f2e2ed851356-etc-machine-id\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.748947 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82nkd\" (UniqueName: \"kubernetes.io/projected/e902e4c2-791d-489a-b2b7-f2e2ed851356-kube-api-access-82nkd\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.749008 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e902e4c2-791d-489a-b2b7-f2e2ed851356-config-data\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.749058 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e902e4c2-791d-489a-b2b7-f2e2ed851356-config-data-custom\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.749078 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e902e4c2-791d-489a-b2b7-f2e2ed851356-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.753460 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e902e4c2-791d-489a-b2b7-f2e2ed851356-etc-machine-id\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.754932 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e902e4c2-791d-489a-b2b7-f2e2ed851356-logs\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.757601 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e902e4c2-791d-489a-b2b7-f2e2ed851356-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.757808 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e902e4c2-791d-489a-b2b7-f2e2ed851356-config-data\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.763583 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e902e4c2-791d-489a-b2b7-f2e2ed851356-scripts\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.787536 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82nkd\" (UniqueName: \"kubernetes.io/projected/e902e4c2-791d-489a-b2b7-f2e2ed851356-kube-api-access-82nkd\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.789594 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e902e4c2-791d-489a-b2b7-f2e2ed851356-config-data-custom\") pod \"manila-api-0\" (UID: \"e902e4c2-791d-489a-b2b7-f2e2ed851356\") " pod="openstack/manila-api-0" Oct 02 03:26:18 crc kubenswrapper[4775]: I1002 03:26:18.805880 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 02 03:26:19 crc kubenswrapper[4775]: I1002 03:26:19.266454 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 02 03:26:19 crc kubenswrapper[4775]: I1002 03:26:19.275425 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 03:26:19 crc kubenswrapper[4775]: I1002 03:26:19.810632 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 02 03:26:19 crc kubenswrapper[4775]: I1002 03:26:19.870540 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-867c9c4db9-qs7q5"] Oct 02 03:26:19 crc kubenswrapper[4775]: I1002 03:26:19.949827 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 02 03:26:20 crc kubenswrapper[4775]: I1002 03:26:20.046463 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"2faaa667-8d44-47aa-ab2f-27b1d49e6d70","Type":"ContainerStarted","Data":"9c4ad6d544e2ef07ed393a9802810592abc1049d7637b103735b1199f322fb6f"} Oct 02 03:26:20 crc kubenswrapper[4775]: I1002 03:26:20.047655 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"e902e4c2-791d-489a-b2b7-f2e2ed851356","Type":"ContainerStarted","Data":"20c2db6521554f5c3a2f059b8747abf58104afa4db203c0d9afa8b03be6ca33f"} Oct 02 03:26:20 crc kubenswrapper[4775]: I1002 03:26:20.052802 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" event={"ID":"35407ca5-0ace-441b-8d6f-e01db17cfa26","Type":"ContainerStarted","Data":"0e404a3aeaa17dd6288c794962a8ba5495fbe3ddd157ed5443cd779806e14d86"} Oct 02 03:26:21 crc kubenswrapper[4775]: I1002 03:26:21.070519 4775 generic.go:334] "Generic (PLEG): container finished" podID="35407ca5-0ace-441b-8d6f-e01db17cfa26" containerID="8c0c9d999b3388fdb88cb503479899dcf800ed33c8e7e29850a38208bf2fc888" exitCode=0 Oct 02 03:26:21 crc kubenswrapper[4775]: I1002 03:26:21.070704 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" event={"ID":"35407ca5-0ace-441b-8d6f-e01db17cfa26","Type":"ContainerDied","Data":"8c0c9d999b3388fdb88cb503479899dcf800ed33c8e7e29850a38208bf2fc888"} Oct 02 03:26:21 crc kubenswrapper[4775]: I1002 03:26:21.073210 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"6355bb29-85fc-474d-8370-af3339ff22c8","Type":"ContainerStarted","Data":"4e15a9569e846aa73fcf2824268811b0867f63aa78396b16d36fe6d250ee8ece"} Oct 02 03:26:21 crc kubenswrapper[4775]: I1002 03:26:21.075448 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"2faaa667-8d44-47aa-ab2f-27b1d49e6d70","Type":"ContainerStarted","Data":"31feb2482be985a7563ff63f085b820679f8c0b45619c85ae2e91deda7a023d3"} Oct 02 03:26:21 crc kubenswrapper[4775]: I1002 03:26:21.089003 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"e902e4c2-791d-489a-b2b7-f2e2ed851356","Type":"ContainerStarted","Data":"5bc5ca8c6193b147a313aa587c3c763211b03a4aae5b7135d665e491f374fc8c"} Oct 02 03:26:21 crc kubenswrapper[4775]: I1002 03:26:21.089042 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"e902e4c2-791d-489a-b2b7-f2e2ed851356","Type":"ContainerStarted","Data":"e130382ea8df93620ee339cec37fadc2dc10d1297a3dcbb1daddcebd0f3c52de"} Oct 02 03:26:21 crc kubenswrapper[4775]: I1002 03:26:21.089403 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 02 03:26:21 crc kubenswrapper[4775]: I1002 03:26:21.110079 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.110059891 podStartE2EDuration="3.110059891s" podCreationTimestamp="2025-10-02 03:26:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:26:21.108745427 +0000 UTC m=+6318.275489467" watchObservedRunningTime="2025-10-02 03:26:21.110059891 +0000 UTC m=+6318.276803931" Oct 02 03:26:22 crc kubenswrapper[4775]: I1002 03:26:22.103584 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" event={"ID":"35407ca5-0ace-441b-8d6f-e01db17cfa26","Type":"ContainerStarted","Data":"a2451c31506939f6408df0cf03f544f3c2df48ec00792ef47aebcdd8d3b26c03"} Oct 02 03:26:22 crc kubenswrapper[4775]: I1002 03:26:22.103870 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:22 crc kubenswrapper[4775]: I1002 03:26:22.107166 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"2faaa667-8d44-47aa-ab2f-27b1d49e6d70","Type":"ContainerStarted","Data":"34f973ac329c31a6a40e32d2d0a6cfd06979c3ca4dadbfc2652173e032088f3b"} Oct 02 03:26:22 crc kubenswrapper[4775]: I1002 03:26:22.131337 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" podStartSLOduration=4.131312661 podStartE2EDuration="4.131312661s" podCreationTimestamp="2025-10-02 03:26:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:26:22.125082269 +0000 UTC m=+6319.291826319" watchObservedRunningTime="2025-10-02 03:26:22.131312661 +0000 UTC m=+6319.298056701" Oct 02 03:26:22 crc kubenswrapper[4775]: I1002 03:26:22.156608 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.228758928 podStartE2EDuration="4.156583606s" podCreationTimestamp="2025-10-02 03:26:18 +0000 UTC" firstStartedPulling="2025-10-02 03:26:19.275161052 +0000 UTC m=+6316.441905092" lastFinishedPulling="2025-10-02 03:26:20.20298573 +0000 UTC m=+6317.369729770" observedRunningTime="2025-10-02 03:26:22.146520035 +0000 UTC m=+6319.313264075" watchObservedRunningTime="2025-10-02 03:26:22.156583606 +0000 UTC m=+6319.323327656" Oct 02 03:26:27 crc kubenswrapper[4775]: I1002 03:26:27.174903 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"6355bb29-85fc-474d-8370-af3339ff22c8","Type":"ContainerStarted","Data":"122f7b5a1fd1cc375f8c55daff7065f6f57ee70056287e388983fbf270e3f015"} Oct 02 03:26:27 crc kubenswrapper[4775]: I1002 03:26:27.175604 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"6355bb29-85fc-474d-8370-af3339ff22c8","Type":"ContainerStarted","Data":"d7fd033324356b50bd4275a5658721099eec3a7769df0dc0bc24afbd2052fd32"} Oct 02 03:26:27 crc kubenswrapper[4775]: I1002 03:26:27.203032 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.527028194 podStartE2EDuration="9.20301698s" podCreationTimestamp="2025-10-02 03:26:18 +0000 UTC" firstStartedPulling="2025-10-02 03:26:20.186017141 +0000 UTC m=+6317.352761181" lastFinishedPulling="2025-10-02 03:26:25.862005897 +0000 UTC m=+6323.028749967" observedRunningTime="2025-10-02 03:26:27.199015187 +0000 UTC m=+6324.365759247" watchObservedRunningTime="2025-10-02 03:26:27.20301698 +0000 UTC m=+6324.369761020" Oct 02 03:26:28 crc kubenswrapper[4775]: I1002 03:26:28.583021 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 02 03:26:28 crc kubenswrapper[4775]: I1002 03:26:28.690234 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 02 03:26:28 crc kubenswrapper[4775]: I1002 03:26:28.738399 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:26:28 crc kubenswrapper[4775]: I1002 03:26:28.826263 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bd57d768c-b47z2"] Oct 02 03:26:28 crc kubenswrapper[4775]: I1002 03:26:28.826629 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" podUID="5cbdb932-01a7-409c-bc31-4dceb903ccad" containerName="dnsmasq-dns" containerID="cri-o://0ca58f904e02e5c88e9e6959d4f51da3eb893f3e04ecbd6afcc108c76aa71c10" gracePeriod=10 Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.208249 4775 generic.go:334] "Generic (PLEG): container finished" podID="5cbdb932-01a7-409c-bc31-4dceb903ccad" containerID="0ca58f904e02e5c88e9e6959d4f51da3eb893f3e04ecbd6afcc108c76aa71c10" exitCode=0 Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.209275 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" event={"ID":"5cbdb932-01a7-409c-bc31-4dceb903ccad","Type":"ContainerDied","Data":"0ca58f904e02e5c88e9e6959d4f51da3eb893f3e04ecbd6afcc108c76aa71c10"} Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.410569 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.528060 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-config\") pod \"5cbdb932-01a7-409c-bc31-4dceb903ccad\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.528309 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-ovsdbserver-sb\") pod \"5cbdb932-01a7-409c-bc31-4dceb903ccad\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.528340 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-ovsdbserver-nb\") pod \"5cbdb932-01a7-409c-bc31-4dceb903ccad\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.528444 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b482g\" (UniqueName: \"kubernetes.io/projected/5cbdb932-01a7-409c-bc31-4dceb903ccad-kube-api-access-b482g\") pod \"5cbdb932-01a7-409c-bc31-4dceb903ccad\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.528485 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-dns-svc\") pod \"5cbdb932-01a7-409c-bc31-4dceb903ccad\" (UID: \"5cbdb932-01a7-409c-bc31-4dceb903ccad\") " Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.535223 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cbdb932-01a7-409c-bc31-4dceb903ccad-kube-api-access-b482g" (OuterVolumeSpecName: "kube-api-access-b482g") pod "5cbdb932-01a7-409c-bc31-4dceb903ccad" (UID: "5cbdb932-01a7-409c-bc31-4dceb903ccad"). InnerVolumeSpecName "kube-api-access-b482g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.584488 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5cbdb932-01a7-409c-bc31-4dceb903ccad" (UID: "5cbdb932-01a7-409c-bc31-4dceb903ccad"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.589555 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5cbdb932-01a7-409c-bc31-4dceb903ccad" (UID: "5cbdb932-01a7-409c-bc31-4dceb903ccad"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.591887 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5cbdb932-01a7-409c-bc31-4dceb903ccad" (UID: "5cbdb932-01a7-409c-bc31-4dceb903ccad"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.599683 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-config" (OuterVolumeSpecName: "config") pod "5cbdb932-01a7-409c-bc31-4dceb903ccad" (UID: "5cbdb932-01a7-409c-bc31-4dceb903ccad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.630995 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b482g\" (UniqueName: \"kubernetes.io/projected/5cbdb932-01a7-409c-bc31-4dceb903ccad-kube-api-access-b482g\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.631030 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.631040 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.631050 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:29 crc kubenswrapper[4775]: I1002 03:26:29.631058 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5cbdb932-01a7-409c-bc31-4dceb903ccad-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:30 crc kubenswrapper[4775]: I1002 03:26:30.221466 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" event={"ID":"5cbdb932-01a7-409c-bc31-4dceb903ccad","Type":"ContainerDied","Data":"51729983878fe4d5fb3aeb93797a9a0e42a49ff8fce583d8be4667f55125ba91"} Oct 02 03:26:30 crc kubenswrapper[4775]: I1002 03:26:30.221516 4775 scope.go:117] "RemoveContainer" containerID="0ca58f904e02e5c88e9e6959d4f51da3eb893f3e04ecbd6afcc108c76aa71c10" Oct 02 03:26:30 crc kubenswrapper[4775]: I1002 03:26:30.221585 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bd57d768c-b47z2" Oct 02 03:26:30 crc kubenswrapper[4775]: I1002 03:26:30.266240 4775 scope.go:117] "RemoveContainer" containerID="7a6dd5f8679e03c842934fac8237531a771fe43ac712c7b34032696f3841d01a" Oct 02 03:26:30 crc kubenswrapper[4775]: I1002 03:26:30.271135 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bd57d768c-b47z2"] Oct 02 03:26:30 crc kubenswrapper[4775]: I1002 03:26:30.304022 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bd57d768c-b47z2"] Oct 02 03:26:31 crc kubenswrapper[4775]: I1002 03:26:31.781981 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cbdb932-01a7-409c-bc31-4dceb903ccad" path="/var/lib/kubelet/pods/5cbdb932-01a7-409c-bc31-4dceb903ccad/volumes" Oct 02 03:26:32 crc kubenswrapper[4775]: I1002 03:26:32.124753 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:26:32 crc kubenswrapper[4775]: I1002 03:26:32.125325 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="ceilometer-central-agent" containerID="cri-o://c2ec351b591ea86e68ce0416c75a9ded5c704d3d92a039b0d9822914929d0f35" gracePeriod=30 Oct 02 03:26:32 crc kubenswrapper[4775]: I1002 03:26:32.125465 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="sg-core" containerID="cri-o://39016a334edcad3d93505de3d6893d73b2f2a4823b34e336905dcb42a0710450" gracePeriod=30 Oct 02 03:26:32 crc kubenswrapper[4775]: I1002 03:26:32.125457 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="ceilometer-notification-agent" containerID="cri-o://ae6c7109b74cbd862d807864adc0a904b82b8c818e01605e3d1ee73d2c502ebe" gracePeriod=30 Oct 02 03:26:32 crc kubenswrapper[4775]: I1002 03:26:32.125500 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="proxy-httpd" containerID="cri-o://cb37ad9fd671f088d7dd365c1700814f6901137af232efe560e1ef4b92ba3a58" gracePeriod=30 Oct 02 03:26:33 crc kubenswrapper[4775]: I1002 03:26:33.252440 4775 generic.go:334] "Generic (PLEG): container finished" podID="3788020a-9d4d-4abd-8752-33df16542a43" containerID="cb37ad9fd671f088d7dd365c1700814f6901137af232efe560e1ef4b92ba3a58" exitCode=0 Oct 02 03:26:33 crc kubenswrapper[4775]: I1002 03:26:33.252772 4775 generic.go:334] "Generic (PLEG): container finished" podID="3788020a-9d4d-4abd-8752-33df16542a43" containerID="39016a334edcad3d93505de3d6893d73b2f2a4823b34e336905dcb42a0710450" exitCode=2 Oct 02 03:26:33 crc kubenswrapper[4775]: I1002 03:26:33.252782 4775 generic.go:334] "Generic (PLEG): container finished" podID="3788020a-9d4d-4abd-8752-33df16542a43" containerID="c2ec351b591ea86e68ce0416c75a9ded5c704d3d92a039b0d9822914929d0f35" exitCode=0 Oct 02 03:26:33 crc kubenswrapper[4775]: I1002 03:26:33.252512 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3788020a-9d4d-4abd-8752-33df16542a43","Type":"ContainerDied","Data":"cb37ad9fd671f088d7dd365c1700814f6901137af232efe560e1ef4b92ba3a58"} Oct 02 03:26:33 crc kubenswrapper[4775]: I1002 03:26:33.252819 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3788020a-9d4d-4abd-8752-33df16542a43","Type":"ContainerDied","Data":"39016a334edcad3d93505de3d6893d73b2f2a4823b34e336905dcb42a0710450"} Oct 02 03:26:33 crc kubenswrapper[4775]: I1002 03:26:33.252834 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3788020a-9d4d-4abd-8752-33df16542a43","Type":"ContainerDied","Data":"c2ec351b591ea86e68ce0416c75a9ded5c704d3d92a039b0d9822914929d0f35"} Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.233275 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.234064 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.323770 4775 generic.go:334] "Generic (PLEG): container finished" podID="3788020a-9d4d-4abd-8752-33df16542a43" containerID="ae6c7109b74cbd862d807864adc0a904b82b8c818e01605e3d1ee73d2c502ebe" exitCode=0 Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.323810 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3788020a-9d4d-4abd-8752-33df16542a43","Type":"ContainerDied","Data":"ae6c7109b74cbd862d807864adc0a904b82b8c818e01605e3d1ee73d2c502ebe"} Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.659039 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.829603 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-combined-ca-bundle\") pod \"3788020a-9d4d-4abd-8752-33df16542a43\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.829724 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-config-data\") pod \"3788020a-9d4d-4abd-8752-33df16542a43\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.829755 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-sg-core-conf-yaml\") pod \"3788020a-9d4d-4abd-8752-33df16542a43\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.829833 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3788020a-9d4d-4abd-8752-33df16542a43-run-httpd\") pod \"3788020a-9d4d-4abd-8752-33df16542a43\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.829909 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3788020a-9d4d-4abd-8752-33df16542a43-log-httpd\") pod \"3788020a-9d4d-4abd-8752-33df16542a43\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.829927 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-scripts\") pod \"3788020a-9d4d-4abd-8752-33df16542a43\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.830041 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgqzj\" (UniqueName: \"kubernetes.io/projected/3788020a-9d4d-4abd-8752-33df16542a43-kube-api-access-bgqzj\") pod \"3788020a-9d4d-4abd-8752-33df16542a43\" (UID: \"3788020a-9d4d-4abd-8752-33df16542a43\") " Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.830446 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3788020a-9d4d-4abd-8752-33df16542a43-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3788020a-9d4d-4abd-8752-33df16542a43" (UID: "3788020a-9d4d-4abd-8752-33df16542a43"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.830757 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3788020a-9d4d-4abd-8752-33df16542a43-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3788020a-9d4d-4abd-8752-33df16542a43" (UID: "3788020a-9d4d-4abd-8752-33df16542a43"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.831261 4775 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3788020a-9d4d-4abd-8752-33df16542a43-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.831281 4775 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3788020a-9d4d-4abd-8752-33df16542a43-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.836014 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-scripts" (OuterVolumeSpecName: "scripts") pod "3788020a-9d4d-4abd-8752-33df16542a43" (UID: "3788020a-9d4d-4abd-8752-33df16542a43"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.840186 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3788020a-9d4d-4abd-8752-33df16542a43-kube-api-access-bgqzj" (OuterVolumeSpecName: "kube-api-access-bgqzj") pod "3788020a-9d4d-4abd-8752-33df16542a43" (UID: "3788020a-9d4d-4abd-8752-33df16542a43"). InnerVolumeSpecName "kube-api-access-bgqzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.866143 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3788020a-9d4d-4abd-8752-33df16542a43" (UID: "3788020a-9d4d-4abd-8752-33df16542a43"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.933536 4775 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.933839 4775 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-scripts\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.933854 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgqzj\" (UniqueName: \"kubernetes.io/projected/3788020a-9d4d-4abd-8752-33df16542a43-kube-api-access-bgqzj\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.938315 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3788020a-9d4d-4abd-8752-33df16542a43" (UID: "3788020a-9d4d-4abd-8752-33df16542a43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:26:37 crc kubenswrapper[4775]: I1002 03:26:37.978540 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-config-data" (OuterVolumeSpecName: "config-data") pod "3788020a-9d4d-4abd-8752-33df16542a43" (UID: "3788020a-9d4d-4abd-8752-33df16542a43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.036799 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.036830 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3788020a-9d4d-4abd-8752-33df16542a43-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.340064 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3788020a-9d4d-4abd-8752-33df16542a43","Type":"ContainerDied","Data":"2a3d00738a89b38d77d27d1be7e2e78f33ad8fb2315e7bb5925f69fe95521bb2"} Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.340137 4775 scope.go:117] "RemoveContainer" containerID="cb37ad9fd671f088d7dd365c1700814f6901137af232efe560e1ef4b92ba3a58" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.340336 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.389313 4775 scope.go:117] "RemoveContainer" containerID="39016a334edcad3d93505de3d6893d73b2f2a4823b34e336905dcb42a0710450" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.400147 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.413161 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.430707 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:26:38 crc kubenswrapper[4775]: E1002 03:26:38.431279 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="ceilometer-notification-agent" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.431302 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="ceilometer-notification-agent" Oct 02 03:26:38 crc kubenswrapper[4775]: E1002 03:26:38.431324 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="proxy-httpd" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.431337 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="proxy-httpd" Oct 02 03:26:38 crc kubenswrapper[4775]: E1002 03:26:38.431365 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="ceilometer-central-agent" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.431373 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="ceilometer-central-agent" Oct 02 03:26:38 crc kubenswrapper[4775]: E1002 03:26:38.431391 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbdb932-01a7-409c-bc31-4dceb903ccad" containerName="dnsmasq-dns" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.431400 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbdb932-01a7-409c-bc31-4dceb903ccad" containerName="dnsmasq-dns" Oct 02 03:26:38 crc kubenswrapper[4775]: E1002 03:26:38.431433 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="sg-core" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.431441 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="sg-core" Oct 02 03:26:38 crc kubenswrapper[4775]: E1002 03:26:38.431467 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cbdb932-01a7-409c-bc31-4dceb903ccad" containerName="init" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.431475 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cbdb932-01a7-409c-bc31-4dceb903ccad" containerName="init" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.431719 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cbdb932-01a7-409c-bc31-4dceb903ccad" containerName="dnsmasq-dns" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.431736 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="ceilometer-notification-agent" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.431759 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="ceilometer-central-agent" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.431777 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="sg-core" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.431801 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3788020a-9d4d-4abd-8752-33df16542a43" containerName="proxy-httpd" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.434319 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.437638 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.437791 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.445597 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.447418 4775 scope.go:117] "RemoveContainer" containerID="ae6c7109b74cbd862d807864adc0a904b82b8c818e01605e3d1ee73d2c502ebe" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.484319 4775 scope.go:117] "RemoveContainer" containerID="c2ec351b591ea86e68ce0416c75a9ded5c704d3d92a039b0d9822914929d0f35" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.548680 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbwdb\" (UniqueName: \"kubernetes.io/projected/c7172345-ba1a-4856-8b88-009379970b34-kube-api-access-nbwdb\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.548740 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7172345-ba1a-4856-8b88-009379970b34-run-httpd\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.548767 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7172345-ba1a-4856-8b88-009379970b34-scripts\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.548886 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7172345-ba1a-4856-8b88-009379970b34-config-data\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.548946 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7172345-ba1a-4856-8b88-009379970b34-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.549016 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7172345-ba1a-4856-8b88-009379970b34-log-httpd\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.549053 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7172345-ba1a-4856-8b88-009379970b34-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.650754 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7172345-ba1a-4856-8b88-009379970b34-log-httpd\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.650862 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7172345-ba1a-4856-8b88-009379970b34-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.651033 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbwdb\" (UniqueName: \"kubernetes.io/projected/c7172345-ba1a-4856-8b88-009379970b34-kube-api-access-nbwdb\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.651080 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7172345-ba1a-4856-8b88-009379970b34-scripts\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.651109 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7172345-ba1a-4856-8b88-009379970b34-run-httpd\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.651215 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7172345-ba1a-4856-8b88-009379970b34-config-data\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.651799 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7172345-ba1a-4856-8b88-009379970b34-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.651862 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7172345-ba1a-4856-8b88-009379970b34-run-httpd\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.652119 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c7172345-ba1a-4856-8b88-009379970b34-log-httpd\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.657300 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c7172345-ba1a-4856-8b88-009379970b34-scripts\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.657333 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c7172345-ba1a-4856-8b88-009379970b34-config-data\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.658189 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c7172345-ba1a-4856-8b88-009379970b34-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.659830 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c7172345-ba1a-4856-8b88-009379970b34-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.669818 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbwdb\" (UniqueName: \"kubernetes.io/projected/c7172345-ba1a-4856-8b88-009379970b34-kube-api-access-nbwdb\") pod \"ceilometer-0\" (UID: \"c7172345-ba1a-4856-8b88-009379970b34\") " pod="openstack/ceilometer-0" Oct 02 03:26:38 crc kubenswrapper[4775]: I1002 03:26:38.764508 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 02 03:26:39 crc kubenswrapper[4775]: I1002 03:26:39.331894 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 02 03:26:39 crc kubenswrapper[4775]: W1002 03:26:39.336661 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7172345_ba1a_4856_8b88_009379970b34.slice/crio-797ca6da3c87227080fbd4e55489c1f2ed535a12c5de13efa3a245a4d8a0583f WatchSource:0}: Error finding container 797ca6da3c87227080fbd4e55489c1f2ed535a12c5de13efa3a245a4d8a0583f: Status 404 returned error can't find the container with id 797ca6da3c87227080fbd4e55489c1f2ed535a12c5de13efa3a245a4d8a0583f Oct 02 03:26:39 crc kubenswrapper[4775]: I1002 03:26:39.350290 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7172345-ba1a-4856-8b88-009379970b34","Type":"ContainerStarted","Data":"797ca6da3c87227080fbd4e55489c1f2ed535a12c5de13efa3a245a4d8a0583f"} Oct 02 03:26:39 crc kubenswrapper[4775]: I1002 03:26:39.777070 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3788020a-9d4d-4abd-8752-33df16542a43" path="/var/lib/kubelet/pods/3788020a-9d4d-4abd-8752-33df16542a43/volumes" Oct 02 03:26:40 crc kubenswrapper[4775]: I1002 03:26:40.036001 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 02 03:26:40 crc kubenswrapper[4775]: I1002 03:26:40.173292 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 02 03:26:40 crc kubenswrapper[4775]: I1002 03:26:40.212573 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 02 03:26:41 crc kubenswrapper[4775]: I1002 03:26:41.373508 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7172345-ba1a-4856-8b88-009379970b34","Type":"ContainerStarted","Data":"84aaa8ee4802cf4f509e94b0390f2f07fbb89c89018e9f406127d1917b65e802"} Oct 02 03:26:42 crc kubenswrapper[4775]: I1002 03:26:42.383456 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7172345-ba1a-4856-8b88-009379970b34","Type":"ContainerStarted","Data":"c33c8deee76675dfda8e40655f7c3cd67aa98d1f1d17669a5aa48dd89592509e"} Oct 02 03:26:43 crc kubenswrapper[4775]: I1002 03:26:43.394872 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7172345-ba1a-4856-8b88-009379970b34","Type":"ContainerStarted","Data":"b4384b652dd59205bbd37024fd066f2800d02d001c7a511eb713b115b6dfb884"} Oct 02 03:26:45 crc kubenswrapper[4775]: I1002 03:26:45.423124 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c7172345-ba1a-4856-8b88-009379970b34","Type":"ContainerStarted","Data":"1411c581d6a9da0415adbc8b18c345a8a2fc51a37873589dcdb8212fd425db7d"} Oct 02 03:26:45 crc kubenswrapper[4775]: I1002 03:26:45.424943 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 02 03:26:45 crc kubenswrapper[4775]: I1002 03:26:45.471892 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.972318104 podStartE2EDuration="7.471865678s" podCreationTimestamp="2025-10-02 03:26:38 +0000 UTC" firstStartedPulling="2025-10-02 03:26:39.339094898 +0000 UTC m=+6336.505838938" lastFinishedPulling="2025-10-02 03:26:44.838642432 +0000 UTC m=+6342.005386512" observedRunningTime="2025-10-02 03:26:45.449534729 +0000 UTC m=+6342.616278789" watchObservedRunningTime="2025-10-02 03:26:45.471865678 +0000 UTC m=+6342.638609758" Oct 02 03:27:07 crc kubenswrapper[4775]: I1002 03:27:07.233343 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:27:07 crc kubenswrapper[4775]: I1002 03:27:07.234038 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:27:07 crc kubenswrapper[4775]: I1002 03:27:07.234092 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 03:27:07 crc kubenswrapper[4775]: I1002 03:27:07.235052 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 03:27:07 crc kubenswrapper[4775]: I1002 03:27:07.235167 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" gracePeriod=600 Oct 02 03:27:07 crc kubenswrapper[4775]: E1002 03:27:07.368488 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:27:07 crc kubenswrapper[4775]: I1002 03:27:07.731645 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769"} Oct 02 03:27:07 crc kubenswrapper[4775]: I1002 03:27:07.731721 4775 scope.go:117] "RemoveContainer" containerID="c4930148cbbbfb59b99fb94d29b145a85b1cfbffe3a70db63ae0fd48271fb062" Oct 02 03:27:07 crc kubenswrapper[4775]: I1002 03:27:07.732846 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:27:07 crc kubenswrapper[4775]: E1002 03:27:07.733416 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:27:07 crc kubenswrapper[4775]: I1002 03:27:07.731594 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" exitCode=0 Oct 02 03:27:08 crc kubenswrapper[4775]: I1002 03:27:08.771232 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 02 03:27:19 crc kubenswrapper[4775]: I1002 03:27:19.766371 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:27:19 crc kubenswrapper[4775]: E1002 03:27:19.767423 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.334241 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d6c869-m2kg7"] Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.340873 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.343012 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.355112 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d6c869-m2kg7"] Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.459594 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.460123 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.460313 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x68z\" (UniqueName: \"kubernetes.io/projected/86bb6a99-2caa-478e-b73f-7c6e9b24808a-kube-api-access-9x68z\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.460361 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-dns-svc\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.460622 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-openstack-cell1\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.460667 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-config\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.562246 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-openstack-cell1\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.562291 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-config\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.562379 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.562416 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.562492 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x68z\" (UniqueName: \"kubernetes.io/projected/86bb6a99-2caa-478e-b73f-7c6e9b24808a-kube-api-access-9x68z\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.562512 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-dns-svc\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.563589 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.563619 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-dns-svc\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.563833 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-openstack-cell1\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.565402 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.565998 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-config\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.583355 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x68z\" (UniqueName: \"kubernetes.io/projected/86bb6a99-2caa-478e-b73f-7c6e9b24808a-kube-api-access-9x68z\") pod \"dnsmasq-dns-5c79d6c869-m2kg7\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:29 crc kubenswrapper[4775]: I1002 03:27:29.668861 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:30 crc kubenswrapper[4775]: I1002 03:27:30.177356 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d6c869-m2kg7"] Oct 02 03:27:30 crc kubenswrapper[4775]: I1002 03:27:30.765484 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:27:30 crc kubenswrapper[4775]: E1002 03:27:30.766311 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:27:31 crc kubenswrapper[4775]: I1002 03:27:31.080905 4775 generic.go:334] "Generic (PLEG): container finished" podID="86bb6a99-2caa-478e-b73f-7c6e9b24808a" containerID="95aeb45a504e567784b7144ed8cc19a83837748dc8e65efbb3ac700260b7133a" exitCode=0 Oct 02 03:27:31 crc kubenswrapper[4775]: I1002 03:27:31.081012 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" event={"ID":"86bb6a99-2caa-478e-b73f-7c6e9b24808a","Type":"ContainerDied","Data":"95aeb45a504e567784b7144ed8cc19a83837748dc8e65efbb3ac700260b7133a"} Oct 02 03:27:31 crc kubenswrapper[4775]: I1002 03:27:31.081085 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" event={"ID":"86bb6a99-2caa-478e-b73f-7c6e9b24808a","Type":"ContainerStarted","Data":"e60bf3d19a61de1ae31dab8790406cfedfe9d50d14c735d12eee42779276bf94"} Oct 02 03:27:32 crc kubenswrapper[4775]: I1002 03:27:32.095164 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" event={"ID":"86bb6a99-2caa-478e-b73f-7c6e9b24808a","Type":"ContainerStarted","Data":"cb59be74b5f644a82a2f7f2898b6681b2938fc9c818cea6e2f9fa2e0d3233de9"} Oct 02 03:27:32 crc kubenswrapper[4775]: I1002 03:27:32.097629 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:32 crc kubenswrapper[4775]: I1002 03:27:32.145737 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" podStartSLOduration=3.145716179 podStartE2EDuration="3.145716179s" podCreationTimestamp="2025-10-02 03:27:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:27:32.125522146 +0000 UTC m=+6389.292266226" watchObservedRunningTime="2025-10-02 03:27:32.145716179 +0000 UTC m=+6389.312460229" Oct 02 03:27:39 crc kubenswrapper[4775]: I1002 03:27:39.672208 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:39 crc kubenswrapper[4775]: I1002 03:27:39.753718 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-867c9c4db9-qs7q5"] Oct 02 03:27:39 crc kubenswrapper[4775]: I1002 03:27:39.755364 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" podUID="35407ca5-0ace-441b-8d6f-e01db17cfa26" containerName="dnsmasq-dns" containerID="cri-o://a2451c31506939f6408df0cf03f544f3c2df48ec00792ef47aebcdd8d3b26c03" gracePeriod=10 Oct 02 03:27:39 crc kubenswrapper[4775]: I1002 03:27:39.887019 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54d96fb68f-shvgf"] Oct 02 03:27:39 crc kubenswrapper[4775]: I1002 03:27:39.889208 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:39 crc kubenswrapper[4775]: I1002 03:27:39.895643 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54d96fb68f-shvgf"] Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.059232 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-ovsdbserver-sb\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.059322 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-config\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.059341 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-openstack-cell1\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.059371 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-dns-svc\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.059399 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-ovsdbserver-nb\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.059497 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nnsc\" (UniqueName: \"kubernetes.io/projected/4808d138-0235-4e3e-97fe-8362aa73d26f-kube-api-access-2nnsc\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.161367 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nnsc\" (UniqueName: \"kubernetes.io/projected/4808d138-0235-4e3e-97fe-8362aa73d26f-kube-api-access-2nnsc\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.161697 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-ovsdbserver-sb\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.161785 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-config\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.161884 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-openstack-cell1\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.162001 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-dns-svc\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.162112 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-ovsdbserver-nb\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.162896 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-openstack-cell1\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.162909 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-dns-svc\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.163639 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-ovsdbserver-nb\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.166756 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-ovsdbserver-sb\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.172477 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4808d138-0235-4e3e-97fe-8362aa73d26f-config\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.196940 4775 generic.go:334] "Generic (PLEG): container finished" podID="35407ca5-0ace-441b-8d6f-e01db17cfa26" containerID="a2451c31506939f6408df0cf03f544f3c2df48ec00792ef47aebcdd8d3b26c03" exitCode=0 Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.196996 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" event={"ID":"35407ca5-0ace-441b-8d6f-e01db17cfa26","Type":"ContainerDied","Data":"a2451c31506939f6408df0cf03f544f3c2df48ec00792ef47aebcdd8d3b26c03"} Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.204748 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nnsc\" (UniqueName: \"kubernetes.io/projected/4808d138-0235-4e3e-97fe-8362aa73d26f-kube-api-access-2nnsc\") pod \"dnsmasq-dns-54d96fb68f-shvgf\" (UID: \"4808d138-0235-4e3e-97fe-8362aa73d26f\") " pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.239839 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.422304 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.472771 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-ovsdbserver-sb\") pod \"35407ca5-0ace-441b-8d6f-e01db17cfa26\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.472848 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-ovsdbserver-nb\") pod \"35407ca5-0ace-441b-8d6f-e01db17cfa26\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.473035 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-dns-svc\") pod \"35407ca5-0ace-441b-8d6f-e01db17cfa26\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.473094 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68pjf\" (UniqueName: \"kubernetes.io/projected/35407ca5-0ace-441b-8d6f-e01db17cfa26-kube-api-access-68pjf\") pod \"35407ca5-0ace-441b-8d6f-e01db17cfa26\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.473188 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-config\") pod \"35407ca5-0ace-441b-8d6f-e01db17cfa26\" (UID: \"35407ca5-0ace-441b-8d6f-e01db17cfa26\") " Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.493237 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35407ca5-0ace-441b-8d6f-e01db17cfa26-kube-api-access-68pjf" (OuterVolumeSpecName: "kube-api-access-68pjf") pod "35407ca5-0ace-441b-8d6f-e01db17cfa26" (UID: "35407ca5-0ace-441b-8d6f-e01db17cfa26"). InnerVolumeSpecName "kube-api-access-68pjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.536565 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "35407ca5-0ace-441b-8d6f-e01db17cfa26" (UID: "35407ca5-0ace-441b-8d6f-e01db17cfa26"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.543787 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-config" (OuterVolumeSpecName: "config") pod "35407ca5-0ace-441b-8d6f-e01db17cfa26" (UID: "35407ca5-0ace-441b-8d6f-e01db17cfa26"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.556379 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "35407ca5-0ace-441b-8d6f-e01db17cfa26" (UID: "35407ca5-0ace-441b-8d6f-e01db17cfa26"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.558406 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "35407ca5-0ace-441b-8d6f-e01db17cfa26" (UID: "35407ca5-0ace-441b-8d6f-e01db17cfa26"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.574921 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.574967 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.574978 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68pjf\" (UniqueName: \"kubernetes.io/projected/35407ca5-0ace-441b-8d6f-e01db17cfa26-kube-api-access-68pjf\") on node \"crc\" DevicePath \"\"" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.574988 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.574996 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35407ca5-0ace-441b-8d6f-e01db17cfa26-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 03:27:40 crc kubenswrapper[4775]: I1002 03:27:40.731208 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54d96fb68f-shvgf"] Oct 02 03:27:41 crc kubenswrapper[4775]: I1002 03:27:41.216716 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" Oct 02 03:27:41 crc kubenswrapper[4775]: I1002 03:27:41.216625 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-867c9c4db9-qs7q5" event={"ID":"35407ca5-0ace-441b-8d6f-e01db17cfa26","Type":"ContainerDied","Data":"0e404a3aeaa17dd6288c794962a8ba5495fbe3ddd157ed5443cd779806e14d86"} Oct 02 03:27:41 crc kubenswrapper[4775]: I1002 03:27:41.217365 4775 scope.go:117] "RemoveContainer" containerID="a2451c31506939f6408df0cf03f544f3c2df48ec00792ef47aebcdd8d3b26c03" Oct 02 03:27:41 crc kubenswrapper[4775]: I1002 03:27:41.218844 4775 generic.go:334] "Generic (PLEG): container finished" podID="4808d138-0235-4e3e-97fe-8362aa73d26f" containerID="dedac25e7def4ea3bc75074fa10d62f86ba8f235386359649265f895d4b07573" exitCode=0 Oct 02 03:27:41 crc kubenswrapper[4775]: I1002 03:27:41.218923 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" event={"ID":"4808d138-0235-4e3e-97fe-8362aa73d26f","Type":"ContainerDied","Data":"dedac25e7def4ea3bc75074fa10d62f86ba8f235386359649265f895d4b07573"} Oct 02 03:27:41 crc kubenswrapper[4775]: I1002 03:27:41.218945 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" event={"ID":"4808d138-0235-4e3e-97fe-8362aa73d26f","Type":"ContainerStarted","Data":"8ff2497a0ad25aa7c3b4d482edc4a8b6349d29c7d67f61351447372378b05f20"} Oct 02 03:27:41 crc kubenswrapper[4775]: I1002 03:27:41.247663 4775 scope.go:117] "RemoveContainer" containerID="8c0c9d999b3388fdb88cb503479899dcf800ed33c8e7e29850a38208bf2fc888" Oct 02 03:27:41 crc kubenswrapper[4775]: I1002 03:27:41.471648 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-867c9c4db9-qs7q5"] Oct 02 03:27:41 crc kubenswrapper[4775]: I1002 03:27:41.479980 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-867c9c4db9-qs7q5"] Oct 02 03:27:41 crc kubenswrapper[4775]: I1002 03:27:41.766049 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:27:41 crc kubenswrapper[4775]: E1002 03:27:41.766629 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:27:41 crc kubenswrapper[4775]: I1002 03:27:41.778088 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35407ca5-0ace-441b-8d6f-e01db17cfa26" path="/var/lib/kubelet/pods/35407ca5-0ace-441b-8d6f-e01db17cfa26/volumes" Oct 02 03:27:42 crc kubenswrapper[4775]: I1002 03:27:42.237503 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" event={"ID":"4808d138-0235-4e3e-97fe-8362aa73d26f","Type":"ContainerStarted","Data":"4aada528845ef757f0975a28ba89c9c8a685a6e9c11d8623a75d893b7b777fe1"} Oct 02 03:27:42 crc kubenswrapper[4775]: I1002 03:27:42.238710 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:42 crc kubenswrapper[4775]: I1002 03:27:42.275528 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" podStartSLOduration=3.275499566 podStartE2EDuration="3.275499566s" podCreationTimestamp="2025-10-02 03:27:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 03:27:42.267992321 +0000 UTC m=+6399.434736421" watchObservedRunningTime="2025-10-02 03:27:42.275499566 +0000 UTC m=+6399.442243656" Oct 02 03:27:50 crc kubenswrapper[4775]: I1002 03:27:50.243194 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54d96fb68f-shvgf" Oct 02 03:27:50 crc kubenswrapper[4775]: I1002 03:27:50.337886 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d6c869-m2kg7"] Oct 02 03:27:50 crc kubenswrapper[4775]: I1002 03:27:50.338828 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" podUID="86bb6a99-2caa-478e-b73f-7c6e9b24808a" containerName="dnsmasq-dns" containerID="cri-o://cb59be74b5f644a82a2f7f2898b6681b2938fc9c818cea6e2f9fa2e0d3233de9" gracePeriod=10 Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.048344 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.068410 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-dns-svc\") pod \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.068896 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-openstack-cell1\") pod \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.069019 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9x68z\" (UniqueName: \"kubernetes.io/projected/86bb6a99-2caa-478e-b73f-7c6e9b24808a-kube-api-access-9x68z\") pod \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.069114 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-config\") pod \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.069861 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-ovsdbserver-sb\") pod \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.070042 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-ovsdbserver-nb\") pod \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\" (UID: \"86bb6a99-2caa-478e-b73f-7c6e9b24808a\") " Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.082604 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86bb6a99-2caa-478e-b73f-7c6e9b24808a-kube-api-access-9x68z" (OuterVolumeSpecName: "kube-api-access-9x68z") pod "86bb6a99-2caa-478e-b73f-7c6e9b24808a" (UID: "86bb6a99-2caa-478e-b73f-7c6e9b24808a"). InnerVolumeSpecName "kube-api-access-9x68z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.156697 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-config" (OuterVolumeSpecName: "config") pod "86bb6a99-2caa-478e-b73f-7c6e9b24808a" (UID: "86bb6a99-2caa-478e-b73f-7c6e9b24808a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.161142 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "86bb6a99-2caa-478e-b73f-7c6e9b24808a" (UID: "86bb6a99-2caa-478e-b73f-7c6e9b24808a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.164826 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "86bb6a99-2caa-478e-b73f-7c6e9b24808a" (UID: "86bb6a99-2caa-478e-b73f-7c6e9b24808a"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.174008 4775 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.174044 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9x68z\" (UniqueName: \"kubernetes.io/projected/86bb6a99-2caa-478e-b73f-7c6e9b24808a-kube-api-access-9x68z\") on node \"crc\" DevicePath \"\"" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.174061 4775 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-config\") on node \"crc\" DevicePath \"\"" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.174096 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.174348 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "86bb6a99-2caa-478e-b73f-7c6e9b24808a" (UID: "86bb6a99-2caa-478e-b73f-7c6e9b24808a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.174961 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "86bb6a99-2caa-478e-b73f-7c6e9b24808a" (UID: "86bb6a99-2caa-478e-b73f-7c6e9b24808a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.275640 4775 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.275674 4775 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86bb6a99-2caa-478e-b73f-7c6e9b24808a-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.374556 4775 generic.go:334] "Generic (PLEG): container finished" podID="86bb6a99-2caa-478e-b73f-7c6e9b24808a" containerID="cb59be74b5f644a82a2f7f2898b6681b2938fc9c818cea6e2f9fa2e0d3233de9" exitCode=0 Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.374612 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" event={"ID":"86bb6a99-2caa-478e-b73f-7c6e9b24808a","Type":"ContainerDied","Data":"cb59be74b5f644a82a2f7f2898b6681b2938fc9c818cea6e2f9fa2e0d3233de9"} Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.374621 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.374644 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d6c869-m2kg7" event={"ID":"86bb6a99-2caa-478e-b73f-7c6e9b24808a","Type":"ContainerDied","Data":"e60bf3d19a61de1ae31dab8790406cfedfe9d50d14c735d12eee42779276bf94"} Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.374683 4775 scope.go:117] "RemoveContainer" containerID="cb59be74b5f644a82a2f7f2898b6681b2938fc9c818cea6e2f9fa2e0d3233de9" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.409701 4775 scope.go:117] "RemoveContainer" containerID="95aeb45a504e567784b7144ed8cc19a83837748dc8e65efbb3ac700260b7133a" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.420637 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d6c869-m2kg7"] Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.429649 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d6c869-m2kg7"] Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.434167 4775 scope.go:117] "RemoveContainer" containerID="cb59be74b5f644a82a2f7f2898b6681b2938fc9c818cea6e2f9fa2e0d3233de9" Oct 02 03:27:51 crc kubenswrapper[4775]: E1002 03:27:51.434842 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb59be74b5f644a82a2f7f2898b6681b2938fc9c818cea6e2f9fa2e0d3233de9\": container with ID starting with cb59be74b5f644a82a2f7f2898b6681b2938fc9c818cea6e2f9fa2e0d3233de9 not found: ID does not exist" containerID="cb59be74b5f644a82a2f7f2898b6681b2938fc9c818cea6e2f9fa2e0d3233de9" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.434877 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb59be74b5f644a82a2f7f2898b6681b2938fc9c818cea6e2f9fa2e0d3233de9"} err="failed to get container status \"cb59be74b5f644a82a2f7f2898b6681b2938fc9c818cea6e2f9fa2e0d3233de9\": rpc error: code = NotFound desc = could not find container \"cb59be74b5f644a82a2f7f2898b6681b2938fc9c818cea6e2f9fa2e0d3233de9\": container with ID starting with cb59be74b5f644a82a2f7f2898b6681b2938fc9c818cea6e2f9fa2e0d3233de9 not found: ID does not exist" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.434905 4775 scope.go:117] "RemoveContainer" containerID="95aeb45a504e567784b7144ed8cc19a83837748dc8e65efbb3ac700260b7133a" Oct 02 03:27:51 crc kubenswrapper[4775]: E1002 03:27:51.435389 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95aeb45a504e567784b7144ed8cc19a83837748dc8e65efbb3ac700260b7133a\": container with ID starting with 95aeb45a504e567784b7144ed8cc19a83837748dc8e65efbb3ac700260b7133a not found: ID does not exist" containerID="95aeb45a504e567784b7144ed8cc19a83837748dc8e65efbb3ac700260b7133a" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.435422 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95aeb45a504e567784b7144ed8cc19a83837748dc8e65efbb3ac700260b7133a"} err="failed to get container status \"95aeb45a504e567784b7144ed8cc19a83837748dc8e65efbb3ac700260b7133a\": rpc error: code = NotFound desc = could not find container \"95aeb45a504e567784b7144ed8cc19a83837748dc8e65efbb3ac700260b7133a\": container with ID starting with 95aeb45a504e567784b7144ed8cc19a83837748dc8e65efbb3ac700260b7133a not found: ID does not exist" Oct 02 03:27:51 crc kubenswrapper[4775]: I1002 03:27:51.787741 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86bb6a99-2caa-478e-b73f-7c6e9b24808a" path="/var/lib/kubelet/pods/86bb6a99-2caa-478e-b73f-7c6e9b24808a/volumes" Oct 02 03:27:54 crc kubenswrapper[4775]: I1002 03:27:54.765713 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:27:54 crc kubenswrapper[4775]: E1002 03:27:54.766525 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.664316 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp"] Oct 02 03:28:01 crc kubenswrapper[4775]: E1002 03:28:01.665498 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35407ca5-0ace-441b-8d6f-e01db17cfa26" containerName="dnsmasq-dns" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.665514 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="35407ca5-0ace-441b-8d6f-e01db17cfa26" containerName="dnsmasq-dns" Oct 02 03:28:01 crc kubenswrapper[4775]: E1002 03:28:01.665535 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86bb6a99-2caa-478e-b73f-7c6e9b24808a" containerName="dnsmasq-dns" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.665544 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="86bb6a99-2caa-478e-b73f-7c6e9b24808a" containerName="dnsmasq-dns" Oct 02 03:28:01 crc kubenswrapper[4775]: E1002 03:28:01.665585 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86bb6a99-2caa-478e-b73f-7c6e9b24808a" containerName="init" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.665594 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="86bb6a99-2caa-478e-b73f-7c6e9b24808a" containerName="init" Oct 02 03:28:01 crc kubenswrapper[4775]: E1002 03:28:01.665616 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35407ca5-0ace-441b-8d6f-e01db17cfa26" containerName="init" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.665623 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="35407ca5-0ace-441b-8d6f-e01db17cfa26" containerName="init" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.665935 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="35407ca5-0ace-441b-8d6f-e01db17cfa26" containerName="dnsmasq-dns" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.665975 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="86bb6a99-2caa-478e-b73f-7c6e9b24808a" containerName="dnsmasq-dns" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.667069 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.669299 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.669612 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.672885 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.679054 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.681841 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp"] Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.693050 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.693100 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.693132 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zm27w\" (UniqueName: \"kubernetes.io/projected/e6b88bf2-a076-4957-8cd5-7777afcceead-kube-api-access-zm27w\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.693173 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.693224 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.794249 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.794336 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.794452 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.794491 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.794515 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zm27w\" (UniqueName: \"kubernetes.io/projected/e6b88bf2-a076-4957-8cd5-7777afcceead-kube-api-access-zm27w\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.800791 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.801414 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.802034 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.803188 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:01 crc kubenswrapper[4775]: I1002 03:28:01.811325 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zm27w\" (UniqueName: \"kubernetes.io/projected/e6b88bf2-a076-4957-8cd5-7777afcceead-kube-api-access-zm27w\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:02 crc kubenswrapper[4775]: I1002 03:28:02.034404 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:02 crc kubenswrapper[4775]: I1002 03:28:02.667824 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp"] Oct 02 03:28:02 crc kubenswrapper[4775]: W1002 03:28:02.668169 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6b88bf2_a076_4957_8cd5_7777afcceead.slice/crio-1cacaafb8d6c6cce48159078bd95fb9e4d59896ab67f4497c1257a57e6dd087c WatchSource:0}: Error finding container 1cacaafb8d6c6cce48159078bd95fb9e4d59896ab67f4497c1257a57e6dd087c: Status 404 returned error can't find the container with id 1cacaafb8d6c6cce48159078bd95fb9e4d59896ab67f4497c1257a57e6dd087c Oct 02 03:28:03 crc kubenswrapper[4775]: I1002 03:28:03.520251 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" event={"ID":"e6b88bf2-a076-4957-8cd5-7777afcceead","Type":"ContainerStarted","Data":"1cacaafb8d6c6cce48159078bd95fb9e4d59896ab67f4497c1257a57e6dd087c"} Oct 02 03:28:07 crc kubenswrapper[4775]: I1002 03:28:07.765100 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:28:07 crc kubenswrapper[4775]: E1002 03:28:07.766155 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:28:11 crc kubenswrapper[4775]: I1002 03:28:11.092719 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:28:11 crc kubenswrapper[4775]: I1002 03:28:11.609356 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" event={"ID":"e6b88bf2-a076-4957-8cd5-7777afcceead","Type":"ContainerStarted","Data":"7b19596a0461efa488c7d039f5d744f6e3f3811889c3375a7c4b0850039d1f50"} Oct 02 03:28:11 crc kubenswrapper[4775]: I1002 03:28:11.637144 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" podStartSLOduration=2.219853489 podStartE2EDuration="10.637117698s" podCreationTimestamp="2025-10-02 03:28:01 +0000 UTC" firstStartedPulling="2025-10-02 03:28:02.671675731 +0000 UTC m=+6419.838419761" lastFinishedPulling="2025-10-02 03:28:11.08893989 +0000 UTC m=+6428.255683970" observedRunningTime="2025-10-02 03:28:11.634532151 +0000 UTC m=+6428.801276231" watchObservedRunningTime="2025-10-02 03:28:11.637117698 +0000 UTC m=+6428.803861778" Oct 02 03:28:14 crc kubenswrapper[4775]: I1002 03:28:14.808835 4775 scope.go:117] "RemoveContainer" containerID="08daed43b5b2c3af8cd8d2ed33d2ab6c8d4215a06a0afa0b06eac6c17223cc76" Oct 02 03:28:14 crc kubenswrapper[4775]: I1002 03:28:14.840056 4775 scope.go:117] "RemoveContainer" containerID="093c7b5016f58ac468b44bcde981ff59684d781a1871d62c5b8003687c1c1678" Oct 02 03:28:14 crc kubenswrapper[4775]: I1002 03:28:14.866489 4775 scope.go:117] "RemoveContainer" containerID="c906ec36c1f6343283414031f74bada9b06dccd28fcf9b63b7e95d29c96e0463" Oct 02 03:28:15 crc kubenswrapper[4775]: I1002 03:28:15.118299 4775 scope.go:117] "RemoveContainer" containerID="64326c41f019909555f9a29048ce5f52f58aa35c55b23859039ccdd6fecbaf50" Oct 02 03:28:17 crc kubenswrapper[4775]: I1002 03:28:17.070499 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-fvb7k"] Oct 02 03:28:17 crc kubenswrapper[4775]: I1002 03:28:17.087315 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-fvb7k"] Oct 02 03:28:17 crc kubenswrapper[4775]: I1002 03:28:17.789529 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2dc1393d-8d30-4fd9-9e47-a6f5d0df2314" path="/var/lib/kubelet/pods/2dc1393d-8d30-4fd9-9e47-a6f5d0df2314/volumes" Oct 02 03:28:21 crc kubenswrapper[4775]: I1002 03:28:21.766031 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:28:21 crc kubenswrapper[4775]: E1002 03:28:21.766842 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:28:25 crc kubenswrapper[4775]: I1002 03:28:25.806049 4775 generic.go:334] "Generic (PLEG): container finished" podID="e6b88bf2-a076-4957-8cd5-7777afcceead" containerID="7b19596a0461efa488c7d039f5d744f6e3f3811889c3375a7c4b0850039d1f50" exitCode=0 Oct 02 03:28:25 crc kubenswrapper[4775]: I1002 03:28:25.806174 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" event={"ID":"e6b88bf2-a076-4957-8cd5-7777afcceead","Type":"ContainerDied","Data":"7b19596a0461efa488c7d039f5d744f6e3f3811889c3375a7c4b0850039d1f50"} Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.389526 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.552603 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-ceph\") pod \"e6b88bf2-a076-4957-8cd5-7777afcceead\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.552869 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-pre-adoption-validation-combined-ca-bundle\") pod \"e6b88bf2-a076-4957-8cd5-7777afcceead\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.553262 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zm27w\" (UniqueName: \"kubernetes.io/projected/e6b88bf2-a076-4957-8cd5-7777afcceead-kube-api-access-zm27w\") pod \"e6b88bf2-a076-4957-8cd5-7777afcceead\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.553395 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-inventory\") pod \"e6b88bf2-a076-4957-8cd5-7777afcceead\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.553615 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-ssh-key\") pod \"e6b88bf2-a076-4957-8cd5-7777afcceead\" (UID: \"e6b88bf2-a076-4957-8cd5-7777afcceead\") " Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.558310 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-ceph" (OuterVolumeSpecName: "ceph") pod "e6b88bf2-a076-4957-8cd5-7777afcceead" (UID: "e6b88bf2-a076-4957-8cd5-7777afcceead"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.558848 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6b88bf2-a076-4957-8cd5-7777afcceead-kube-api-access-zm27w" (OuterVolumeSpecName: "kube-api-access-zm27w") pod "e6b88bf2-a076-4957-8cd5-7777afcceead" (UID: "e6b88bf2-a076-4957-8cd5-7777afcceead"). InnerVolumeSpecName "kube-api-access-zm27w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.567082 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "e6b88bf2-a076-4957-8cd5-7777afcceead" (UID: "e6b88bf2-a076-4957-8cd5-7777afcceead"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.583214 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-inventory" (OuterVolumeSpecName: "inventory") pod "e6b88bf2-a076-4957-8cd5-7777afcceead" (UID: "e6b88bf2-a076-4957-8cd5-7777afcceead"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.592348 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e6b88bf2-a076-4957-8cd5-7777afcceead" (UID: "e6b88bf2-a076-4957-8cd5-7777afcceead"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.656149 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.656418 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.656428 4775 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.656442 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zm27w\" (UniqueName: \"kubernetes.io/projected/e6b88bf2-a076-4957-8cd5-7777afcceead-kube-api-access-zm27w\") on node \"crc\" DevicePath \"\"" Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.656453 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6b88bf2-a076-4957-8cd5-7777afcceead-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.843768 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" event={"ID":"e6b88bf2-a076-4957-8cd5-7777afcceead","Type":"ContainerDied","Data":"1cacaafb8d6c6cce48159078bd95fb9e4d59896ab67f4497c1257a57e6dd087c"} Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.843815 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cacaafb8d6c6cce48159078bd95fb9e4d59896ab67f4497c1257a57e6dd087c" Oct 02 03:28:27 crc kubenswrapper[4775]: I1002 03:28:27.843897 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp" Oct 02 03:28:30 crc kubenswrapper[4775]: I1002 03:28:30.057858 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-e022-account-create-ggfk2"] Oct 02 03:28:30 crc kubenswrapper[4775]: I1002 03:28:30.071803 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-e022-account-create-ggfk2"] Oct 02 03:28:31 crc kubenswrapper[4775]: I1002 03:28:31.787562 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad" path="/var/lib/kubelet/pods/ebbd91b0-4bb4-4edc-a7fe-2486fe8899ad/volumes" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.285402 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m"] Oct 02 03:28:35 crc kubenswrapper[4775]: E1002 03:28:35.286700 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6b88bf2-a076-4957-8cd5-7777afcceead" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.286724 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6b88bf2-a076-4957-8cd5-7777afcceead" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.287124 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6b88bf2-a076-4957-8cd5-7777afcceead" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.288413 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.291253 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.291552 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.294411 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.301409 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.306813 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m"] Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.451487 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m96fp\" (UniqueName: \"kubernetes.io/projected/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-kube-api-access-m96fp\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.451599 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.451668 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.451845 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.452231 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.554537 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.554692 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.554748 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m96fp\" (UniqueName: \"kubernetes.io/projected/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-kube-api-access-m96fp\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.554779 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.554805 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.562900 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.564045 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.564717 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.576235 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.583029 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m96fp\" (UniqueName: \"kubernetes.io/projected/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-kube-api-access-m96fp\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:35 crc kubenswrapper[4775]: I1002 03:28:35.614876 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:28:36 crc kubenswrapper[4775]: I1002 03:28:36.259230 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m"] Oct 02 03:28:36 crc kubenswrapper[4775]: W1002 03:28:36.263517 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3959dab0_5e51_4cb7_a540_96ae7c7a28b9.slice/crio-146139aee9b93023ca907ac1413f1961f88bc23385ea78adfa041e4ad4fdf75f WatchSource:0}: Error finding container 146139aee9b93023ca907ac1413f1961f88bc23385ea78adfa041e4ad4fdf75f: Status 404 returned error can't find the container with id 146139aee9b93023ca907ac1413f1961f88bc23385ea78adfa041e4ad4fdf75f Oct 02 03:28:36 crc kubenswrapper[4775]: I1002 03:28:36.765457 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:28:36 crc kubenswrapper[4775]: E1002 03:28:36.766103 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:28:36 crc kubenswrapper[4775]: I1002 03:28:36.952083 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" event={"ID":"3959dab0-5e51-4cb7-a540-96ae7c7a28b9","Type":"ContainerStarted","Data":"146139aee9b93023ca907ac1413f1961f88bc23385ea78adfa041e4ad4fdf75f"} Oct 02 03:28:37 crc kubenswrapper[4775]: I1002 03:28:37.043035 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-hkplh"] Oct 02 03:28:37 crc kubenswrapper[4775]: I1002 03:28:37.056131 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-hkplh"] Oct 02 03:28:37 crc kubenswrapper[4775]: I1002 03:28:37.782895 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9f570ab-6a73-47ee-8b8b-64c3f78a9d34" path="/var/lib/kubelet/pods/c9f570ab-6a73-47ee-8b8b-64c3f78a9d34/volumes" Oct 02 03:28:37 crc kubenswrapper[4775]: I1002 03:28:37.966577 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" event={"ID":"3959dab0-5e51-4cb7-a540-96ae7c7a28b9","Type":"ContainerStarted","Data":"7747f9a3c676ae1eb0c97b3d9d7209af2dd7ca4986e4721c56050fea17e6a762"} Oct 02 03:28:38 crc kubenswrapper[4775]: I1002 03:28:38.003386 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" podStartSLOduration=2.427620044 podStartE2EDuration="3.003335075s" podCreationTimestamp="2025-10-02 03:28:35 +0000 UTC" firstStartedPulling="2025-10-02 03:28:36.268414161 +0000 UTC m=+6453.435158201" lastFinishedPulling="2025-10-02 03:28:36.844129182 +0000 UTC m=+6454.010873232" observedRunningTime="2025-10-02 03:28:37.988769618 +0000 UTC m=+6455.155513698" watchObservedRunningTime="2025-10-02 03:28:38.003335075 +0000 UTC m=+6455.170079145" Oct 02 03:28:48 crc kubenswrapper[4775]: I1002 03:28:48.057202 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-69a7-account-create-5hjm6"] Oct 02 03:28:48 crc kubenswrapper[4775]: I1002 03:28:48.073831 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-69a7-account-create-5hjm6"] Oct 02 03:28:49 crc kubenswrapper[4775]: I1002 03:28:49.786996 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c212a5c8-90ef-45dc-9da1-762c24eafd75" path="/var/lib/kubelet/pods/c212a5c8-90ef-45dc-9da1-762c24eafd75/volumes" Oct 02 03:28:50 crc kubenswrapper[4775]: I1002 03:28:50.766028 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:28:50 crc kubenswrapper[4775]: E1002 03:28:50.766841 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:29:03 crc kubenswrapper[4775]: I1002 03:29:03.781379 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:29:03 crc kubenswrapper[4775]: E1002 03:29:03.782878 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:29:14 crc kubenswrapper[4775]: I1002 03:29:14.765851 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:29:14 crc kubenswrapper[4775]: E1002 03:29:14.767055 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:29:15 crc kubenswrapper[4775]: I1002 03:29:15.524448 4775 scope.go:117] "RemoveContainer" containerID="5dada035b80ae1e48617d87704761e1c28e4e47be96334ff0dc040c835851d7d" Oct 02 03:29:15 crc kubenswrapper[4775]: I1002 03:29:15.562035 4775 scope.go:117] "RemoveContainer" containerID="243512ee77002e8bc897149f111abb5d5fa338b9e7fdf3d46fa9b2b4c746504e" Oct 02 03:29:15 crc kubenswrapper[4775]: I1002 03:29:15.640432 4775 scope.go:117] "RemoveContainer" containerID="3cc458720d01fd6ac7affe6c97741c4468afd8751f16cc478f16c98fb8f9ca57" Oct 02 03:29:15 crc kubenswrapper[4775]: I1002 03:29:15.680409 4775 scope.go:117] "RemoveContainer" containerID="8a22eacafcc1c5d4ea5bcf410fcf7ad2bc798f3cfabcf2fe8fdaaa7eeafd07ff" Oct 02 03:29:27 crc kubenswrapper[4775]: I1002 03:29:27.765809 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:29:27 crc kubenswrapper[4775]: E1002 03:29:27.766824 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:29:29 crc kubenswrapper[4775]: I1002 03:29:29.053943 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-5vw8g"] Oct 02 03:29:29 crc kubenswrapper[4775]: I1002 03:29:29.063398 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-5vw8g"] Oct 02 03:29:29 crc kubenswrapper[4775]: I1002 03:29:29.809417 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b44ecb2-8bf4-4c25-841b-3885dbbd0318" path="/var/lib/kubelet/pods/7b44ecb2-8bf4-4c25-841b-3885dbbd0318/volumes" Oct 02 03:29:39 crc kubenswrapper[4775]: I1002 03:29:39.785190 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:29:39 crc kubenswrapper[4775]: E1002 03:29:39.785885 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:29:53 crc kubenswrapper[4775]: I1002 03:29:53.780028 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:29:53 crc kubenswrapper[4775]: E1002 03:29:53.781026 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.192311 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6"] Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.195308 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.198013 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.198271 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.212633 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6"] Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.273399 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f05c8229-d21d-4c03-abc5-5e5ef54121ee-config-volume\") pod \"collect-profiles-29322930-kv9d6\" (UID: \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.273705 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvtdg\" (UniqueName: \"kubernetes.io/projected/f05c8229-d21d-4c03-abc5-5e5ef54121ee-kube-api-access-dvtdg\") pod \"collect-profiles-29322930-kv9d6\" (UID: \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.273783 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f05c8229-d21d-4c03-abc5-5e5ef54121ee-secret-volume\") pod \"collect-profiles-29322930-kv9d6\" (UID: \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.375758 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f05c8229-d21d-4c03-abc5-5e5ef54121ee-config-volume\") pod \"collect-profiles-29322930-kv9d6\" (UID: \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.375960 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvtdg\" (UniqueName: \"kubernetes.io/projected/f05c8229-d21d-4c03-abc5-5e5ef54121ee-kube-api-access-dvtdg\") pod \"collect-profiles-29322930-kv9d6\" (UID: \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.376061 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f05c8229-d21d-4c03-abc5-5e5ef54121ee-secret-volume\") pod \"collect-profiles-29322930-kv9d6\" (UID: \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.377041 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f05c8229-d21d-4c03-abc5-5e5ef54121ee-config-volume\") pod \"collect-profiles-29322930-kv9d6\" (UID: \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.386346 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f05c8229-d21d-4c03-abc5-5e5ef54121ee-secret-volume\") pod \"collect-profiles-29322930-kv9d6\" (UID: \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.401110 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvtdg\" (UniqueName: \"kubernetes.io/projected/f05c8229-d21d-4c03-abc5-5e5ef54121ee-kube-api-access-dvtdg\") pod \"collect-profiles-29322930-kv9d6\" (UID: \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" Oct 02 03:30:00 crc kubenswrapper[4775]: I1002 03:30:00.533986 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" Oct 02 03:30:01 crc kubenswrapper[4775]: I1002 03:30:01.067395 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6"] Oct 02 03:30:02 crc kubenswrapper[4775]: I1002 03:30:02.081429 4775 generic.go:334] "Generic (PLEG): container finished" podID="f05c8229-d21d-4c03-abc5-5e5ef54121ee" containerID="89cbeaae642f9ccfa7de8f57341e73a0e28d7ac4555c74e4c752835f00e9761c" exitCode=0 Oct 02 03:30:02 crc kubenswrapper[4775]: I1002 03:30:02.081554 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" event={"ID":"f05c8229-d21d-4c03-abc5-5e5ef54121ee","Type":"ContainerDied","Data":"89cbeaae642f9ccfa7de8f57341e73a0e28d7ac4555c74e4c752835f00e9761c"} Oct 02 03:30:02 crc kubenswrapper[4775]: I1002 03:30:02.083115 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" event={"ID":"f05c8229-d21d-4c03-abc5-5e5ef54121ee","Type":"ContainerStarted","Data":"6efb0fddf81838d857fcfcdf5c9d11dd60eecaf69447493007e2749c87cd3bdc"} Oct 02 03:30:03 crc kubenswrapper[4775]: I1002 03:30:03.677638 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" Oct 02 03:30:03 crc kubenswrapper[4775]: I1002 03:30:03.757426 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvtdg\" (UniqueName: \"kubernetes.io/projected/f05c8229-d21d-4c03-abc5-5e5ef54121ee-kube-api-access-dvtdg\") pod \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\" (UID: \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\") " Oct 02 03:30:03 crc kubenswrapper[4775]: I1002 03:30:03.757493 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f05c8229-d21d-4c03-abc5-5e5ef54121ee-config-volume\") pod \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\" (UID: \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\") " Oct 02 03:30:03 crc kubenswrapper[4775]: I1002 03:30:03.758193 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f05c8229-d21d-4c03-abc5-5e5ef54121ee-secret-volume\") pod \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\" (UID: \"f05c8229-d21d-4c03-abc5-5e5ef54121ee\") " Oct 02 03:30:03 crc kubenswrapper[4775]: I1002 03:30:03.758632 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f05c8229-d21d-4c03-abc5-5e5ef54121ee-config-volume" (OuterVolumeSpecName: "config-volume") pod "f05c8229-d21d-4c03-abc5-5e5ef54121ee" (UID: "f05c8229-d21d-4c03-abc5-5e5ef54121ee"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:30:03 crc kubenswrapper[4775]: I1002 03:30:03.759380 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f05c8229-d21d-4c03-abc5-5e5ef54121ee-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 03:30:03 crc kubenswrapper[4775]: I1002 03:30:03.764480 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f05c8229-d21d-4c03-abc5-5e5ef54121ee-kube-api-access-dvtdg" (OuterVolumeSpecName: "kube-api-access-dvtdg") pod "f05c8229-d21d-4c03-abc5-5e5ef54121ee" (UID: "f05c8229-d21d-4c03-abc5-5e5ef54121ee"). InnerVolumeSpecName "kube-api-access-dvtdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:30:03 crc kubenswrapper[4775]: I1002 03:30:03.766492 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f05c8229-d21d-4c03-abc5-5e5ef54121ee-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f05c8229-d21d-4c03-abc5-5e5ef54121ee" (UID: "f05c8229-d21d-4c03-abc5-5e5ef54121ee"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:30:03 crc kubenswrapper[4775]: I1002 03:30:03.861211 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvtdg\" (UniqueName: \"kubernetes.io/projected/f05c8229-d21d-4c03-abc5-5e5ef54121ee-kube-api-access-dvtdg\") on node \"crc\" DevicePath \"\"" Oct 02 03:30:03 crc kubenswrapper[4775]: I1002 03:30:03.861533 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f05c8229-d21d-4c03-abc5-5e5ef54121ee-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 03:30:04 crc kubenswrapper[4775]: I1002 03:30:04.112505 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" event={"ID":"f05c8229-d21d-4c03-abc5-5e5ef54121ee","Type":"ContainerDied","Data":"6efb0fddf81838d857fcfcdf5c9d11dd60eecaf69447493007e2749c87cd3bdc"} Oct 02 03:30:04 crc kubenswrapper[4775]: I1002 03:30:04.112562 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6efb0fddf81838d857fcfcdf5c9d11dd60eecaf69447493007e2749c87cd3bdc" Oct 02 03:30:04 crc kubenswrapper[4775]: I1002 03:30:04.112696 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6" Oct 02 03:30:04 crc kubenswrapper[4775]: I1002 03:30:04.771285 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl"] Oct 02 03:30:04 crc kubenswrapper[4775]: I1002 03:30:04.779247 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322885-mptjl"] Oct 02 03:30:05 crc kubenswrapper[4775]: I1002 03:30:05.803555 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db403783-c3a4-4423-8cc9-16d1b49c2dd3" path="/var/lib/kubelet/pods/db403783-c3a4-4423-8cc9-16d1b49c2dd3/volumes" Oct 02 03:30:07 crc kubenswrapper[4775]: I1002 03:30:07.766434 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:30:07 crc kubenswrapper[4775]: E1002 03:30:07.767314 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:30:15 crc kubenswrapper[4775]: I1002 03:30:15.822624 4775 scope.go:117] "RemoveContainer" containerID="40b8b46191a6a080a0e34cb85e82b0ecebd35223ea9e8b197cd83e77f16af080" Oct 02 03:30:15 crc kubenswrapper[4775]: I1002 03:30:15.869370 4775 scope.go:117] "RemoveContainer" containerID="5f56e5ffaef8e27022e4e93c19e7765239c07b45e7a61987b5e7c35cfa99c796" Oct 02 03:30:15 crc kubenswrapper[4775]: I1002 03:30:15.929141 4775 scope.go:117] "RemoveContainer" containerID="f243ea03d6cd70473c444c5b6cfa5bf7f79cb3ced96e5c68bc0c93d48e079f4e" Oct 02 03:30:19 crc kubenswrapper[4775]: I1002 03:30:19.765855 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:30:19 crc kubenswrapper[4775]: E1002 03:30:19.767006 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:30:32 crc kubenswrapper[4775]: I1002 03:30:32.766654 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:30:32 crc kubenswrapper[4775]: E1002 03:30:32.768138 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:30:39 crc kubenswrapper[4775]: I1002 03:30:39.988008 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hhknv"] Oct 02 03:30:39 crc kubenswrapper[4775]: E1002 03:30:39.989315 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f05c8229-d21d-4c03-abc5-5e5ef54121ee" containerName="collect-profiles" Oct 02 03:30:39 crc kubenswrapper[4775]: I1002 03:30:39.989332 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f05c8229-d21d-4c03-abc5-5e5ef54121ee" containerName="collect-profiles" Oct 02 03:30:39 crc kubenswrapper[4775]: I1002 03:30:39.989560 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f05c8229-d21d-4c03-abc5-5e5ef54121ee" containerName="collect-profiles" Oct 02 03:30:39 crc kubenswrapper[4775]: I1002 03:30:39.991605 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:40 crc kubenswrapper[4775]: I1002 03:30:40.002894 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhknv"] Oct 02 03:30:40 crc kubenswrapper[4775]: I1002 03:30:40.095768 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq7m8\" (UniqueName: \"kubernetes.io/projected/24a3b415-2d1f-4fec-a623-424f746acfb8-kube-api-access-nq7m8\") pod \"redhat-marketplace-hhknv\" (UID: \"24a3b415-2d1f-4fec-a623-424f746acfb8\") " pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:40 crc kubenswrapper[4775]: I1002 03:30:40.095837 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24a3b415-2d1f-4fec-a623-424f746acfb8-catalog-content\") pod \"redhat-marketplace-hhknv\" (UID: \"24a3b415-2d1f-4fec-a623-424f746acfb8\") " pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:40 crc kubenswrapper[4775]: I1002 03:30:40.096804 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24a3b415-2d1f-4fec-a623-424f746acfb8-utilities\") pod \"redhat-marketplace-hhknv\" (UID: \"24a3b415-2d1f-4fec-a623-424f746acfb8\") " pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:40 crc kubenswrapper[4775]: I1002 03:30:40.198554 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24a3b415-2d1f-4fec-a623-424f746acfb8-utilities\") pod \"redhat-marketplace-hhknv\" (UID: \"24a3b415-2d1f-4fec-a623-424f746acfb8\") " pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:40 crc kubenswrapper[4775]: I1002 03:30:40.198727 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq7m8\" (UniqueName: \"kubernetes.io/projected/24a3b415-2d1f-4fec-a623-424f746acfb8-kube-api-access-nq7m8\") pod \"redhat-marketplace-hhknv\" (UID: \"24a3b415-2d1f-4fec-a623-424f746acfb8\") " pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:40 crc kubenswrapper[4775]: I1002 03:30:40.198782 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24a3b415-2d1f-4fec-a623-424f746acfb8-catalog-content\") pod \"redhat-marketplace-hhknv\" (UID: \"24a3b415-2d1f-4fec-a623-424f746acfb8\") " pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:40 crc kubenswrapper[4775]: I1002 03:30:40.199113 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24a3b415-2d1f-4fec-a623-424f746acfb8-utilities\") pod \"redhat-marketplace-hhknv\" (UID: \"24a3b415-2d1f-4fec-a623-424f746acfb8\") " pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:40 crc kubenswrapper[4775]: I1002 03:30:40.199482 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24a3b415-2d1f-4fec-a623-424f746acfb8-catalog-content\") pod \"redhat-marketplace-hhknv\" (UID: \"24a3b415-2d1f-4fec-a623-424f746acfb8\") " pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:40 crc kubenswrapper[4775]: I1002 03:30:40.218792 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq7m8\" (UniqueName: \"kubernetes.io/projected/24a3b415-2d1f-4fec-a623-424f746acfb8-kube-api-access-nq7m8\") pod \"redhat-marketplace-hhknv\" (UID: \"24a3b415-2d1f-4fec-a623-424f746acfb8\") " pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:40 crc kubenswrapper[4775]: I1002 03:30:40.350887 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:40 crc kubenswrapper[4775]: I1002 03:30:40.840589 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhknv"] Oct 02 03:30:40 crc kubenswrapper[4775]: W1002 03:30:40.847677 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24a3b415_2d1f_4fec_a623_424f746acfb8.slice/crio-b162cbb88300d239c1fc8a7e88b8f131b189459d0ef63de70803f724019bc2f7 WatchSource:0}: Error finding container b162cbb88300d239c1fc8a7e88b8f131b189459d0ef63de70803f724019bc2f7: Status 404 returned error can't find the container with id b162cbb88300d239c1fc8a7e88b8f131b189459d0ef63de70803f724019bc2f7 Oct 02 03:30:41 crc kubenswrapper[4775]: I1002 03:30:41.605920 4775 generic.go:334] "Generic (PLEG): container finished" podID="24a3b415-2d1f-4fec-a623-424f746acfb8" containerID="42841f9f186038d0b0943fc6279725abcd344f7aa357633638da2eb31cae0721" exitCode=0 Oct 02 03:30:41 crc kubenswrapper[4775]: I1002 03:30:41.606036 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhknv" event={"ID":"24a3b415-2d1f-4fec-a623-424f746acfb8","Type":"ContainerDied","Data":"42841f9f186038d0b0943fc6279725abcd344f7aa357633638da2eb31cae0721"} Oct 02 03:30:41 crc kubenswrapper[4775]: I1002 03:30:41.608042 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhknv" event={"ID":"24a3b415-2d1f-4fec-a623-424f746acfb8","Type":"ContainerStarted","Data":"b162cbb88300d239c1fc8a7e88b8f131b189459d0ef63de70803f724019bc2f7"} Oct 02 03:30:43 crc kubenswrapper[4775]: I1002 03:30:43.633042 4775 generic.go:334] "Generic (PLEG): container finished" podID="24a3b415-2d1f-4fec-a623-424f746acfb8" containerID="b9a3a9e2bc5ed238cbf33f13bba9725320e8a6dd6629c2e8f6f850ce228a5917" exitCode=0 Oct 02 03:30:43 crc kubenswrapper[4775]: I1002 03:30:43.633181 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhknv" event={"ID":"24a3b415-2d1f-4fec-a623-424f746acfb8","Type":"ContainerDied","Data":"b9a3a9e2bc5ed238cbf33f13bba9725320e8a6dd6629c2e8f6f850ce228a5917"} Oct 02 03:30:44 crc kubenswrapper[4775]: I1002 03:30:44.660302 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhknv" event={"ID":"24a3b415-2d1f-4fec-a623-424f746acfb8","Type":"ContainerStarted","Data":"0ada6f881c4144cfa5c461a2a843979a2bde01cab43aceb55a7049673046dd16"} Oct 02 03:30:44 crc kubenswrapper[4775]: I1002 03:30:44.696356 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hhknv" podStartSLOduration=3.034029773 podStartE2EDuration="5.696335218s" podCreationTimestamp="2025-10-02 03:30:39 +0000 UTC" firstStartedPulling="2025-10-02 03:30:41.60969023 +0000 UTC m=+6578.776434280" lastFinishedPulling="2025-10-02 03:30:44.271995655 +0000 UTC m=+6581.438739725" observedRunningTime="2025-10-02 03:30:44.68018292 +0000 UTC m=+6581.846926970" watchObservedRunningTime="2025-10-02 03:30:44.696335218 +0000 UTC m=+6581.863079268" Oct 02 03:30:46 crc kubenswrapper[4775]: I1002 03:30:46.765811 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:30:46 crc kubenswrapper[4775]: E1002 03:30:46.766435 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:30:50 crc kubenswrapper[4775]: I1002 03:30:50.351886 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:50 crc kubenswrapper[4775]: I1002 03:30:50.352613 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:50 crc kubenswrapper[4775]: I1002 03:30:50.434324 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:50 crc kubenswrapper[4775]: I1002 03:30:50.822014 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:50 crc kubenswrapper[4775]: I1002 03:30:50.896347 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhknv"] Oct 02 03:30:52 crc kubenswrapper[4775]: I1002 03:30:52.766859 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hhknv" podUID="24a3b415-2d1f-4fec-a623-424f746acfb8" containerName="registry-server" containerID="cri-o://0ada6f881c4144cfa5c461a2a843979a2bde01cab43aceb55a7049673046dd16" gracePeriod=2 Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.296213 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.456783 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24a3b415-2d1f-4fec-a623-424f746acfb8-catalog-content\") pod \"24a3b415-2d1f-4fec-a623-424f746acfb8\" (UID: \"24a3b415-2d1f-4fec-a623-424f746acfb8\") " Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.457089 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24a3b415-2d1f-4fec-a623-424f746acfb8-utilities\") pod \"24a3b415-2d1f-4fec-a623-424f746acfb8\" (UID: \"24a3b415-2d1f-4fec-a623-424f746acfb8\") " Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.457146 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq7m8\" (UniqueName: \"kubernetes.io/projected/24a3b415-2d1f-4fec-a623-424f746acfb8-kube-api-access-nq7m8\") pod \"24a3b415-2d1f-4fec-a623-424f746acfb8\" (UID: \"24a3b415-2d1f-4fec-a623-424f746acfb8\") " Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.458572 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24a3b415-2d1f-4fec-a623-424f746acfb8-utilities" (OuterVolumeSpecName: "utilities") pod "24a3b415-2d1f-4fec-a623-424f746acfb8" (UID: "24a3b415-2d1f-4fec-a623-424f746acfb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.464400 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24a3b415-2d1f-4fec-a623-424f746acfb8-kube-api-access-nq7m8" (OuterVolumeSpecName: "kube-api-access-nq7m8") pod "24a3b415-2d1f-4fec-a623-424f746acfb8" (UID: "24a3b415-2d1f-4fec-a623-424f746acfb8"). InnerVolumeSpecName "kube-api-access-nq7m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.484843 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24a3b415-2d1f-4fec-a623-424f746acfb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24a3b415-2d1f-4fec-a623-424f746acfb8" (UID: "24a3b415-2d1f-4fec-a623-424f746acfb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.560002 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24a3b415-2d1f-4fec-a623-424f746acfb8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.560033 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24a3b415-2d1f-4fec-a623-424f746acfb8-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.560043 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq7m8\" (UniqueName: \"kubernetes.io/projected/24a3b415-2d1f-4fec-a623-424f746acfb8-kube-api-access-nq7m8\") on node \"crc\" DevicePath \"\"" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.785461 4775 generic.go:334] "Generic (PLEG): container finished" podID="24a3b415-2d1f-4fec-a623-424f746acfb8" containerID="0ada6f881c4144cfa5c461a2a843979a2bde01cab43aceb55a7049673046dd16" exitCode=0 Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.785580 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hhknv" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.797408 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhknv" event={"ID":"24a3b415-2d1f-4fec-a623-424f746acfb8","Type":"ContainerDied","Data":"0ada6f881c4144cfa5c461a2a843979a2bde01cab43aceb55a7049673046dd16"} Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.797461 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hhknv" event={"ID":"24a3b415-2d1f-4fec-a623-424f746acfb8","Type":"ContainerDied","Data":"b162cbb88300d239c1fc8a7e88b8f131b189459d0ef63de70803f724019bc2f7"} Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.797487 4775 scope.go:117] "RemoveContainer" containerID="0ada6f881c4144cfa5c461a2a843979a2bde01cab43aceb55a7049673046dd16" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.837432 4775 scope.go:117] "RemoveContainer" containerID="b9a3a9e2bc5ed238cbf33f13bba9725320e8a6dd6629c2e8f6f850ce228a5917" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.855481 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhknv"] Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.870674 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hhknv"] Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.883278 4775 scope.go:117] "RemoveContainer" containerID="42841f9f186038d0b0943fc6279725abcd344f7aa357633638da2eb31cae0721" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.946266 4775 scope.go:117] "RemoveContainer" containerID="0ada6f881c4144cfa5c461a2a843979a2bde01cab43aceb55a7049673046dd16" Oct 02 03:30:53 crc kubenswrapper[4775]: E1002 03:30:53.946881 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ada6f881c4144cfa5c461a2a843979a2bde01cab43aceb55a7049673046dd16\": container with ID starting with 0ada6f881c4144cfa5c461a2a843979a2bde01cab43aceb55a7049673046dd16 not found: ID does not exist" containerID="0ada6f881c4144cfa5c461a2a843979a2bde01cab43aceb55a7049673046dd16" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.946935 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ada6f881c4144cfa5c461a2a843979a2bde01cab43aceb55a7049673046dd16"} err="failed to get container status \"0ada6f881c4144cfa5c461a2a843979a2bde01cab43aceb55a7049673046dd16\": rpc error: code = NotFound desc = could not find container \"0ada6f881c4144cfa5c461a2a843979a2bde01cab43aceb55a7049673046dd16\": container with ID starting with 0ada6f881c4144cfa5c461a2a843979a2bde01cab43aceb55a7049673046dd16 not found: ID does not exist" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.946987 4775 scope.go:117] "RemoveContainer" containerID="b9a3a9e2bc5ed238cbf33f13bba9725320e8a6dd6629c2e8f6f850ce228a5917" Oct 02 03:30:53 crc kubenswrapper[4775]: E1002 03:30:53.947382 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9a3a9e2bc5ed238cbf33f13bba9725320e8a6dd6629c2e8f6f850ce228a5917\": container with ID starting with b9a3a9e2bc5ed238cbf33f13bba9725320e8a6dd6629c2e8f6f850ce228a5917 not found: ID does not exist" containerID="b9a3a9e2bc5ed238cbf33f13bba9725320e8a6dd6629c2e8f6f850ce228a5917" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.947404 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9a3a9e2bc5ed238cbf33f13bba9725320e8a6dd6629c2e8f6f850ce228a5917"} err="failed to get container status \"b9a3a9e2bc5ed238cbf33f13bba9725320e8a6dd6629c2e8f6f850ce228a5917\": rpc error: code = NotFound desc = could not find container \"b9a3a9e2bc5ed238cbf33f13bba9725320e8a6dd6629c2e8f6f850ce228a5917\": container with ID starting with b9a3a9e2bc5ed238cbf33f13bba9725320e8a6dd6629c2e8f6f850ce228a5917 not found: ID does not exist" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.947415 4775 scope.go:117] "RemoveContainer" containerID="42841f9f186038d0b0943fc6279725abcd344f7aa357633638da2eb31cae0721" Oct 02 03:30:53 crc kubenswrapper[4775]: E1002 03:30:53.947770 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42841f9f186038d0b0943fc6279725abcd344f7aa357633638da2eb31cae0721\": container with ID starting with 42841f9f186038d0b0943fc6279725abcd344f7aa357633638da2eb31cae0721 not found: ID does not exist" containerID="42841f9f186038d0b0943fc6279725abcd344f7aa357633638da2eb31cae0721" Oct 02 03:30:53 crc kubenswrapper[4775]: I1002 03:30:53.947800 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42841f9f186038d0b0943fc6279725abcd344f7aa357633638da2eb31cae0721"} err="failed to get container status \"42841f9f186038d0b0943fc6279725abcd344f7aa357633638da2eb31cae0721\": rpc error: code = NotFound desc = could not find container \"42841f9f186038d0b0943fc6279725abcd344f7aa357633638da2eb31cae0721\": container with ID starting with 42841f9f186038d0b0943fc6279725abcd344f7aa357633638da2eb31cae0721 not found: ID does not exist" Oct 02 03:30:55 crc kubenswrapper[4775]: I1002 03:30:55.796525 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24a3b415-2d1f-4fec-a623-424f746acfb8" path="/var/lib/kubelet/pods/24a3b415-2d1f-4fec-a623-424f746acfb8/volumes" Oct 02 03:30:58 crc kubenswrapper[4775]: I1002 03:30:58.765799 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:30:58 crc kubenswrapper[4775]: E1002 03:30:58.766928 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:31:11 crc kubenswrapper[4775]: I1002 03:31:11.765605 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:31:11 crc kubenswrapper[4775]: E1002 03:31:11.766582 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:31:22 crc kubenswrapper[4775]: I1002 03:31:22.765320 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:31:22 crc kubenswrapper[4775]: E1002 03:31:22.766423 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:31:35 crc kubenswrapper[4775]: I1002 03:31:35.766579 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:31:35 crc kubenswrapper[4775]: E1002 03:31:35.768006 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.176322 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-crhfr"] Oct 02 03:31:39 crc kubenswrapper[4775]: E1002 03:31:39.177272 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24a3b415-2d1f-4fec-a623-424f746acfb8" containerName="extract-utilities" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.177287 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="24a3b415-2d1f-4fec-a623-424f746acfb8" containerName="extract-utilities" Oct 02 03:31:39 crc kubenswrapper[4775]: E1002 03:31:39.177311 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24a3b415-2d1f-4fec-a623-424f746acfb8" containerName="registry-server" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.177319 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="24a3b415-2d1f-4fec-a623-424f746acfb8" containerName="registry-server" Oct 02 03:31:39 crc kubenswrapper[4775]: E1002 03:31:39.177365 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24a3b415-2d1f-4fec-a623-424f746acfb8" containerName="extract-content" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.177375 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="24a3b415-2d1f-4fec-a623-424f746acfb8" containerName="extract-content" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.177628 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="24a3b415-2d1f-4fec-a623-424f746acfb8" containerName="registry-server" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.179477 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.193779 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-crhfr"] Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.288434 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-catalog-content\") pod \"community-operators-crhfr\" (UID: \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\") " pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.288487 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xsr5\" (UniqueName: \"kubernetes.io/projected/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-kube-api-access-7xsr5\") pod \"community-operators-crhfr\" (UID: \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\") " pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.288540 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-utilities\") pod \"community-operators-crhfr\" (UID: \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\") " pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.391445 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-utilities\") pod \"community-operators-crhfr\" (UID: \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\") " pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.391891 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-catalog-content\") pod \"community-operators-crhfr\" (UID: \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\") " pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.391934 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7xsr5\" (UniqueName: \"kubernetes.io/projected/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-kube-api-access-7xsr5\") pod \"community-operators-crhfr\" (UID: \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\") " pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.392060 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-utilities\") pod \"community-operators-crhfr\" (UID: \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\") " pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.392788 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-catalog-content\") pod \"community-operators-crhfr\" (UID: \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\") " pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.434764 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xsr5\" (UniqueName: \"kubernetes.io/projected/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-kube-api-access-7xsr5\") pod \"community-operators-crhfr\" (UID: \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\") " pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:39 crc kubenswrapper[4775]: I1002 03:31:39.549006 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:40 crc kubenswrapper[4775]: I1002 03:31:40.029186 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-crhfr"] Oct 02 03:31:40 crc kubenswrapper[4775]: W1002 03:31:40.036617 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b2d2dd3_8907_4f79_b3a8_78797278ddfd.slice/crio-72d5af72329036b7398b41b229a0bc477f010a1b64f3f8b3c106a6a18efb7573 WatchSource:0}: Error finding container 72d5af72329036b7398b41b229a0bc477f010a1b64f3f8b3c106a6a18efb7573: Status 404 returned error can't find the container with id 72d5af72329036b7398b41b229a0bc477f010a1b64f3f8b3c106a6a18efb7573 Oct 02 03:31:40 crc kubenswrapper[4775]: I1002 03:31:40.440589 4775 generic.go:334] "Generic (PLEG): container finished" podID="8b2d2dd3-8907-4f79-b3a8-78797278ddfd" containerID="8492378e496a06061eb599dd2263dfb8c144f4392ffbf68fdd074fd00d20a040" exitCode=0 Oct 02 03:31:40 crc kubenswrapper[4775]: I1002 03:31:40.440737 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crhfr" event={"ID":"8b2d2dd3-8907-4f79-b3a8-78797278ddfd","Type":"ContainerDied","Data":"8492378e496a06061eb599dd2263dfb8c144f4392ffbf68fdd074fd00d20a040"} Oct 02 03:31:40 crc kubenswrapper[4775]: I1002 03:31:40.440905 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crhfr" event={"ID":"8b2d2dd3-8907-4f79-b3a8-78797278ddfd","Type":"ContainerStarted","Data":"72d5af72329036b7398b41b229a0bc477f010a1b64f3f8b3c106a6a18efb7573"} Oct 02 03:31:40 crc kubenswrapper[4775]: I1002 03:31:40.444071 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 03:31:42 crc kubenswrapper[4775]: I1002 03:31:42.498893 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crhfr" event={"ID":"8b2d2dd3-8907-4f79-b3a8-78797278ddfd","Type":"ContainerStarted","Data":"61667108bdb507ee02e79bee8cb8ef4c3c8faba827192e8a9d616db5891cd25d"} Oct 02 03:31:43 crc kubenswrapper[4775]: I1002 03:31:43.511510 4775 generic.go:334] "Generic (PLEG): container finished" podID="8b2d2dd3-8907-4f79-b3a8-78797278ddfd" containerID="61667108bdb507ee02e79bee8cb8ef4c3c8faba827192e8a9d616db5891cd25d" exitCode=0 Oct 02 03:31:43 crc kubenswrapper[4775]: I1002 03:31:43.511557 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crhfr" event={"ID":"8b2d2dd3-8907-4f79-b3a8-78797278ddfd","Type":"ContainerDied","Data":"61667108bdb507ee02e79bee8cb8ef4c3c8faba827192e8a9d616db5891cd25d"} Oct 02 03:31:44 crc kubenswrapper[4775]: I1002 03:31:44.528790 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crhfr" event={"ID":"8b2d2dd3-8907-4f79-b3a8-78797278ddfd","Type":"ContainerStarted","Data":"3271be86c3ac0af202ee038ced04896637ad51510c4ab9dc82b2f7884784f8e1"} Oct 02 03:31:44 crc kubenswrapper[4775]: I1002 03:31:44.561578 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-crhfr" podStartSLOduration=2.060194752 podStartE2EDuration="5.561550683s" podCreationTimestamp="2025-10-02 03:31:39 +0000 UTC" firstStartedPulling="2025-10-02 03:31:40.443388627 +0000 UTC m=+6637.610132717" lastFinishedPulling="2025-10-02 03:31:43.944744578 +0000 UTC m=+6641.111488648" observedRunningTime="2025-10-02 03:31:44.555873916 +0000 UTC m=+6641.722617996" watchObservedRunningTime="2025-10-02 03:31:44.561550683 +0000 UTC m=+6641.728294733" Oct 02 03:31:47 crc kubenswrapper[4775]: I1002 03:31:47.767124 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:31:47 crc kubenswrapper[4775]: E1002 03:31:47.768122 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:31:49 crc kubenswrapper[4775]: I1002 03:31:49.549845 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:49 crc kubenswrapper[4775]: I1002 03:31:49.550284 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:50 crc kubenswrapper[4775]: I1002 03:31:50.625873 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-crhfr" podUID="8b2d2dd3-8907-4f79-b3a8-78797278ddfd" containerName="registry-server" probeResult="failure" output=< Oct 02 03:31:50 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 03:31:50 crc kubenswrapper[4775]: > Oct 02 03:31:59 crc kubenswrapper[4775]: I1002 03:31:59.628292 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:59 crc kubenswrapper[4775]: I1002 03:31:59.698755 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:31:59 crc kubenswrapper[4775]: I1002 03:31:59.765448 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:31:59 crc kubenswrapper[4775]: E1002 03:31:59.766137 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:31:59 crc kubenswrapper[4775]: I1002 03:31:59.875938 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-crhfr"] Oct 02 03:32:00 crc kubenswrapper[4775]: I1002 03:32:00.745194 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-crhfr" podUID="8b2d2dd3-8907-4f79-b3a8-78797278ddfd" containerName="registry-server" containerID="cri-o://3271be86c3ac0af202ee038ced04896637ad51510c4ab9dc82b2f7884784f8e1" gracePeriod=2 Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.299702 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.386861 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-utilities\") pod \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\" (UID: \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\") " Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.387503 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-catalog-content\") pod \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\" (UID: \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\") " Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.387574 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xsr5\" (UniqueName: \"kubernetes.io/projected/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-kube-api-access-7xsr5\") pod \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\" (UID: \"8b2d2dd3-8907-4f79-b3a8-78797278ddfd\") " Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.387743 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-utilities" (OuterVolumeSpecName: "utilities") pod "8b2d2dd3-8907-4f79-b3a8-78797278ddfd" (UID: "8b2d2dd3-8907-4f79-b3a8-78797278ddfd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.388532 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.395300 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-kube-api-access-7xsr5" (OuterVolumeSpecName: "kube-api-access-7xsr5") pod "8b2d2dd3-8907-4f79-b3a8-78797278ddfd" (UID: "8b2d2dd3-8907-4f79-b3a8-78797278ddfd"). InnerVolumeSpecName "kube-api-access-7xsr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.447798 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b2d2dd3-8907-4f79-b3a8-78797278ddfd" (UID: "8b2d2dd3-8907-4f79-b3a8-78797278ddfd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.490963 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.491007 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7xsr5\" (UniqueName: \"kubernetes.io/projected/8b2d2dd3-8907-4f79-b3a8-78797278ddfd-kube-api-access-7xsr5\") on node \"crc\" DevicePath \"\"" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.755429 4775 generic.go:334] "Generic (PLEG): container finished" podID="8b2d2dd3-8907-4f79-b3a8-78797278ddfd" containerID="3271be86c3ac0af202ee038ced04896637ad51510c4ab9dc82b2f7884784f8e1" exitCode=0 Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.755479 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crhfr" event={"ID":"8b2d2dd3-8907-4f79-b3a8-78797278ddfd","Type":"ContainerDied","Data":"3271be86c3ac0af202ee038ced04896637ad51510c4ab9dc82b2f7884784f8e1"} Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.755493 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-crhfr" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.755512 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-crhfr" event={"ID":"8b2d2dd3-8907-4f79-b3a8-78797278ddfd","Type":"ContainerDied","Data":"72d5af72329036b7398b41b229a0bc477f010a1b64f3f8b3c106a6a18efb7573"} Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.755533 4775 scope.go:117] "RemoveContainer" containerID="3271be86c3ac0af202ee038ced04896637ad51510c4ab9dc82b2f7884784f8e1" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.796196 4775 scope.go:117] "RemoveContainer" containerID="61667108bdb507ee02e79bee8cb8ef4c3c8faba827192e8a9d616db5891cd25d" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.799894 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-crhfr"] Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.809594 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-crhfr"] Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.818086 4775 scope.go:117] "RemoveContainer" containerID="8492378e496a06061eb599dd2263dfb8c144f4392ffbf68fdd074fd00d20a040" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.887427 4775 scope.go:117] "RemoveContainer" containerID="3271be86c3ac0af202ee038ced04896637ad51510c4ab9dc82b2f7884784f8e1" Oct 02 03:32:01 crc kubenswrapper[4775]: E1002 03:32:01.888108 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3271be86c3ac0af202ee038ced04896637ad51510c4ab9dc82b2f7884784f8e1\": container with ID starting with 3271be86c3ac0af202ee038ced04896637ad51510c4ab9dc82b2f7884784f8e1 not found: ID does not exist" containerID="3271be86c3ac0af202ee038ced04896637ad51510c4ab9dc82b2f7884784f8e1" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.888176 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3271be86c3ac0af202ee038ced04896637ad51510c4ab9dc82b2f7884784f8e1"} err="failed to get container status \"3271be86c3ac0af202ee038ced04896637ad51510c4ab9dc82b2f7884784f8e1\": rpc error: code = NotFound desc = could not find container \"3271be86c3ac0af202ee038ced04896637ad51510c4ab9dc82b2f7884784f8e1\": container with ID starting with 3271be86c3ac0af202ee038ced04896637ad51510c4ab9dc82b2f7884784f8e1 not found: ID does not exist" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.888218 4775 scope.go:117] "RemoveContainer" containerID="61667108bdb507ee02e79bee8cb8ef4c3c8faba827192e8a9d616db5891cd25d" Oct 02 03:32:01 crc kubenswrapper[4775]: E1002 03:32:01.888869 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61667108bdb507ee02e79bee8cb8ef4c3c8faba827192e8a9d616db5891cd25d\": container with ID starting with 61667108bdb507ee02e79bee8cb8ef4c3c8faba827192e8a9d616db5891cd25d not found: ID does not exist" containerID="61667108bdb507ee02e79bee8cb8ef4c3c8faba827192e8a9d616db5891cd25d" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.888927 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61667108bdb507ee02e79bee8cb8ef4c3c8faba827192e8a9d616db5891cd25d"} err="failed to get container status \"61667108bdb507ee02e79bee8cb8ef4c3c8faba827192e8a9d616db5891cd25d\": rpc error: code = NotFound desc = could not find container \"61667108bdb507ee02e79bee8cb8ef4c3c8faba827192e8a9d616db5891cd25d\": container with ID starting with 61667108bdb507ee02e79bee8cb8ef4c3c8faba827192e8a9d616db5891cd25d not found: ID does not exist" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.888984 4775 scope.go:117] "RemoveContainer" containerID="8492378e496a06061eb599dd2263dfb8c144f4392ffbf68fdd074fd00d20a040" Oct 02 03:32:01 crc kubenswrapper[4775]: E1002 03:32:01.889402 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8492378e496a06061eb599dd2263dfb8c144f4392ffbf68fdd074fd00d20a040\": container with ID starting with 8492378e496a06061eb599dd2263dfb8c144f4392ffbf68fdd074fd00d20a040 not found: ID does not exist" containerID="8492378e496a06061eb599dd2263dfb8c144f4392ffbf68fdd074fd00d20a040" Oct 02 03:32:01 crc kubenswrapper[4775]: I1002 03:32:01.889460 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8492378e496a06061eb599dd2263dfb8c144f4392ffbf68fdd074fd00d20a040"} err="failed to get container status \"8492378e496a06061eb599dd2263dfb8c144f4392ffbf68fdd074fd00d20a040\": rpc error: code = NotFound desc = could not find container \"8492378e496a06061eb599dd2263dfb8c144f4392ffbf68fdd074fd00d20a040\": container with ID starting with 8492378e496a06061eb599dd2263dfb8c144f4392ffbf68fdd074fd00d20a040 not found: ID does not exist" Oct 02 03:32:03 crc kubenswrapper[4775]: I1002 03:32:03.787478 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b2d2dd3-8907-4f79-b3a8-78797278ddfd" path="/var/lib/kubelet/pods/8b2d2dd3-8907-4f79-b3a8-78797278ddfd/volumes" Oct 02 03:32:11 crc kubenswrapper[4775]: I1002 03:32:11.766043 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:32:12 crc kubenswrapper[4775]: I1002 03:32:12.918163 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"b0facf1c02deffa95f9278c1759d009fd20890fb25c08a88c5815d3264bcb42f"} Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.223473 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5j5zf"] Oct 02 03:32:38 crc kubenswrapper[4775]: E1002 03:32:38.226057 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b2d2dd3-8907-4f79-b3a8-78797278ddfd" containerName="extract-content" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.226095 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b2d2dd3-8907-4f79-b3a8-78797278ddfd" containerName="extract-content" Oct 02 03:32:38 crc kubenswrapper[4775]: E1002 03:32:38.226196 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b2d2dd3-8907-4f79-b3a8-78797278ddfd" containerName="registry-server" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.226216 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b2d2dd3-8907-4f79-b3a8-78797278ddfd" containerName="registry-server" Oct 02 03:32:38 crc kubenswrapper[4775]: E1002 03:32:38.226272 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b2d2dd3-8907-4f79-b3a8-78797278ddfd" containerName="extract-utilities" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.226293 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b2d2dd3-8907-4f79-b3a8-78797278ddfd" containerName="extract-utilities" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.226802 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b2d2dd3-8907-4f79-b3a8-78797278ddfd" containerName="registry-server" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.231051 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.253720 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5j5zf"] Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.340817 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mglrr\" (UniqueName: \"kubernetes.io/projected/08a4849a-219f-4e45-831e-63f54912ebfd-kube-api-access-mglrr\") pod \"certified-operators-5j5zf\" (UID: \"08a4849a-219f-4e45-831e-63f54912ebfd\") " pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.341178 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08a4849a-219f-4e45-831e-63f54912ebfd-catalog-content\") pod \"certified-operators-5j5zf\" (UID: \"08a4849a-219f-4e45-831e-63f54912ebfd\") " pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.341262 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08a4849a-219f-4e45-831e-63f54912ebfd-utilities\") pod \"certified-operators-5j5zf\" (UID: \"08a4849a-219f-4e45-831e-63f54912ebfd\") " pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.443057 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08a4849a-219f-4e45-831e-63f54912ebfd-catalog-content\") pod \"certified-operators-5j5zf\" (UID: \"08a4849a-219f-4e45-831e-63f54912ebfd\") " pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.443128 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08a4849a-219f-4e45-831e-63f54912ebfd-utilities\") pod \"certified-operators-5j5zf\" (UID: \"08a4849a-219f-4e45-831e-63f54912ebfd\") " pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.443308 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mglrr\" (UniqueName: \"kubernetes.io/projected/08a4849a-219f-4e45-831e-63f54912ebfd-kube-api-access-mglrr\") pod \"certified-operators-5j5zf\" (UID: \"08a4849a-219f-4e45-831e-63f54912ebfd\") " pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.443980 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08a4849a-219f-4e45-831e-63f54912ebfd-catalog-content\") pod \"certified-operators-5j5zf\" (UID: \"08a4849a-219f-4e45-831e-63f54912ebfd\") " pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.444085 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08a4849a-219f-4e45-831e-63f54912ebfd-utilities\") pod \"certified-operators-5j5zf\" (UID: \"08a4849a-219f-4e45-831e-63f54912ebfd\") " pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.474430 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mglrr\" (UniqueName: \"kubernetes.io/projected/08a4849a-219f-4e45-831e-63f54912ebfd-kube-api-access-mglrr\") pod \"certified-operators-5j5zf\" (UID: \"08a4849a-219f-4e45-831e-63f54912ebfd\") " pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:38 crc kubenswrapper[4775]: I1002 03:32:38.572715 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:39 crc kubenswrapper[4775]: I1002 03:32:39.077339 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5j5zf"] Oct 02 03:32:39 crc kubenswrapper[4775]: I1002 03:32:39.290392 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j5zf" event={"ID":"08a4849a-219f-4e45-831e-63f54912ebfd","Type":"ContainerStarted","Data":"2ce7257bfe8ab5054b5c9a54ad241c732d0fb7d6e074e2180bda63e53ce41064"} Oct 02 03:32:40 crc kubenswrapper[4775]: I1002 03:32:40.302937 4775 generic.go:334] "Generic (PLEG): container finished" podID="08a4849a-219f-4e45-831e-63f54912ebfd" containerID="fcf159ac85c3d0c5d58a99bf756f9e913bc57e90c512e84a791af83ff17d7701" exitCode=0 Oct 02 03:32:40 crc kubenswrapper[4775]: I1002 03:32:40.303004 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j5zf" event={"ID":"08a4849a-219f-4e45-831e-63f54912ebfd","Type":"ContainerDied","Data":"fcf159ac85c3d0c5d58a99bf756f9e913bc57e90c512e84a791af83ff17d7701"} Oct 02 03:32:42 crc kubenswrapper[4775]: I1002 03:32:42.330134 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j5zf" event={"ID":"08a4849a-219f-4e45-831e-63f54912ebfd","Type":"ContainerStarted","Data":"dc5912dd0b15a7d3d891584ac0e587540bd6966636e227504ffc41ff1e271d2a"} Oct 02 03:32:43 crc kubenswrapper[4775]: I1002 03:32:43.346151 4775 generic.go:334] "Generic (PLEG): container finished" podID="08a4849a-219f-4e45-831e-63f54912ebfd" containerID="dc5912dd0b15a7d3d891584ac0e587540bd6966636e227504ffc41ff1e271d2a" exitCode=0 Oct 02 03:32:43 crc kubenswrapper[4775]: I1002 03:32:43.346211 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j5zf" event={"ID":"08a4849a-219f-4e45-831e-63f54912ebfd","Type":"ContainerDied","Data":"dc5912dd0b15a7d3d891584ac0e587540bd6966636e227504ffc41ff1e271d2a"} Oct 02 03:32:44 crc kubenswrapper[4775]: I1002 03:32:44.365117 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j5zf" event={"ID":"08a4849a-219f-4e45-831e-63f54912ebfd","Type":"ContainerStarted","Data":"1df1e0ed4fa4fab3c887d10c9a0f005ca88699e86efcdd4994871f1e90ebdb34"} Oct 02 03:32:44 crc kubenswrapper[4775]: I1002 03:32:44.399056 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5j5zf" podStartSLOduration=2.837808378 podStartE2EDuration="6.399026629s" podCreationTimestamp="2025-10-02 03:32:38 +0000 UTC" firstStartedPulling="2025-10-02 03:32:40.30507367 +0000 UTC m=+6697.471817710" lastFinishedPulling="2025-10-02 03:32:43.866291921 +0000 UTC m=+6701.033035961" observedRunningTime="2025-10-02 03:32:44.386473954 +0000 UTC m=+6701.553218024" watchObservedRunningTime="2025-10-02 03:32:44.399026629 +0000 UTC m=+6701.565770699" Oct 02 03:32:45 crc kubenswrapper[4775]: I1002 03:32:45.057701 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-ckn24"] Oct 02 03:32:45 crc kubenswrapper[4775]: I1002 03:32:45.069439 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-ckn24"] Oct 02 03:32:45 crc kubenswrapper[4775]: I1002 03:32:45.776409 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c9ebe75-71af-415f-b74a-d24d0e677411" path="/var/lib/kubelet/pods/2c9ebe75-71af-415f-b74a-d24d0e677411/volumes" Oct 02 03:32:48 crc kubenswrapper[4775]: I1002 03:32:48.573173 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:48 crc kubenswrapper[4775]: I1002 03:32:48.574113 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:48 crc kubenswrapper[4775]: I1002 03:32:48.640639 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:49 crc kubenswrapper[4775]: I1002 03:32:49.529840 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:49 crc kubenswrapper[4775]: I1002 03:32:49.598351 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5j5zf"] Oct 02 03:32:51 crc kubenswrapper[4775]: I1002 03:32:51.452340 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5j5zf" podUID="08a4849a-219f-4e45-831e-63f54912ebfd" containerName="registry-server" containerID="cri-o://1df1e0ed4fa4fab3c887d10c9a0f005ca88699e86efcdd4994871f1e90ebdb34" gracePeriod=2 Oct 02 03:32:52 crc kubenswrapper[4775]: E1002 03:32:52.043523 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08a4849a_219f_4e45_831e_63f54912ebfd.slice/crio-conmon-1df1e0ed4fa4fab3c887d10c9a0f005ca88699e86efcdd4994871f1e90ebdb34.scope\": RecentStats: unable to find data in memory cache]" Oct 02 03:32:52 crc kubenswrapper[4775]: I1002 03:32:52.488938 4775 generic.go:334] "Generic (PLEG): container finished" podID="08a4849a-219f-4e45-831e-63f54912ebfd" containerID="1df1e0ed4fa4fab3c887d10c9a0f005ca88699e86efcdd4994871f1e90ebdb34" exitCode=0 Oct 02 03:32:52 crc kubenswrapper[4775]: I1002 03:32:52.489066 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j5zf" event={"ID":"08a4849a-219f-4e45-831e-63f54912ebfd","Type":"ContainerDied","Data":"1df1e0ed4fa4fab3c887d10c9a0f005ca88699e86efcdd4994871f1e90ebdb34"} Oct 02 03:32:52 crc kubenswrapper[4775]: I1002 03:32:52.594065 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:52 crc kubenswrapper[4775]: I1002 03:32:52.684259 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08a4849a-219f-4e45-831e-63f54912ebfd-utilities\") pod \"08a4849a-219f-4e45-831e-63f54912ebfd\" (UID: \"08a4849a-219f-4e45-831e-63f54912ebfd\") " Oct 02 03:32:52 crc kubenswrapper[4775]: I1002 03:32:52.684537 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08a4849a-219f-4e45-831e-63f54912ebfd-catalog-content\") pod \"08a4849a-219f-4e45-831e-63f54912ebfd\" (UID: \"08a4849a-219f-4e45-831e-63f54912ebfd\") " Oct 02 03:32:52 crc kubenswrapper[4775]: I1002 03:32:52.684616 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mglrr\" (UniqueName: \"kubernetes.io/projected/08a4849a-219f-4e45-831e-63f54912ebfd-kube-api-access-mglrr\") pod \"08a4849a-219f-4e45-831e-63f54912ebfd\" (UID: \"08a4849a-219f-4e45-831e-63f54912ebfd\") " Oct 02 03:32:52 crc kubenswrapper[4775]: I1002 03:32:52.685845 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08a4849a-219f-4e45-831e-63f54912ebfd-utilities" (OuterVolumeSpecName: "utilities") pod "08a4849a-219f-4e45-831e-63f54912ebfd" (UID: "08a4849a-219f-4e45-831e-63f54912ebfd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:32:52 crc kubenswrapper[4775]: I1002 03:32:52.690432 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08a4849a-219f-4e45-831e-63f54912ebfd-kube-api-access-mglrr" (OuterVolumeSpecName: "kube-api-access-mglrr") pod "08a4849a-219f-4e45-831e-63f54912ebfd" (UID: "08a4849a-219f-4e45-831e-63f54912ebfd"). InnerVolumeSpecName "kube-api-access-mglrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:32:52 crc kubenswrapper[4775]: I1002 03:32:52.737976 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08a4849a-219f-4e45-831e-63f54912ebfd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08a4849a-219f-4e45-831e-63f54912ebfd" (UID: "08a4849a-219f-4e45-831e-63f54912ebfd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:32:52 crc kubenswrapper[4775]: I1002 03:32:52.786562 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08a4849a-219f-4e45-831e-63f54912ebfd-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:32:52 crc kubenswrapper[4775]: I1002 03:32:52.786595 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08a4849a-219f-4e45-831e-63f54912ebfd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:32:52 crc kubenswrapper[4775]: I1002 03:32:52.786606 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mglrr\" (UniqueName: \"kubernetes.io/projected/08a4849a-219f-4e45-831e-63f54912ebfd-kube-api-access-mglrr\") on node \"crc\" DevicePath \"\"" Oct 02 03:32:53 crc kubenswrapper[4775]: I1002 03:32:53.509033 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5j5zf" event={"ID":"08a4849a-219f-4e45-831e-63f54912ebfd","Type":"ContainerDied","Data":"2ce7257bfe8ab5054b5c9a54ad241c732d0fb7d6e074e2180bda63e53ce41064"} Oct 02 03:32:53 crc kubenswrapper[4775]: I1002 03:32:53.509125 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5j5zf" Oct 02 03:32:53 crc kubenswrapper[4775]: I1002 03:32:53.509563 4775 scope.go:117] "RemoveContainer" containerID="1df1e0ed4fa4fab3c887d10c9a0f005ca88699e86efcdd4994871f1e90ebdb34" Oct 02 03:32:53 crc kubenswrapper[4775]: I1002 03:32:53.556344 4775 scope.go:117] "RemoveContainer" containerID="dc5912dd0b15a7d3d891584ac0e587540bd6966636e227504ffc41ff1e271d2a" Oct 02 03:32:53 crc kubenswrapper[4775]: I1002 03:32:53.573425 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5j5zf"] Oct 02 03:32:53 crc kubenswrapper[4775]: I1002 03:32:53.590565 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5j5zf"] Oct 02 03:32:53 crc kubenswrapper[4775]: I1002 03:32:53.599896 4775 scope.go:117] "RemoveContainer" containerID="fcf159ac85c3d0c5d58a99bf756f9e913bc57e90c512e84a791af83ff17d7701" Oct 02 03:32:53 crc kubenswrapper[4775]: I1002 03:32:53.785657 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08a4849a-219f-4e45-831e-63f54912ebfd" path="/var/lib/kubelet/pods/08a4849a-219f-4e45-831e-63f54912ebfd/volumes" Oct 02 03:32:56 crc kubenswrapper[4775]: I1002 03:32:56.109783 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-53fd-account-create-px9tb"] Oct 02 03:32:56 crc kubenswrapper[4775]: I1002 03:32:56.122117 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-53fd-account-create-px9tb"] Oct 02 03:32:57 crc kubenswrapper[4775]: I1002 03:32:57.784191 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="394faa8e-649e-4dc0-8b2c-ed1e9523e241" path="/var/lib/kubelet/pods/394faa8e-649e-4dc0-8b2c-ed1e9523e241/volumes" Oct 02 03:33:09 crc kubenswrapper[4775]: I1002 03:33:09.058836 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-25qfl"] Oct 02 03:33:09 crc kubenswrapper[4775]: I1002 03:33:09.076193 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-25qfl"] Oct 02 03:33:09 crc kubenswrapper[4775]: I1002 03:33:09.778925 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bc90542-eaa4-4cb7-b041-8222fef25ceb" path="/var/lib/kubelet/pods/7bc90542-eaa4-4cb7-b041-8222fef25ceb/volumes" Oct 02 03:33:16 crc kubenswrapper[4775]: I1002 03:33:16.138518 4775 scope.go:117] "RemoveContainer" containerID="9098f3167b6e483dcd46e2cfcb81813b0005e374075fe2ffdfd2c1c1d0d5dfd9" Oct 02 03:33:16 crc kubenswrapper[4775]: I1002 03:33:16.183228 4775 scope.go:117] "RemoveContainer" containerID="48c43a6bd7ba4e609dec4579a08cf76cedf58641bac90e753d4e5bb4bb3abce8" Oct 02 03:33:16 crc kubenswrapper[4775]: I1002 03:33:16.300896 4775 scope.go:117] "RemoveContainer" containerID="c007d3ba64951b37481cc0a06ad17d7b0b7cf1181c9e44a84424b285a67ecdf0" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.287283 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qlgzm"] Oct 02 03:33:20 crc kubenswrapper[4775]: E1002 03:33:20.288897 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08a4849a-219f-4e45-831e-63f54912ebfd" containerName="extract-content" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.288920 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="08a4849a-219f-4e45-831e-63f54912ebfd" containerName="extract-content" Oct 02 03:33:20 crc kubenswrapper[4775]: E1002 03:33:20.288988 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08a4849a-219f-4e45-831e-63f54912ebfd" containerName="registry-server" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.289002 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="08a4849a-219f-4e45-831e-63f54912ebfd" containerName="registry-server" Oct 02 03:33:20 crc kubenswrapper[4775]: E1002 03:33:20.289106 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08a4849a-219f-4e45-831e-63f54912ebfd" containerName="extract-utilities" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.289121 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="08a4849a-219f-4e45-831e-63f54912ebfd" containerName="extract-utilities" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.289711 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="08a4849a-219f-4e45-831e-63f54912ebfd" containerName="registry-server" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.294337 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.305693 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qlgzm"] Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.485503 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-catalog-content\") pod \"redhat-operators-qlgzm\" (UID: \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\") " pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.485627 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpdlz\" (UniqueName: \"kubernetes.io/projected/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-kube-api-access-mpdlz\") pod \"redhat-operators-qlgzm\" (UID: \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\") " pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.485721 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-utilities\") pod \"redhat-operators-qlgzm\" (UID: \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\") " pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.587204 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpdlz\" (UniqueName: \"kubernetes.io/projected/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-kube-api-access-mpdlz\") pod \"redhat-operators-qlgzm\" (UID: \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\") " pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.587324 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-utilities\") pod \"redhat-operators-qlgzm\" (UID: \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\") " pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.587567 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-catalog-content\") pod \"redhat-operators-qlgzm\" (UID: \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\") " pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.588070 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-utilities\") pod \"redhat-operators-qlgzm\" (UID: \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\") " pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.588187 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-catalog-content\") pod \"redhat-operators-qlgzm\" (UID: \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\") " pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.619776 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpdlz\" (UniqueName: \"kubernetes.io/projected/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-kube-api-access-mpdlz\") pod \"redhat-operators-qlgzm\" (UID: \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\") " pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:33:20 crc kubenswrapper[4775]: I1002 03:33:20.649471 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:33:21 crc kubenswrapper[4775]: W1002 03:33:21.166507 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9afc6fa_f61b_4c8a_8f03_c8b65664662b.slice/crio-3526b79ba0ff0054b07f1b0b09bc2bd0e66f293a98dc8e782578f9972126101d WatchSource:0}: Error finding container 3526b79ba0ff0054b07f1b0b09bc2bd0e66f293a98dc8e782578f9972126101d: Status 404 returned error can't find the container with id 3526b79ba0ff0054b07f1b0b09bc2bd0e66f293a98dc8e782578f9972126101d Oct 02 03:33:21 crc kubenswrapper[4775]: I1002 03:33:21.173437 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qlgzm"] Oct 02 03:33:21 crc kubenswrapper[4775]: I1002 03:33:21.882097 4775 generic.go:334] "Generic (PLEG): container finished" podID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" containerID="a65b38cb228e1449fca36626ebc4139c04783db246bb863c01838ab33da06918" exitCode=0 Oct 02 03:33:21 crc kubenswrapper[4775]: I1002 03:33:21.882177 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qlgzm" event={"ID":"f9afc6fa-f61b-4c8a-8f03-c8b65664662b","Type":"ContainerDied","Data":"a65b38cb228e1449fca36626ebc4139c04783db246bb863c01838ab33da06918"} Oct 02 03:33:21 crc kubenswrapper[4775]: I1002 03:33:21.882414 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qlgzm" event={"ID":"f9afc6fa-f61b-4c8a-8f03-c8b65664662b","Type":"ContainerStarted","Data":"3526b79ba0ff0054b07f1b0b09bc2bd0e66f293a98dc8e782578f9972126101d"} Oct 02 03:33:23 crc kubenswrapper[4775]: I1002 03:33:23.908765 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qlgzm" event={"ID":"f9afc6fa-f61b-4c8a-8f03-c8b65664662b","Type":"ContainerStarted","Data":"8227be6daac3ed4c507b5910eef3a174339849fd748a871a3d580c38a2b9b493"} Oct 02 03:33:27 crc kubenswrapper[4775]: I1002 03:33:27.958762 4775 generic.go:334] "Generic (PLEG): container finished" podID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" containerID="8227be6daac3ed4c507b5910eef3a174339849fd748a871a3d580c38a2b9b493" exitCode=0 Oct 02 03:33:27 crc kubenswrapper[4775]: I1002 03:33:27.958885 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qlgzm" event={"ID":"f9afc6fa-f61b-4c8a-8f03-c8b65664662b","Type":"ContainerDied","Data":"8227be6daac3ed4c507b5910eef3a174339849fd748a871a3d580c38a2b9b493"} Oct 02 03:33:28 crc kubenswrapper[4775]: I1002 03:33:28.975523 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qlgzm" event={"ID":"f9afc6fa-f61b-4c8a-8f03-c8b65664662b","Type":"ContainerStarted","Data":"c7094e38bdcd0ed70afec44fb09d8fe3f78e83b5f48c95c4b9a5f747a9d79f15"} Oct 02 03:33:29 crc kubenswrapper[4775]: I1002 03:33:29.003829 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qlgzm" podStartSLOduration=2.498605067 podStartE2EDuration="9.003803504s" podCreationTimestamp="2025-10-02 03:33:20 +0000 UTC" firstStartedPulling="2025-10-02 03:33:21.885189801 +0000 UTC m=+6739.051933871" lastFinishedPulling="2025-10-02 03:33:28.390388258 +0000 UTC m=+6745.557132308" observedRunningTime="2025-10-02 03:33:29.000030936 +0000 UTC m=+6746.166774976" watchObservedRunningTime="2025-10-02 03:33:29.003803504 +0000 UTC m=+6746.170547584" Oct 02 03:33:30 crc kubenswrapper[4775]: I1002 03:33:30.651013 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:33:30 crc kubenswrapper[4775]: I1002 03:33:30.651291 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:33:31 crc kubenswrapper[4775]: I1002 03:33:31.708983 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qlgzm" podUID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" containerName="registry-server" probeResult="failure" output=< Oct 02 03:33:31 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 03:33:31 crc kubenswrapper[4775]: > Oct 02 03:33:41 crc kubenswrapper[4775]: I1002 03:33:41.706457 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qlgzm" podUID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" containerName="registry-server" probeResult="failure" output=< Oct 02 03:33:41 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 03:33:41 crc kubenswrapper[4775]: > Oct 02 03:33:51 crc kubenswrapper[4775]: I1002 03:33:51.698055 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qlgzm" podUID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" containerName="registry-server" probeResult="failure" output=< Oct 02 03:33:51 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 03:33:51 crc kubenswrapper[4775]: > Oct 02 03:34:00 crc kubenswrapper[4775]: I1002 03:34:00.739107 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:34:00 crc kubenswrapper[4775]: I1002 03:34:00.834407 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:34:00 crc kubenswrapper[4775]: I1002 03:34:00.989997 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qlgzm"] Oct 02 03:34:02 crc kubenswrapper[4775]: I1002 03:34:02.366119 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qlgzm" podUID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" containerName="registry-server" containerID="cri-o://c7094e38bdcd0ed70afec44fb09d8fe3f78e83b5f48c95c4b9a5f747a9d79f15" gracePeriod=2 Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.066342 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.245792 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-utilities\") pod \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\" (UID: \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\") " Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.245865 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpdlz\" (UniqueName: \"kubernetes.io/projected/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-kube-api-access-mpdlz\") pod \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\" (UID: \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\") " Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.246130 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-catalog-content\") pod \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\" (UID: \"f9afc6fa-f61b-4c8a-8f03-c8b65664662b\") " Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.247916 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-utilities" (OuterVolumeSpecName: "utilities") pod "f9afc6fa-f61b-4c8a-8f03-c8b65664662b" (UID: "f9afc6fa-f61b-4c8a-8f03-c8b65664662b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.256172 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-kube-api-access-mpdlz" (OuterVolumeSpecName: "kube-api-access-mpdlz") pod "f9afc6fa-f61b-4c8a-8f03-c8b65664662b" (UID: "f9afc6fa-f61b-4c8a-8f03-c8b65664662b"). InnerVolumeSpecName "kube-api-access-mpdlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.335081 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f9afc6fa-f61b-4c8a-8f03-c8b65664662b" (UID: "f9afc6fa-f61b-4c8a-8f03-c8b65664662b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.349833 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.349915 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.349935 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpdlz\" (UniqueName: \"kubernetes.io/projected/f9afc6fa-f61b-4c8a-8f03-c8b65664662b-kube-api-access-mpdlz\") on node \"crc\" DevicePath \"\"" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.380889 4775 generic.go:334] "Generic (PLEG): container finished" podID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" containerID="c7094e38bdcd0ed70afec44fb09d8fe3f78e83b5f48c95c4b9a5f747a9d79f15" exitCode=0 Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.380994 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qlgzm" event={"ID":"f9afc6fa-f61b-4c8a-8f03-c8b65664662b","Type":"ContainerDied","Data":"c7094e38bdcd0ed70afec44fb09d8fe3f78e83b5f48c95c4b9a5f747a9d79f15"} Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.381037 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qlgzm" event={"ID":"f9afc6fa-f61b-4c8a-8f03-c8b65664662b","Type":"ContainerDied","Data":"3526b79ba0ff0054b07f1b0b09bc2bd0e66f293a98dc8e782578f9972126101d"} Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.381067 4775 scope.go:117] "RemoveContainer" containerID="c7094e38bdcd0ed70afec44fb09d8fe3f78e83b5f48c95c4b9a5f747a9d79f15" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.381277 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qlgzm" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.428556 4775 scope.go:117] "RemoveContainer" containerID="8227be6daac3ed4c507b5910eef3a174339849fd748a871a3d580c38a2b9b493" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.437104 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qlgzm"] Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.462651 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qlgzm"] Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.467724 4775 scope.go:117] "RemoveContainer" containerID="a65b38cb228e1449fca36626ebc4139c04783db246bb863c01838ab33da06918" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.512037 4775 scope.go:117] "RemoveContainer" containerID="c7094e38bdcd0ed70afec44fb09d8fe3f78e83b5f48c95c4b9a5f747a9d79f15" Oct 02 03:34:03 crc kubenswrapper[4775]: E1002 03:34:03.512713 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7094e38bdcd0ed70afec44fb09d8fe3f78e83b5f48c95c4b9a5f747a9d79f15\": container with ID starting with c7094e38bdcd0ed70afec44fb09d8fe3f78e83b5f48c95c4b9a5f747a9d79f15 not found: ID does not exist" containerID="c7094e38bdcd0ed70afec44fb09d8fe3f78e83b5f48c95c4b9a5f747a9d79f15" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.512750 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7094e38bdcd0ed70afec44fb09d8fe3f78e83b5f48c95c4b9a5f747a9d79f15"} err="failed to get container status \"c7094e38bdcd0ed70afec44fb09d8fe3f78e83b5f48c95c4b9a5f747a9d79f15\": rpc error: code = NotFound desc = could not find container \"c7094e38bdcd0ed70afec44fb09d8fe3f78e83b5f48c95c4b9a5f747a9d79f15\": container with ID starting with c7094e38bdcd0ed70afec44fb09d8fe3f78e83b5f48c95c4b9a5f747a9d79f15 not found: ID does not exist" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.512776 4775 scope.go:117] "RemoveContainer" containerID="8227be6daac3ed4c507b5910eef3a174339849fd748a871a3d580c38a2b9b493" Oct 02 03:34:03 crc kubenswrapper[4775]: E1002 03:34:03.513268 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8227be6daac3ed4c507b5910eef3a174339849fd748a871a3d580c38a2b9b493\": container with ID starting with 8227be6daac3ed4c507b5910eef3a174339849fd748a871a3d580c38a2b9b493 not found: ID does not exist" containerID="8227be6daac3ed4c507b5910eef3a174339849fd748a871a3d580c38a2b9b493" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.513341 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8227be6daac3ed4c507b5910eef3a174339849fd748a871a3d580c38a2b9b493"} err="failed to get container status \"8227be6daac3ed4c507b5910eef3a174339849fd748a871a3d580c38a2b9b493\": rpc error: code = NotFound desc = could not find container \"8227be6daac3ed4c507b5910eef3a174339849fd748a871a3d580c38a2b9b493\": container with ID starting with 8227be6daac3ed4c507b5910eef3a174339849fd748a871a3d580c38a2b9b493 not found: ID does not exist" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.513410 4775 scope.go:117] "RemoveContainer" containerID="a65b38cb228e1449fca36626ebc4139c04783db246bb863c01838ab33da06918" Oct 02 03:34:03 crc kubenswrapper[4775]: E1002 03:34:03.513986 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a65b38cb228e1449fca36626ebc4139c04783db246bb863c01838ab33da06918\": container with ID starting with a65b38cb228e1449fca36626ebc4139c04783db246bb863c01838ab33da06918 not found: ID does not exist" containerID="a65b38cb228e1449fca36626ebc4139c04783db246bb863c01838ab33da06918" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.514048 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a65b38cb228e1449fca36626ebc4139c04783db246bb863c01838ab33da06918"} err="failed to get container status \"a65b38cb228e1449fca36626ebc4139c04783db246bb863c01838ab33da06918\": rpc error: code = NotFound desc = could not find container \"a65b38cb228e1449fca36626ebc4139c04783db246bb863c01838ab33da06918\": container with ID starting with a65b38cb228e1449fca36626ebc4139c04783db246bb863c01838ab33da06918 not found: ID does not exist" Oct 02 03:34:03 crc kubenswrapper[4775]: I1002 03:34:03.784749 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" path="/var/lib/kubelet/pods/f9afc6fa-f61b-4c8a-8f03-c8b65664662b/volumes" Oct 02 03:34:37 crc kubenswrapper[4775]: I1002 03:34:37.234243 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:34:37 crc kubenswrapper[4775]: I1002 03:34:37.234998 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:35:07 crc kubenswrapper[4775]: I1002 03:35:07.234181 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:35:07 crc kubenswrapper[4775]: I1002 03:35:07.235283 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:35:13 crc kubenswrapper[4775]: I1002 03:35:13.058233 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-tl697"] Oct 02 03:35:13 crc kubenswrapper[4775]: I1002 03:35:13.073601 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-tl697"] Oct 02 03:35:13 crc kubenswrapper[4775]: I1002 03:35:13.791428 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42bb3fa6-b8c1-478f-9b7e-7986a17501c3" path="/var/lib/kubelet/pods/42bb3fa6-b8c1-478f-9b7e-7986a17501c3/volumes" Oct 02 03:35:16 crc kubenswrapper[4775]: I1002 03:35:16.505630 4775 scope.go:117] "RemoveContainer" containerID="52680f1826704472e47edf9cbbb4bf827264949fa8bce2bc4fea413c313edc5c" Oct 02 03:35:23 crc kubenswrapper[4775]: I1002 03:35:23.057437 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0319-account-create-fcjrh"] Oct 02 03:35:23 crc kubenswrapper[4775]: I1002 03:35:23.068195 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0319-account-create-fcjrh"] Oct 02 03:35:23 crc kubenswrapper[4775]: I1002 03:35:23.787725 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2006decc-1f50-44fe-ad50-b3dd82beb654" path="/var/lib/kubelet/pods/2006decc-1f50-44fe-ad50-b3dd82beb654/volumes" Oct 02 03:35:36 crc kubenswrapper[4775]: I1002 03:35:36.039663 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-j97tn"] Oct 02 03:35:36 crc kubenswrapper[4775]: I1002 03:35:36.050050 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-j97tn"] Oct 02 03:35:37 crc kubenswrapper[4775]: I1002 03:35:37.233532 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:35:37 crc kubenswrapper[4775]: I1002 03:35:37.233618 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:35:37 crc kubenswrapper[4775]: I1002 03:35:37.233674 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 03:35:37 crc kubenswrapper[4775]: I1002 03:35:37.234948 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b0facf1c02deffa95f9278c1759d009fd20890fb25c08a88c5815d3264bcb42f"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 03:35:37 crc kubenswrapper[4775]: I1002 03:35:37.235275 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://b0facf1c02deffa95f9278c1759d009fd20890fb25c08a88c5815d3264bcb42f" gracePeriod=600 Oct 02 03:35:37 crc kubenswrapper[4775]: I1002 03:35:37.578694 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="b0facf1c02deffa95f9278c1759d009fd20890fb25c08a88c5815d3264bcb42f" exitCode=0 Oct 02 03:35:37 crc kubenswrapper[4775]: I1002 03:35:37.578826 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"b0facf1c02deffa95f9278c1759d009fd20890fb25c08a88c5815d3264bcb42f"} Oct 02 03:35:37 crc kubenswrapper[4775]: I1002 03:35:37.579426 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0"} Oct 02 03:35:37 crc kubenswrapper[4775]: I1002 03:35:37.579465 4775 scope.go:117] "RemoveContainer" containerID="a9b433e49b56c0b5c305bb849993a08bc6baac8219386c490bc0bebdb9bba769" Oct 02 03:35:37 crc kubenswrapper[4775]: I1002 03:35:37.784737 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="903ed25a-6e5c-46c9-8aac-04c071cc2067" path="/var/lib/kubelet/pods/903ed25a-6e5c-46c9-8aac-04c071cc2067/volumes" Oct 02 03:35:56 crc kubenswrapper[4775]: I1002 03:35:56.058930 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-fjmh2"] Oct 02 03:35:56 crc kubenswrapper[4775]: I1002 03:35:56.074756 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-fjmh2"] Oct 02 03:35:57 crc kubenswrapper[4775]: I1002 03:35:57.778889 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e7ec537-ac4e-4f99-b9d0-e5758f145445" path="/var/lib/kubelet/pods/5e7ec537-ac4e-4f99-b9d0-e5758f145445/volumes" Oct 02 03:36:06 crc kubenswrapper[4775]: I1002 03:36:06.044384 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-1375-account-create-hln6z"] Oct 02 03:36:06 crc kubenswrapper[4775]: I1002 03:36:06.056245 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-1375-account-create-hln6z"] Oct 02 03:36:07 crc kubenswrapper[4775]: I1002 03:36:07.788526 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf" path="/var/lib/kubelet/pods/e4f1a84b-a72b-4380-9bf6-3ba00d10cbdf/volumes" Oct 02 03:36:16 crc kubenswrapper[4775]: I1002 03:36:16.609815 4775 scope.go:117] "RemoveContainer" containerID="903bbca96494b81e50dbce334a55e19d2a4174f186fd33c486b7b166908a930f" Oct 02 03:36:16 crc kubenswrapper[4775]: I1002 03:36:16.676483 4775 scope.go:117] "RemoveContainer" containerID="85d10a99025652d3b33c7723f4114227c286c946ea888f1de64ff5f5218c2717" Oct 02 03:36:16 crc kubenswrapper[4775]: I1002 03:36:16.742802 4775 scope.go:117] "RemoveContainer" containerID="e4b66c8cb8a64de4a941e5d2473508aa6b43c17777fe3600d699c1bcbafd5880" Oct 02 03:36:16 crc kubenswrapper[4775]: I1002 03:36:16.778990 4775 scope.go:117] "RemoveContainer" containerID="e4cd09bc36b03426ae6bbcca6fc8d9574b4c0345d2367a3b930032ff24f4c3b8" Oct 02 03:36:18 crc kubenswrapper[4775]: I1002 03:36:18.034305 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-xlbkk"] Oct 02 03:36:18 crc kubenswrapper[4775]: I1002 03:36:18.039993 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-xlbkk"] Oct 02 03:36:19 crc kubenswrapper[4775]: I1002 03:36:19.782115 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0987ddf-aa6c-4f61-9fcd-42370ad9a697" path="/var/lib/kubelet/pods/b0987ddf-aa6c-4f61-9fcd-42370ad9a697/volumes" Oct 02 03:37:16 crc kubenswrapper[4775]: I1002 03:37:16.943466 4775 scope.go:117] "RemoveContainer" containerID="3e49c5e9b5524d8baf15da9a9f6ef660fb6a545a205679265737adee43c43f76" Oct 02 03:37:37 crc kubenswrapper[4775]: I1002 03:37:37.233524 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:37:37 crc kubenswrapper[4775]: I1002 03:37:37.235091 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:38:07 crc kubenswrapper[4775]: I1002 03:38:07.233320 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:38:07 crc kubenswrapper[4775]: I1002 03:38:07.234155 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:38:37 crc kubenswrapper[4775]: I1002 03:38:37.234349 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:38:37 crc kubenswrapper[4775]: I1002 03:38:37.235225 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:38:37 crc kubenswrapper[4775]: I1002 03:38:37.235302 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 03:38:37 crc kubenswrapper[4775]: I1002 03:38:37.236615 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 03:38:37 crc kubenswrapper[4775]: I1002 03:38:37.237733 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" gracePeriod=600 Oct 02 03:38:37 crc kubenswrapper[4775]: E1002 03:38:37.434548 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:38:37 crc kubenswrapper[4775]: I1002 03:38:37.903522 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" exitCode=0 Oct 02 03:38:37 crc kubenswrapper[4775]: I1002 03:38:37.903577 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0"} Oct 02 03:38:37 crc kubenswrapper[4775]: I1002 03:38:37.903617 4775 scope.go:117] "RemoveContainer" containerID="b0facf1c02deffa95f9278c1759d009fd20890fb25c08a88c5815d3264bcb42f" Oct 02 03:38:37 crc kubenswrapper[4775]: I1002 03:38:37.905330 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:38:37 crc kubenswrapper[4775]: E1002 03:38:37.906413 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:38:52 crc kubenswrapper[4775]: I1002 03:38:52.766213 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:38:52 crc kubenswrapper[4775]: E1002 03:38:52.767514 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:39:05 crc kubenswrapper[4775]: I1002 03:39:05.766848 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:39:05 crc kubenswrapper[4775]: E1002 03:39:05.769933 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:39:16 crc kubenswrapper[4775]: I1002 03:39:16.766449 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:39:16 crc kubenswrapper[4775]: E1002 03:39:16.767692 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:39:30 crc kubenswrapper[4775]: I1002 03:39:30.766311 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:39:30 crc kubenswrapper[4775]: E1002 03:39:30.767417 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:39:45 crc kubenswrapper[4775]: I1002 03:39:45.766938 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:39:45 crc kubenswrapper[4775]: E1002 03:39:45.767890 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:39:59 crc kubenswrapper[4775]: I1002 03:39:59.767562 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:39:59 crc kubenswrapper[4775]: E1002 03:39:59.768732 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:40:14 crc kubenswrapper[4775]: I1002 03:40:14.766279 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:40:14 crc kubenswrapper[4775]: E1002 03:40:14.768232 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:40:26 crc kubenswrapper[4775]: I1002 03:40:26.766882 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:40:26 crc kubenswrapper[4775]: E1002 03:40:26.768219 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:40:38 crc kubenswrapper[4775]: I1002 03:40:38.765252 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:40:38 crc kubenswrapper[4775]: E1002 03:40:38.765945 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:40:48 crc kubenswrapper[4775]: I1002 03:40:48.494728 4775 generic.go:334] "Generic (PLEG): container finished" podID="3959dab0-5e51-4cb7-a540-96ae7c7a28b9" containerID="7747f9a3c676ae1eb0c97b3d9d7209af2dd7ca4986e4721c56050fea17e6a762" exitCode=0 Oct 02 03:40:48 crc kubenswrapper[4775]: I1002 03:40:48.495336 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" event={"ID":"3959dab0-5e51-4cb7-a540-96ae7c7a28b9","Type":"ContainerDied","Data":"7747f9a3c676ae1eb0c97b3d9d7209af2dd7ca4986e4721c56050fea17e6a762"} Oct 02 03:40:48 crc kubenswrapper[4775]: I1002 03:40:48.772478 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-spmnd"] Oct 02 03:40:48 crc kubenswrapper[4775]: E1002 03:40:48.772937 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" containerName="extract-content" Oct 02 03:40:48 crc kubenswrapper[4775]: I1002 03:40:48.773038 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" containerName="extract-content" Oct 02 03:40:48 crc kubenswrapper[4775]: E1002 03:40:48.773073 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" containerName="registry-server" Oct 02 03:40:48 crc kubenswrapper[4775]: I1002 03:40:48.773083 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" containerName="registry-server" Oct 02 03:40:48 crc kubenswrapper[4775]: E1002 03:40:48.773095 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" containerName="extract-utilities" Oct 02 03:40:48 crc kubenswrapper[4775]: I1002 03:40:48.773104 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" containerName="extract-utilities" Oct 02 03:40:48 crc kubenswrapper[4775]: I1002 03:40:48.773384 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9afc6fa-f61b-4c8a-8f03-c8b65664662b" containerName="registry-server" Oct 02 03:40:48 crc kubenswrapper[4775]: I1002 03:40:48.775380 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:48 crc kubenswrapper[4775]: I1002 03:40:48.788606 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-spmnd"] Oct 02 03:40:48 crc kubenswrapper[4775]: I1002 03:40:48.951306 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6afff7-1988-4403-9fcf-9705120e1a43-catalog-content\") pod \"redhat-marketplace-spmnd\" (UID: \"ba6afff7-1988-4403-9fcf-9705120e1a43\") " pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:48 crc kubenswrapper[4775]: I1002 03:40:48.951355 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f22p7\" (UniqueName: \"kubernetes.io/projected/ba6afff7-1988-4403-9fcf-9705120e1a43-kube-api-access-f22p7\") pod \"redhat-marketplace-spmnd\" (UID: \"ba6afff7-1988-4403-9fcf-9705120e1a43\") " pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:48 crc kubenswrapper[4775]: I1002 03:40:48.951534 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6afff7-1988-4403-9fcf-9705120e1a43-utilities\") pod \"redhat-marketplace-spmnd\" (UID: \"ba6afff7-1988-4403-9fcf-9705120e1a43\") " pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:49 crc kubenswrapper[4775]: I1002 03:40:49.053131 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6afff7-1988-4403-9fcf-9705120e1a43-catalog-content\") pod \"redhat-marketplace-spmnd\" (UID: \"ba6afff7-1988-4403-9fcf-9705120e1a43\") " pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:49 crc kubenswrapper[4775]: I1002 03:40:49.053185 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f22p7\" (UniqueName: \"kubernetes.io/projected/ba6afff7-1988-4403-9fcf-9705120e1a43-kube-api-access-f22p7\") pod \"redhat-marketplace-spmnd\" (UID: \"ba6afff7-1988-4403-9fcf-9705120e1a43\") " pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:49 crc kubenswrapper[4775]: I1002 03:40:49.053338 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6afff7-1988-4403-9fcf-9705120e1a43-utilities\") pod \"redhat-marketplace-spmnd\" (UID: \"ba6afff7-1988-4403-9fcf-9705120e1a43\") " pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:49 crc kubenswrapper[4775]: I1002 03:40:49.053890 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6afff7-1988-4403-9fcf-9705120e1a43-utilities\") pod \"redhat-marketplace-spmnd\" (UID: \"ba6afff7-1988-4403-9fcf-9705120e1a43\") " pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:49 crc kubenswrapper[4775]: I1002 03:40:49.054186 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6afff7-1988-4403-9fcf-9705120e1a43-catalog-content\") pod \"redhat-marketplace-spmnd\" (UID: \"ba6afff7-1988-4403-9fcf-9705120e1a43\") " pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:49 crc kubenswrapper[4775]: I1002 03:40:49.088756 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f22p7\" (UniqueName: \"kubernetes.io/projected/ba6afff7-1988-4403-9fcf-9705120e1a43-kube-api-access-f22p7\") pod \"redhat-marketplace-spmnd\" (UID: \"ba6afff7-1988-4403-9fcf-9705120e1a43\") " pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:49 crc kubenswrapper[4775]: I1002 03:40:49.141343 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:49 crc kubenswrapper[4775]: I1002 03:40:49.675405 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-spmnd"] Oct 02 03:40:49 crc kubenswrapper[4775]: I1002 03:40:49.955715 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.076931 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-ceph\") pod \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.077002 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m96fp\" (UniqueName: \"kubernetes.io/projected/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-kube-api-access-m96fp\") pod \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.077137 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-ssh-key\") pod \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.077181 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-tripleo-cleanup-combined-ca-bundle\") pod \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.077216 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-inventory\") pod \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\" (UID: \"3959dab0-5e51-4cb7-a540-96ae7c7a28b9\") " Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.082751 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-ceph" (OuterVolumeSpecName: "ceph") pod "3959dab0-5e51-4cb7-a540-96ae7c7a28b9" (UID: "3959dab0-5e51-4cb7-a540-96ae7c7a28b9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.083271 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "3959dab0-5e51-4cb7-a540-96ae7c7a28b9" (UID: "3959dab0-5e51-4cb7-a540-96ae7c7a28b9"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.083931 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-kube-api-access-m96fp" (OuterVolumeSpecName: "kube-api-access-m96fp") pod "3959dab0-5e51-4cb7-a540-96ae7c7a28b9" (UID: "3959dab0-5e51-4cb7-a540-96ae7c7a28b9"). InnerVolumeSpecName "kube-api-access-m96fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.105789 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3959dab0-5e51-4cb7-a540-96ae7c7a28b9" (UID: "3959dab0-5e51-4cb7-a540-96ae7c7a28b9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.127834 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-inventory" (OuterVolumeSpecName: "inventory") pod "3959dab0-5e51-4cb7-a540-96ae7c7a28b9" (UID: "3959dab0-5e51-4cb7-a540-96ae7c7a28b9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.180529 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.180564 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m96fp\" (UniqueName: \"kubernetes.io/projected/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-kube-api-access-m96fp\") on node \"crc\" DevicePath \"\"" Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.180575 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.180584 4775 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.180594 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3959dab0-5e51-4cb7-a540-96ae7c7a28b9-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.518037 4775 generic.go:334] "Generic (PLEG): container finished" podID="ba6afff7-1988-4403-9fcf-9705120e1a43" containerID="95908addeb0156aff505db56e4d81d9faeec25a43e4031e16bd61adaf827f539" exitCode=0 Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.518140 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-spmnd" event={"ID":"ba6afff7-1988-4403-9fcf-9705120e1a43","Type":"ContainerDied","Data":"95908addeb0156aff505db56e4d81d9faeec25a43e4031e16bd61adaf827f539"} Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.518427 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-spmnd" event={"ID":"ba6afff7-1988-4403-9fcf-9705120e1a43","Type":"ContainerStarted","Data":"f52449c21c3e93e3480d6510083e1390ee5cab873c6c9cba9a01526d182c23ed"} Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.521339 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" event={"ID":"3959dab0-5e51-4cb7-a540-96ae7c7a28b9","Type":"ContainerDied","Data":"146139aee9b93023ca907ac1413f1961f88bc23385ea78adfa041e4ad4fdf75f"} Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.521377 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="146139aee9b93023ca907ac1413f1961f88bc23385ea78adfa041e4ad4fdf75f" Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.521456 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m" Oct 02 03:40:50 crc kubenswrapper[4775]: I1002 03:40:50.522999 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.472999 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-rldqh"] Oct 02 03:40:51 crc kubenswrapper[4775]: E1002 03:40:51.474035 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3959dab0-5e51-4cb7-a540-96ae7c7a28b9" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.474057 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3959dab0-5e51-4cb7-a540-96ae7c7a28b9" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.474603 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3959dab0-5e51-4cb7-a540-96ae7c7a28b9" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.478239 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.481701 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.481862 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.482402 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.483901 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.494496 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-rldqh"] Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.517967 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-inventory\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.518017 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-ceph\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.518110 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrqtk\" (UniqueName: \"kubernetes.io/projected/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-kube-api-access-nrqtk\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.518165 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.518197 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.534883 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-spmnd" event={"ID":"ba6afff7-1988-4403-9fcf-9705120e1a43","Type":"ContainerStarted","Data":"7c4691d90e5e9b1e8dba0250a7f767e17bd2bd918d73b68350e7d621410eaa15"} Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.621937 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrqtk\" (UniqueName: \"kubernetes.io/projected/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-kube-api-access-nrqtk\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.622031 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.622065 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.622213 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-inventory\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.622240 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-ceph\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.628929 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.632833 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-ceph\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.637303 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.643099 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrqtk\" (UniqueName: \"kubernetes.io/projected/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-kube-api-access-nrqtk\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.648239 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-inventory\") pod \"bootstrap-openstack-openstack-cell1-rldqh\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:51 crc kubenswrapper[4775]: I1002 03:40:51.801246 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:40:52 crc kubenswrapper[4775]: I1002 03:40:52.359442 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-rldqh"] Oct 02 03:40:52 crc kubenswrapper[4775]: I1002 03:40:52.551073 4775 generic.go:334] "Generic (PLEG): container finished" podID="ba6afff7-1988-4403-9fcf-9705120e1a43" containerID="7c4691d90e5e9b1e8dba0250a7f767e17bd2bd918d73b68350e7d621410eaa15" exitCode=0 Oct 02 03:40:52 crc kubenswrapper[4775]: I1002 03:40:52.551137 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-spmnd" event={"ID":"ba6afff7-1988-4403-9fcf-9705120e1a43","Type":"ContainerDied","Data":"7c4691d90e5e9b1e8dba0250a7f767e17bd2bd918d73b68350e7d621410eaa15"} Oct 02 03:40:52 crc kubenswrapper[4775]: I1002 03:40:52.554200 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" event={"ID":"5032d81f-b10d-4bf0-89f1-a6ee328ceaec","Type":"ContainerStarted","Data":"7bd11041aab6e73b6deab2305cb6461460e5889cf5ba289f68825eadbe9e6fed"} Oct 02 03:40:53 crc kubenswrapper[4775]: I1002 03:40:53.564550 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" event={"ID":"5032d81f-b10d-4bf0-89f1-a6ee328ceaec","Type":"ContainerStarted","Data":"cc770e3aefa525af5f92132e6a150235cbb6881c64531140f56f6bae32991c2d"} Oct 02 03:40:53 crc kubenswrapper[4775]: I1002 03:40:53.568150 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-spmnd" event={"ID":"ba6afff7-1988-4403-9fcf-9705120e1a43","Type":"ContainerStarted","Data":"f47a1f96bdea4604932d13145ea661386c623a0adc9e5dedc7f72010e86d4996"} Oct 02 03:40:53 crc kubenswrapper[4775]: I1002 03:40:53.590084 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" podStartSLOduration=1.955933404 podStartE2EDuration="2.590063555s" podCreationTimestamp="2025-10-02 03:40:51 +0000 UTC" firstStartedPulling="2025-10-02 03:40:52.377208885 +0000 UTC m=+7189.543952935" lastFinishedPulling="2025-10-02 03:40:53.011339006 +0000 UTC m=+7190.178083086" observedRunningTime="2025-10-02 03:40:53.589423498 +0000 UTC m=+7190.756167538" watchObservedRunningTime="2025-10-02 03:40:53.590063555 +0000 UTC m=+7190.756807605" Oct 02 03:40:53 crc kubenswrapper[4775]: I1002 03:40:53.619595 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-spmnd" podStartSLOduration=2.999750193 podStartE2EDuration="5.619568289s" podCreationTimestamp="2025-10-02 03:40:48 +0000 UTC" firstStartedPulling="2025-10-02 03:40:50.522438099 +0000 UTC m=+7187.689182169" lastFinishedPulling="2025-10-02 03:40:53.142256225 +0000 UTC m=+7190.309000265" observedRunningTime="2025-10-02 03:40:53.604719044 +0000 UTC m=+7190.771463094" watchObservedRunningTime="2025-10-02 03:40:53.619568289 +0000 UTC m=+7190.786312359" Oct 02 03:40:53 crc kubenswrapper[4775]: I1002 03:40:53.774005 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:40:53 crc kubenswrapper[4775]: E1002 03:40:53.774258 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:40:59 crc kubenswrapper[4775]: I1002 03:40:59.142723 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:59 crc kubenswrapper[4775]: I1002 03:40:59.144643 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:59 crc kubenswrapper[4775]: I1002 03:40:59.237811 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:59 crc kubenswrapper[4775]: I1002 03:40:59.737861 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:40:59 crc kubenswrapper[4775]: I1002 03:40:59.800277 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-spmnd"] Oct 02 03:41:01 crc kubenswrapper[4775]: I1002 03:41:01.674980 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-spmnd" podUID="ba6afff7-1988-4403-9fcf-9705120e1a43" containerName="registry-server" containerID="cri-o://f47a1f96bdea4604932d13145ea661386c623a0adc9e5dedc7f72010e86d4996" gracePeriod=2 Oct 02 03:41:02 crc kubenswrapper[4775]: I1002 03:41:02.706934 4775 generic.go:334] "Generic (PLEG): container finished" podID="ba6afff7-1988-4403-9fcf-9705120e1a43" containerID="f47a1f96bdea4604932d13145ea661386c623a0adc9e5dedc7f72010e86d4996" exitCode=0 Oct 02 03:41:02 crc kubenswrapper[4775]: I1002 03:41:02.707232 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-spmnd" event={"ID":"ba6afff7-1988-4403-9fcf-9705120e1a43","Type":"ContainerDied","Data":"f47a1f96bdea4604932d13145ea661386c623a0adc9e5dedc7f72010e86d4996"} Oct 02 03:41:08 crc kubenswrapper[4775]: I1002 03:41:08.766586 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:41:08 crc kubenswrapper[4775]: E1002 03:41:08.767483 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:41:09 crc kubenswrapper[4775]: E1002 03:41:09.142930 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f47a1f96bdea4604932d13145ea661386c623a0adc9e5dedc7f72010e86d4996 is running failed: container process not found" containerID="f47a1f96bdea4604932d13145ea661386c623a0adc9e5dedc7f72010e86d4996" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 03:41:09 crc kubenswrapper[4775]: E1002 03:41:09.143744 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f47a1f96bdea4604932d13145ea661386c623a0adc9e5dedc7f72010e86d4996 is running failed: container process not found" containerID="f47a1f96bdea4604932d13145ea661386c623a0adc9e5dedc7f72010e86d4996" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 03:41:09 crc kubenswrapper[4775]: E1002 03:41:09.144381 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f47a1f96bdea4604932d13145ea661386c623a0adc9e5dedc7f72010e86d4996 is running failed: container process not found" containerID="f47a1f96bdea4604932d13145ea661386c623a0adc9e5dedc7f72010e86d4996" cmd=["grpc_health_probe","-addr=:50051"] Oct 02 03:41:09 crc kubenswrapper[4775]: E1002 03:41:09.144429 4775 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f47a1f96bdea4604932d13145ea661386c623a0adc9e5dedc7f72010e86d4996 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-spmnd" podUID="ba6afff7-1988-4403-9fcf-9705120e1a43" containerName="registry-server" Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.448589 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.603374 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6afff7-1988-4403-9fcf-9705120e1a43-catalog-content\") pod \"ba6afff7-1988-4403-9fcf-9705120e1a43\" (UID: \"ba6afff7-1988-4403-9fcf-9705120e1a43\") " Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.603745 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6afff7-1988-4403-9fcf-9705120e1a43-utilities\") pod \"ba6afff7-1988-4403-9fcf-9705120e1a43\" (UID: \"ba6afff7-1988-4403-9fcf-9705120e1a43\") " Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.603777 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f22p7\" (UniqueName: \"kubernetes.io/projected/ba6afff7-1988-4403-9fcf-9705120e1a43-kube-api-access-f22p7\") pod \"ba6afff7-1988-4403-9fcf-9705120e1a43\" (UID: \"ba6afff7-1988-4403-9fcf-9705120e1a43\") " Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.604799 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba6afff7-1988-4403-9fcf-9705120e1a43-utilities" (OuterVolumeSpecName: "utilities") pod "ba6afff7-1988-4403-9fcf-9705120e1a43" (UID: "ba6afff7-1988-4403-9fcf-9705120e1a43"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.610339 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba6afff7-1988-4403-9fcf-9705120e1a43-kube-api-access-f22p7" (OuterVolumeSpecName: "kube-api-access-f22p7") pod "ba6afff7-1988-4403-9fcf-9705120e1a43" (UID: "ba6afff7-1988-4403-9fcf-9705120e1a43"). InnerVolumeSpecName "kube-api-access-f22p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.619258 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba6afff7-1988-4403-9fcf-9705120e1a43-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba6afff7-1988-4403-9fcf-9705120e1a43" (UID: "ba6afff7-1988-4403-9fcf-9705120e1a43"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.706918 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba6afff7-1988-4403-9fcf-9705120e1a43-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.706969 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba6afff7-1988-4403-9fcf-9705120e1a43-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.706985 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f22p7\" (UniqueName: \"kubernetes.io/projected/ba6afff7-1988-4403-9fcf-9705120e1a43-kube-api-access-f22p7\") on node \"crc\" DevicePath \"\"" Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.818141 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-spmnd" event={"ID":"ba6afff7-1988-4403-9fcf-9705120e1a43","Type":"ContainerDied","Data":"f52449c21c3e93e3480d6510083e1390ee5cab873c6c9cba9a01526d182c23ed"} Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.818205 4775 scope.go:117] "RemoveContainer" containerID="f47a1f96bdea4604932d13145ea661386c623a0adc9e5dedc7f72010e86d4996" Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.818290 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-spmnd" Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.856689 4775 scope.go:117] "RemoveContainer" containerID="7c4691d90e5e9b1e8dba0250a7f767e17bd2bd918d73b68350e7d621410eaa15" Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.887262 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-spmnd"] Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.904737 4775 scope.go:117] "RemoveContainer" containerID="95908addeb0156aff505db56e4d81d9faeec25a43e4031e16bd61adaf827f539" Oct 02 03:41:10 crc kubenswrapper[4775]: I1002 03:41:10.905184 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-spmnd"] Oct 02 03:41:11 crc kubenswrapper[4775]: I1002 03:41:11.783144 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba6afff7-1988-4403-9fcf-9705120e1a43" path="/var/lib/kubelet/pods/ba6afff7-1988-4403-9fcf-9705120e1a43/volumes" Oct 02 03:41:23 crc kubenswrapper[4775]: I1002 03:41:23.781176 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:41:23 crc kubenswrapper[4775]: E1002 03:41:23.782230 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:41:36 crc kubenswrapper[4775]: I1002 03:41:36.766109 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:41:36 crc kubenswrapper[4775]: E1002 03:41:36.766926 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:41:48 crc kubenswrapper[4775]: I1002 03:41:48.766209 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:41:48 crc kubenswrapper[4775]: E1002 03:41:48.767642 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:42:03 crc kubenswrapper[4775]: I1002 03:42:03.774875 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:42:03 crc kubenswrapper[4775]: E1002 03:42:03.776102 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:42:16 crc kubenswrapper[4775]: I1002 03:42:16.766877 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:42:16 crc kubenswrapper[4775]: E1002 03:42:16.768053 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:42:28 crc kubenswrapper[4775]: I1002 03:42:28.767561 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:42:28 crc kubenswrapper[4775]: E1002 03:42:28.768877 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:42:41 crc kubenswrapper[4775]: I1002 03:42:41.767095 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:42:41 crc kubenswrapper[4775]: E1002 03:42:41.768062 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:42:55 crc kubenswrapper[4775]: I1002 03:42:55.768433 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:42:55 crc kubenswrapper[4775]: E1002 03:42:55.787715 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.556112 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gdzlr"] Oct 02 03:42:56 crc kubenswrapper[4775]: E1002 03:42:56.557072 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba6afff7-1988-4403-9fcf-9705120e1a43" containerName="extract-utilities" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.557101 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6afff7-1988-4403-9fcf-9705120e1a43" containerName="extract-utilities" Oct 02 03:42:56 crc kubenswrapper[4775]: E1002 03:42:56.557160 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba6afff7-1988-4403-9fcf-9705120e1a43" containerName="registry-server" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.557174 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6afff7-1988-4403-9fcf-9705120e1a43" containerName="registry-server" Oct 02 03:42:56 crc kubenswrapper[4775]: E1002 03:42:56.557215 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba6afff7-1988-4403-9fcf-9705120e1a43" containerName="extract-content" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.557230 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6afff7-1988-4403-9fcf-9705120e1a43" containerName="extract-content" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.557627 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba6afff7-1988-4403-9fcf-9705120e1a43" containerName="registry-server" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.561387 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.578756 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gdzlr"] Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.634378 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7frk\" (UniqueName: \"kubernetes.io/projected/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-kube-api-access-h7frk\") pod \"community-operators-gdzlr\" (UID: \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\") " pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.634678 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-catalog-content\") pod \"community-operators-gdzlr\" (UID: \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\") " pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.635495 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-utilities\") pod \"community-operators-gdzlr\" (UID: \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\") " pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.738541 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-utilities\") pod \"community-operators-gdzlr\" (UID: \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\") " pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.738637 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7frk\" (UniqueName: \"kubernetes.io/projected/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-kube-api-access-h7frk\") pod \"community-operators-gdzlr\" (UID: \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\") " pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.738686 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-catalog-content\") pod \"community-operators-gdzlr\" (UID: \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\") " pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.739096 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-utilities\") pod \"community-operators-gdzlr\" (UID: \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\") " pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.739151 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-catalog-content\") pod \"community-operators-gdzlr\" (UID: \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\") " pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.773112 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7frk\" (UniqueName: \"kubernetes.io/projected/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-kube-api-access-h7frk\") pod \"community-operators-gdzlr\" (UID: \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\") " pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:42:56 crc kubenswrapper[4775]: I1002 03:42:56.955612 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:42:57 crc kubenswrapper[4775]: I1002 03:42:57.482816 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gdzlr"] Oct 02 03:42:58 crc kubenswrapper[4775]: I1002 03:42:58.125939 4775 generic.go:334] "Generic (PLEG): container finished" podID="dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" containerID="49756d218b8bdff2a809d681adcf2b406910897ea2d92a1efc2a9d0da2939da1" exitCode=0 Oct 02 03:42:58 crc kubenswrapper[4775]: I1002 03:42:58.126011 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdzlr" event={"ID":"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26","Type":"ContainerDied","Data":"49756d218b8bdff2a809d681adcf2b406910897ea2d92a1efc2a9d0da2939da1"} Oct 02 03:42:58 crc kubenswrapper[4775]: I1002 03:42:58.126043 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdzlr" event={"ID":"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26","Type":"ContainerStarted","Data":"7376e58d6c050233cfe1021c5ec9aa0c72972a1c8f6aa3b223001d0a86c8ce1d"} Oct 02 03:42:59 crc kubenswrapper[4775]: I1002 03:42:59.143269 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdzlr" event={"ID":"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26","Type":"ContainerStarted","Data":"97cd9e090ac7bd2d7af3f452b07404a9a68eddc9b9a15009983ec93286c7ae8a"} Oct 02 03:43:01 crc kubenswrapper[4775]: I1002 03:43:01.171769 4775 generic.go:334] "Generic (PLEG): container finished" podID="dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" containerID="97cd9e090ac7bd2d7af3f452b07404a9a68eddc9b9a15009983ec93286c7ae8a" exitCode=0 Oct 02 03:43:01 crc kubenswrapper[4775]: I1002 03:43:01.171861 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdzlr" event={"ID":"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26","Type":"ContainerDied","Data":"97cd9e090ac7bd2d7af3f452b07404a9a68eddc9b9a15009983ec93286c7ae8a"} Oct 02 03:43:02 crc kubenswrapper[4775]: I1002 03:43:02.186329 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdzlr" event={"ID":"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26","Type":"ContainerStarted","Data":"a12b3eb71b1e6494fd9f7e4dfa33ab30e869878bf9d0d785cdc528bbb202fc14"} Oct 02 03:43:02 crc kubenswrapper[4775]: I1002 03:43:02.215398 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gdzlr" podStartSLOduration=2.761445458 podStartE2EDuration="6.215377082s" podCreationTimestamp="2025-10-02 03:42:56 +0000 UTC" firstStartedPulling="2025-10-02 03:42:58.127766758 +0000 UTC m=+7315.294510808" lastFinishedPulling="2025-10-02 03:43:01.581698392 +0000 UTC m=+7318.748442432" observedRunningTime="2025-10-02 03:43:02.210931657 +0000 UTC m=+7319.377675727" watchObservedRunningTime="2025-10-02 03:43:02.215377082 +0000 UTC m=+7319.382121132" Oct 02 03:43:06 crc kubenswrapper[4775]: I1002 03:43:06.956912 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:43:06 crc kubenswrapper[4775]: I1002 03:43:06.957539 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:43:07 crc kubenswrapper[4775]: I1002 03:43:07.087310 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:43:07 crc kubenswrapper[4775]: I1002 03:43:07.313869 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:43:07 crc kubenswrapper[4775]: I1002 03:43:07.383563 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gdzlr"] Oct 02 03:43:07 crc kubenswrapper[4775]: I1002 03:43:07.766477 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:43:07 crc kubenswrapper[4775]: E1002 03:43:07.770452 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:43:09 crc kubenswrapper[4775]: I1002 03:43:09.277316 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gdzlr" podUID="dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" containerName="registry-server" containerID="cri-o://a12b3eb71b1e6494fd9f7e4dfa33ab30e869878bf9d0d785cdc528bbb202fc14" gracePeriod=2 Oct 02 03:43:09 crc kubenswrapper[4775]: I1002 03:43:09.875501 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:43:09 crc kubenswrapper[4775]: I1002 03:43:09.993234 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-catalog-content\") pod \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\" (UID: \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\") " Oct 02 03:43:09 crc kubenswrapper[4775]: I1002 03:43:09.993379 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-utilities\") pod \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\" (UID: \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\") " Oct 02 03:43:09 crc kubenswrapper[4775]: I1002 03:43:09.993445 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7frk\" (UniqueName: \"kubernetes.io/projected/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-kube-api-access-h7frk\") pod \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\" (UID: \"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26\") " Oct 02 03:43:09 crc kubenswrapper[4775]: I1002 03:43:09.994233 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-utilities" (OuterVolumeSpecName: "utilities") pod "dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" (UID: "dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.007276 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-kube-api-access-h7frk" (OuterVolumeSpecName: "kube-api-access-h7frk") pod "dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" (UID: "dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26"). InnerVolumeSpecName "kube-api-access-h7frk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.065216 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" (UID: "dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.096536 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.096576 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.096590 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7frk\" (UniqueName: \"kubernetes.io/projected/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26-kube-api-access-h7frk\") on node \"crc\" DevicePath \"\"" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.295066 4775 generic.go:334] "Generic (PLEG): container finished" podID="dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" containerID="a12b3eb71b1e6494fd9f7e4dfa33ab30e869878bf9d0d785cdc528bbb202fc14" exitCode=0 Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.295135 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdzlr" event={"ID":"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26","Type":"ContainerDied","Data":"a12b3eb71b1e6494fd9f7e4dfa33ab30e869878bf9d0d785cdc528bbb202fc14"} Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.295141 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gdzlr" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.295184 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gdzlr" event={"ID":"dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26","Type":"ContainerDied","Data":"7376e58d6c050233cfe1021c5ec9aa0c72972a1c8f6aa3b223001d0a86c8ce1d"} Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.295215 4775 scope.go:117] "RemoveContainer" containerID="a12b3eb71b1e6494fd9f7e4dfa33ab30e869878bf9d0d785cdc528bbb202fc14" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.319418 4775 scope.go:117] "RemoveContainer" containerID="97cd9e090ac7bd2d7af3f452b07404a9a68eddc9b9a15009983ec93286c7ae8a" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.333035 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gdzlr"] Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.343134 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gdzlr"] Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.360607 4775 scope.go:117] "RemoveContainer" containerID="49756d218b8bdff2a809d681adcf2b406910897ea2d92a1efc2a9d0da2939da1" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.403693 4775 scope.go:117] "RemoveContainer" containerID="a12b3eb71b1e6494fd9f7e4dfa33ab30e869878bf9d0d785cdc528bbb202fc14" Oct 02 03:43:10 crc kubenswrapper[4775]: E1002 03:43:10.404466 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a12b3eb71b1e6494fd9f7e4dfa33ab30e869878bf9d0d785cdc528bbb202fc14\": container with ID starting with a12b3eb71b1e6494fd9f7e4dfa33ab30e869878bf9d0d785cdc528bbb202fc14 not found: ID does not exist" containerID="a12b3eb71b1e6494fd9f7e4dfa33ab30e869878bf9d0d785cdc528bbb202fc14" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.404553 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a12b3eb71b1e6494fd9f7e4dfa33ab30e869878bf9d0d785cdc528bbb202fc14"} err="failed to get container status \"a12b3eb71b1e6494fd9f7e4dfa33ab30e869878bf9d0d785cdc528bbb202fc14\": rpc error: code = NotFound desc = could not find container \"a12b3eb71b1e6494fd9f7e4dfa33ab30e869878bf9d0d785cdc528bbb202fc14\": container with ID starting with a12b3eb71b1e6494fd9f7e4dfa33ab30e869878bf9d0d785cdc528bbb202fc14 not found: ID does not exist" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.404583 4775 scope.go:117] "RemoveContainer" containerID="97cd9e090ac7bd2d7af3f452b07404a9a68eddc9b9a15009983ec93286c7ae8a" Oct 02 03:43:10 crc kubenswrapper[4775]: E1002 03:43:10.404884 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97cd9e090ac7bd2d7af3f452b07404a9a68eddc9b9a15009983ec93286c7ae8a\": container with ID starting with 97cd9e090ac7bd2d7af3f452b07404a9a68eddc9b9a15009983ec93286c7ae8a not found: ID does not exist" containerID="97cd9e090ac7bd2d7af3f452b07404a9a68eddc9b9a15009983ec93286c7ae8a" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.404939 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97cd9e090ac7bd2d7af3f452b07404a9a68eddc9b9a15009983ec93286c7ae8a"} err="failed to get container status \"97cd9e090ac7bd2d7af3f452b07404a9a68eddc9b9a15009983ec93286c7ae8a\": rpc error: code = NotFound desc = could not find container \"97cd9e090ac7bd2d7af3f452b07404a9a68eddc9b9a15009983ec93286c7ae8a\": container with ID starting with 97cd9e090ac7bd2d7af3f452b07404a9a68eddc9b9a15009983ec93286c7ae8a not found: ID does not exist" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.405068 4775 scope.go:117] "RemoveContainer" containerID="49756d218b8bdff2a809d681adcf2b406910897ea2d92a1efc2a9d0da2939da1" Oct 02 03:43:10 crc kubenswrapper[4775]: E1002 03:43:10.405396 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49756d218b8bdff2a809d681adcf2b406910897ea2d92a1efc2a9d0da2939da1\": container with ID starting with 49756d218b8bdff2a809d681adcf2b406910897ea2d92a1efc2a9d0da2939da1 not found: ID does not exist" containerID="49756d218b8bdff2a809d681adcf2b406910897ea2d92a1efc2a9d0da2939da1" Oct 02 03:43:10 crc kubenswrapper[4775]: I1002 03:43:10.405425 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49756d218b8bdff2a809d681adcf2b406910897ea2d92a1efc2a9d0da2939da1"} err="failed to get container status \"49756d218b8bdff2a809d681adcf2b406910897ea2d92a1efc2a9d0da2939da1\": rpc error: code = NotFound desc = could not find container \"49756d218b8bdff2a809d681adcf2b406910897ea2d92a1efc2a9d0da2939da1\": container with ID starting with 49756d218b8bdff2a809d681adcf2b406910897ea2d92a1efc2a9d0da2939da1 not found: ID does not exist" Oct 02 03:43:11 crc kubenswrapper[4775]: I1002 03:43:11.778465 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" path="/var/lib/kubelet/pods/dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26/volumes" Oct 02 03:43:19 crc kubenswrapper[4775]: I1002 03:43:19.766683 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:43:19 crc kubenswrapper[4775]: E1002 03:43:19.767570 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:43:32 crc kubenswrapper[4775]: I1002 03:43:32.765564 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:43:32 crc kubenswrapper[4775]: E1002 03:43:32.766582 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:43:45 crc kubenswrapper[4775]: I1002 03:43:45.766369 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:43:46 crc kubenswrapper[4775]: I1002 03:43:46.786544 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"82e9bf60a688a73f4e67e12ca66bd36d7e6ce4ea01f01845c9edec604f41bbff"} Oct 02 03:44:03 crc kubenswrapper[4775]: I1002 03:44:03.992776 4775 generic.go:334] "Generic (PLEG): container finished" podID="5032d81f-b10d-4bf0-89f1-a6ee328ceaec" containerID="cc770e3aefa525af5f92132e6a150235cbb6881c64531140f56f6bae32991c2d" exitCode=0 Oct 02 03:44:03 crc kubenswrapper[4775]: I1002 03:44:03.993293 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" event={"ID":"5032d81f-b10d-4bf0-89f1-a6ee328ceaec","Type":"ContainerDied","Data":"cc770e3aefa525af5f92132e6a150235cbb6881c64531140f56f6bae32991c2d"} Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.596929 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.697870 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-ssh-key\") pod \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.698019 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-ceph\") pod \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.698065 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-bootstrap-combined-ca-bundle\") pod \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.698139 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrqtk\" (UniqueName: \"kubernetes.io/projected/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-kube-api-access-nrqtk\") pod \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.698211 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-inventory\") pod \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\" (UID: \"5032d81f-b10d-4bf0-89f1-a6ee328ceaec\") " Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.706149 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5032d81f-b10d-4bf0-89f1-a6ee328ceaec" (UID: "5032d81f-b10d-4bf0-89f1-a6ee328ceaec"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.706270 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-kube-api-access-nrqtk" (OuterVolumeSpecName: "kube-api-access-nrqtk") pod "5032d81f-b10d-4bf0-89f1-a6ee328ceaec" (UID: "5032d81f-b10d-4bf0-89f1-a6ee328ceaec"). InnerVolumeSpecName "kube-api-access-nrqtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.706430 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-ceph" (OuterVolumeSpecName: "ceph") pod "5032d81f-b10d-4bf0-89f1-a6ee328ceaec" (UID: "5032d81f-b10d-4bf0-89f1-a6ee328ceaec"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.737455 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5032d81f-b10d-4bf0-89f1-a6ee328ceaec" (UID: "5032d81f-b10d-4bf0-89f1-a6ee328ceaec"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.748926 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-inventory" (OuterVolumeSpecName: "inventory") pod "5032d81f-b10d-4bf0-89f1-a6ee328ceaec" (UID: "5032d81f-b10d-4bf0-89f1-a6ee328ceaec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.800340 4775 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.800375 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrqtk\" (UniqueName: \"kubernetes.io/projected/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-kube-api-access-nrqtk\") on node \"crc\" DevicePath \"\"" Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.800388 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.800401 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:44:05 crc kubenswrapper[4775]: I1002 03:44:05.800413 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5032d81f-b10d-4bf0-89f1-a6ee328ceaec-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.020621 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" event={"ID":"5032d81f-b10d-4bf0-89f1-a6ee328ceaec","Type":"ContainerDied","Data":"7bd11041aab6e73b6deab2305cb6461460e5889cf5ba289f68825eadbe9e6fed"} Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.020665 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bd11041aab6e73b6deab2305cb6461460e5889cf5ba289f68825eadbe9e6fed" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.020688 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-rldqh" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.128064 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-5mg6v"] Oct 02 03:44:06 crc kubenswrapper[4775]: E1002 03:44:06.128790 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5032d81f-b10d-4bf0-89f1-a6ee328ceaec" containerName="bootstrap-openstack-openstack-cell1" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.128810 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5032d81f-b10d-4bf0-89f1-a6ee328ceaec" containerName="bootstrap-openstack-openstack-cell1" Oct 02 03:44:06 crc kubenswrapper[4775]: E1002 03:44:06.129222 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" containerName="extract-content" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.129232 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" containerName="extract-content" Oct 02 03:44:06 crc kubenswrapper[4775]: E1002 03:44:06.129249 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" containerName="registry-server" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.129255 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" containerName="registry-server" Oct 02 03:44:06 crc kubenswrapper[4775]: E1002 03:44:06.129274 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" containerName="extract-utilities" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.129280 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" containerName="extract-utilities" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.129491 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5032d81f-b10d-4bf0-89f1-a6ee328ceaec" containerName="bootstrap-openstack-openstack-cell1" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.129510 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc0ab4b5-c8b4-43e8-8c28-d4f66d9b9d26" containerName="registry-server" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.130690 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.133311 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.133660 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.134468 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.139301 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.143978 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-5mg6v"] Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.311097 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-inventory\") pod \"download-cache-openstack-openstack-cell1-5mg6v\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.311309 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-ssh-key\") pod \"download-cache-openstack-openstack-cell1-5mg6v\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.311357 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-ceph\") pod \"download-cache-openstack-openstack-cell1-5mg6v\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.311452 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9gnc\" (UniqueName: \"kubernetes.io/projected/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-kube-api-access-m9gnc\") pod \"download-cache-openstack-openstack-cell1-5mg6v\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.414498 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-inventory\") pod \"download-cache-openstack-openstack-cell1-5mg6v\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.414585 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-ssh-key\") pod \"download-cache-openstack-openstack-cell1-5mg6v\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.414654 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-ceph\") pod \"download-cache-openstack-openstack-cell1-5mg6v\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.414755 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9gnc\" (UniqueName: \"kubernetes.io/projected/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-kube-api-access-m9gnc\") pod \"download-cache-openstack-openstack-cell1-5mg6v\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.419077 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-inventory\") pod \"download-cache-openstack-openstack-cell1-5mg6v\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.419424 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-ssh-key\") pod \"download-cache-openstack-openstack-cell1-5mg6v\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.420330 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-ceph\") pod \"download-cache-openstack-openstack-cell1-5mg6v\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.436644 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9gnc\" (UniqueName: \"kubernetes.io/projected/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-kube-api-access-m9gnc\") pod \"download-cache-openstack-openstack-cell1-5mg6v\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:44:06 crc kubenswrapper[4775]: I1002 03:44:06.452559 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:44:07 crc kubenswrapper[4775]: I1002 03:44:07.020635 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-5mg6v"] Oct 02 03:44:07 crc kubenswrapper[4775]: I1002 03:44:07.040210 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" event={"ID":"e15f9838-d5f3-4b6c-b968-de51d37ffaf9","Type":"ContainerStarted","Data":"9ee797e2cd2425f65e028d865efb8b49b2f97336d1aae5997a2f6c0b72dc8c72"} Oct 02 03:44:08 crc kubenswrapper[4775]: I1002 03:44:08.056175 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" event={"ID":"e15f9838-d5f3-4b6c-b968-de51d37ffaf9","Type":"ContainerStarted","Data":"fde99f848a13206969111a51ff4d46d89b750075f20eb229463e4a54ab713d5e"} Oct 02 03:44:08 crc kubenswrapper[4775]: I1002 03:44:08.091940 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" podStartSLOduration=1.558828994 podStartE2EDuration="2.091921751s" podCreationTimestamp="2025-10-02 03:44:06 +0000 UTC" firstStartedPulling="2025-10-02 03:44:07.024117844 +0000 UTC m=+7384.190861914" lastFinishedPulling="2025-10-02 03:44:07.557210591 +0000 UTC m=+7384.723954671" observedRunningTime="2025-10-02 03:44:08.08299101 +0000 UTC m=+7385.249735060" watchObservedRunningTime="2025-10-02 03:44:08.091921751 +0000 UTC m=+7385.258665801" Oct 02 03:44:31 crc kubenswrapper[4775]: I1002 03:44:31.817037 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wd7tp"] Oct 02 03:44:31 crc kubenswrapper[4775]: I1002 03:44:31.822046 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:31 crc kubenswrapper[4775]: I1002 03:44:31.833501 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wd7tp"] Oct 02 03:44:31 crc kubenswrapper[4775]: I1002 03:44:31.881919 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c31a9c91-09db-4d7d-a039-5508c7c0216f-utilities\") pod \"redhat-operators-wd7tp\" (UID: \"c31a9c91-09db-4d7d-a039-5508c7c0216f\") " pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:31 crc kubenswrapper[4775]: I1002 03:44:31.882002 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp4g8\" (UniqueName: \"kubernetes.io/projected/c31a9c91-09db-4d7d-a039-5508c7c0216f-kube-api-access-wp4g8\") pod \"redhat-operators-wd7tp\" (UID: \"c31a9c91-09db-4d7d-a039-5508c7c0216f\") " pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:31 crc kubenswrapper[4775]: I1002 03:44:31.882030 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c31a9c91-09db-4d7d-a039-5508c7c0216f-catalog-content\") pod \"redhat-operators-wd7tp\" (UID: \"c31a9c91-09db-4d7d-a039-5508c7c0216f\") " pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:31 crc kubenswrapper[4775]: I1002 03:44:31.983946 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c31a9c91-09db-4d7d-a039-5508c7c0216f-utilities\") pod \"redhat-operators-wd7tp\" (UID: \"c31a9c91-09db-4d7d-a039-5508c7c0216f\") " pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:31 crc kubenswrapper[4775]: I1002 03:44:31.984013 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp4g8\" (UniqueName: \"kubernetes.io/projected/c31a9c91-09db-4d7d-a039-5508c7c0216f-kube-api-access-wp4g8\") pod \"redhat-operators-wd7tp\" (UID: \"c31a9c91-09db-4d7d-a039-5508c7c0216f\") " pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:31 crc kubenswrapper[4775]: I1002 03:44:31.984039 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c31a9c91-09db-4d7d-a039-5508c7c0216f-catalog-content\") pod \"redhat-operators-wd7tp\" (UID: \"c31a9c91-09db-4d7d-a039-5508c7c0216f\") " pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:31 crc kubenswrapper[4775]: I1002 03:44:31.984613 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c31a9c91-09db-4d7d-a039-5508c7c0216f-catalog-content\") pod \"redhat-operators-wd7tp\" (UID: \"c31a9c91-09db-4d7d-a039-5508c7c0216f\") " pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:31 crc kubenswrapper[4775]: I1002 03:44:31.984831 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c31a9c91-09db-4d7d-a039-5508c7c0216f-utilities\") pod \"redhat-operators-wd7tp\" (UID: \"c31a9c91-09db-4d7d-a039-5508c7c0216f\") " pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:32 crc kubenswrapper[4775]: I1002 03:44:32.010363 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp4g8\" (UniqueName: \"kubernetes.io/projected/c31a9c91-09db-4d7d-a039-5508c7c0216f-kube-api-access-wp4g8\") pod \"redhat-operators-wd7tp\" (UID: \"c31a9c91-09db-4d7d-a039-5508c7c0216f\") " pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:32 crc kubenswrapper[4775]: I1002 03:44:32.163768 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:32 crc kubenswrapper[4775]: I1002 03:44:32.640716 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wd7tp"] Oct 02 03:44:33 crc kubenswrapper[4775]: I1002 03:44:33.368206 4775 generic.go:334] "Generic (PLEG): container finished" podID="c31a9c91-09db-4d7d-a039-5508c7c0216f" containerID="3d970d61c00173ef83a2bd892378ea93c744e212513b488720916858b2ba300e" exitCode=0 Oct 02 03:44:33 crc kubenswrapper[4775]: I1002 03:44:33.368671 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wd7tp" event={"ID":"c31a9c91-09db-4d7d-a039-5508c7c0216f","Type":"ContainerDied","Data":"3d970d61c00173ef83a2bd892378ea93c744e212513b488720916858b2ba300e"} Oct 02 03:44:33 crc kubenswrapper[4775]: I1002 03:44:33.368705 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wd7tp" event={"ID":"c31a9c91-09db-4d7d-a039-5508c7c0216f","Type":"ContainerStarted","Data":"098790ee30cae61db4a36f3c8621da814c3ab9c526b797b5e736a31b4a11369b"} Oct 02 03:44:35 crc kubenswrapper[4775]: I1002 03:44:35.392488 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wd7tp" event={"ID":"c31a9c91-09db-4d7d-a039-5508c7c0216f","Type":"ContainerStarted","Data":"2f270f7d9ee2e8deb102348eb6e428a18bb538268764c72049467811b8dd1b3f"} Oct 02 03:44:43 crc kubenswrapper[4775]: I1002 03:44:43.501559 4775 generic.go:334] "Generic (PLEG): container finished" podID="c31a9c91-09db-4d7d-a039-5508c7c0216f" containerID="2f270f7d9ee2e8deb102348eb6e428a18bb538268764c72049467811b8dd1b3f" exitCode=0 Oct 02 03:44:43 crc kubenswrapper[4775]: I1002 03:44:43.503865 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wd7tp" event={"ID":"c31a9c91-09db-4d7d-a039-5508c7c0216f","Type":"ContainerDied","Data":"2f270f7d9ee2e8deb102348eb6e428a18bb538268764c72049467811b8dd1b3f"} Oct 02 03:44:44 crc kubenswrapper[4775]: I1002 03:44:44.529342 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wd7tp" event={"ID":"c31a9c91-09db-4d7d-a039-5508c7c0216f","Type":"ContainerStarted","Data":"5f150526a8cf6bafd6cbf11e79518870413aa554279ee0f71abd57e11eaa4694"} Oct 02 03:44:44 crc kubenswrapper[4775]: I1002 03:44:44.562706 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wd7tp" podStartSLOduration=2.9767665 podStartE2EDuration="13.562686693s" podCreationTimestamp="2025-10-02 03:44:31 +0000 UTC" firstStartedPulling="2025-10-02 03:44:33.370662283 +0000 UTC m=+7410.537406323" lastFinishedPulling="2025-10-02 03:44:43.956582476 +0000 UTC m=+7421.123326516" observedRunningTime="2025-10-02 03:44:44.55637919 +0000 UTC m=+7421.723123240" watchObservedRunningTime="2025-10-02 03:44:44.562686693 +0000 UTC m=+7421.729430743" Oct 02 03:44:52 crc kubenswrapper[4775]: I1002 03:44:52.164271 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:52 crc kubenswrapper[4775]: I1002 03:44:52.164978 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:52 crc kubenswrapper[4775]: I1002 03:44:52.231377 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:52 crc kubenswrapper[4775]: I1002 03:44:52.695519 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:52 crc kubenswrapper[4775]: I1002 03:44:52.763009 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wd7tp"] Oct 02 03:44:54 crc kubenswrapper[4775]: I1002 03:44:54.649043 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wd7tp" podUID="c31a9c91-09db-4d7d-a039-5508c7c0216f" containerName="registry-server" containerID="cri-o://5f150526a8cf6bafd6cbf11e79518870413aa554279ee0f71abd57e11eaa4694" gracePeriod=2 Oct 02 03:44:54 crc kubenswrapper[4775]: E1002 03:44:54.963934 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc31a9c91_09db_4d7d_a039_5508c7c0216f.slice/crio-conmon-5f150526a8cf6bafd6cbf11e79518870413aa554279ee0f71abd57e11eaa4694.scope\": RecentStats: unable to find data in memory cache]" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.257485 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.360347 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c31a9c91-09db-4d7d-a039-5508c7c0216f-utilities\") pod \"c31a9c91-09db-4d7d-a039-5508c7c0216f\" (UID: \"c31a9c91-09db-4d7d-a039-5508c7c0216f\") " Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.360489 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp4g8\" (UniqueName: \"kubernetes.io/projected/c31a9c91-09db-4d7d-a039-5508c7c0216f-kube-api-access-wp4g8\") pod \"c31a9c91-09db-4d7d-a039-5508c7c0216f\" (UID: \"c31a9c91-09db-4d7d-a039-5508c7c0216f\") " Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.360618 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c31a9c91-09db-4d7d-a039-5508c7c0216f-catalog-content\") pod \"c31a9c91-09db-4d7d-a039-5508c7c0216f\" (UID: \"c31a9c91-09db-4d7d-a039-5508c7c0216f\") " Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.361885 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c31a9c91-09db-4d7d-a039-5508c7c0216f-utilities" (OuterVolumeSpecName: "utilities") pod "c31a9c91-09db-4d7d-a039-5508c7c0216f" (UID: "c31a9c91-09db-4d7d-a039-5508c7c0216f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.376027 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c31a9c91-09db-4d7d-a039-5508c7c0216f-kube-api-access-wp4g8" (OuterVolumeSpecName: "kube-api-access-wp4g8") pod "c31a9c91-09db-4d7d-a039-5508c7c0216f" (UID: "c31a9c91-09db-4d7d-a039-5508c7c0216f"). InnerVolumeSpecName "kube-api-access-wp4g8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.460667 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c31a9c91-09db-4d7d-a039-5508c7c0216f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c31a9c91-09db-4d7d-a039-5508c7c0216f" (UID: "c31a9c91-09db-4d7d-a039-5508c7c0216f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.464384 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c31a9c91-09db-4d7d-a039-5508c7c0216f-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.464424 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp4g8\" (UniqueName: \"kubernetes.io/projected/c31a9c91-09db-4d7d-a039-5508c7c0216f-kube-api-access-wp4g8\") on node \"crc\" DevicePath \"\"" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.464445 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c31a9c91-09db-4d7d-a039-5508c7c0216f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.662889 4775 generic.go:334] "Generic (PLEG): container finished" podID="c31a9c91-09db-4d7d-a039-5508c7c0216f" containerID="5f150526a8cf6bafd6cbf11e79518870413aa554279ee0f71abd57e11eaa4694" exitCode=0 Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.662936 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wd7tp" event={"ID":"c31a9c91-09db-4d7d-a039-5508c7c0216f","Type":"ContainerDied","Data":"5f150526a8cf6bafd6cbf11e79518870413aa554279ee0f71abd57e11eaa4694"} Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.662985 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wd7tp" event={"ID":"c31a9c91-09db-4d7d-a039-5508c7c0216f","Type":"ContainerDied","Data":"098790ee30cae61db4a36f3c8621da814c3ab9c526b797b5e736a31b4a11369b"} Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.663006 4775 scope.go:117] "RemoveContainer" containerID="5f150526a8cf6bafd6cbf11e79518870413aa554279ee0f71abd57e11eaa4694" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.663032 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wd7tp" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.715012 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wd7tp"] Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.726025 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wd7tp"] Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.727172 4775 scope.go:117] "RemoveContainer" containerID="2f270f7d9ee2e8deb102348eb6e428a18bb538268764c72049467811b8dd1b3f" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.807077 4775 scope.go:117] "RemoveContainer" containerID="3d970d61c00173ef83a2bd892378ea93c744e212513b488720916858b2ba300e" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.829360 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c31a9c91-09db-4d7d-a039-5508c7c0216f" path="/var/lib/kubelet/pods/c31a9c91-09db-4d7d-a039-5508c7c0216f/volumes" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.887506 4775 scope.go:117] "RemoveContainer" containerID="5f150526a8cf6bafd6cbf11e79518870413aa554279ee0f71abd57e11eaa4694" Oct 02 03:44:55 crc kubenswrapper[4775]: E1002 03:44:55.889371 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f150526a8cf6bafd6cbf11e79518870413aa554279ee0f71abd57e11eaa4694\": container with ID starting with 5f150526a8cf6bafd6cbf11e79518870413aa554279ee0f71abd57e11eaa4694 not found: ID does not exist" containerID="5f150526a8cf6bafd6cbf11e79518870413aa554279ee0f71abd57e11eaa4694" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.889401 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f150526a8cf6bafd6cbf11e79518870413aa554279ee0f71abd57e11eaa4694"} err="failed to get container status \"5f150526a8cf6bafd6cbf11e79518870413aa554279ee0f71abd57e11eaa4694\": rpc error: code = NotFound desc = could not find container \"5f150526a8cf6bafd6cbf11e79518870413aa554279ee0f71abd57e11eaa4694\": container with ID starting with 5f150526a8cf6bafd6cbf11e79518870413aa554279ee0f71abd57e11eaa4694 not found: ID does not exist" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.889421 4775 scope.go:117] "RemoveContainer" containerID="2f270f7d9ee2e8deb102348eb6e428a18bb538268764c72049467811b8dd1b3f" Oct 02 03:44:55 crc kubenswrapper[4775]: E1002 03:44:55.897424 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f270f7d9ee2e8deb102348eb6e428a18bb538268764c72049467811b8dd1b3f\": container with ID starting with 2f270f7d9ee2e8deb102348eb6e428a18bb538268764c72049467811b8dd1b3f not found: ID does not exist" containerID="2f270f7d9ee2e8deb102348eb6e428a18bb538268764c72049467811b8dd1b3f" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.897471 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f270f7d9ee2e8deb102348eb6e428a18bb538268764c72049467811b8dd1b3f"} err="failed to get container status \"2f270f7d9ee2e8deb102348eb6e428a18bb538268764c72049467811b8dd1b3f\": rpc error: code = NotFound desc = could not find container \"2f270f7d9ee2e8deb102348eb6e428a18bb538268764c72049467811b8dd1b3f\": container with ID starting with 2f270f7d9ee2e8deb102348eb6e428a18bb538268764c72049467811b8dd1b3f not found: ID does not exist" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.897494 4775 scope.go:117] "RemoveContainer" containerID="3d970d61c00173ef83a2bd892378ea93c744e212513b488720916858b2ba300e" Oct 02 03:44:55 crc kubenswrapper[4775]: E1002 03:44:55.905633 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d970d61c00173ef83a2bd892378ea93c744e212513b488720916858b2ba300e\": container with ID starting with 3d970d61c00173ef83a2bd892378ea93c744e212513b488720916858b2ba300e not found: ID does not exist" containerID="3d970d61c00173ef83a2bd892378ea93c744e212513b488720916858b2ba300e" Oct 02 03:44:55 crc kubenswrapper[4775]: I1002 03:44:55.905681 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d970d61c00173ef83a2bd892378ea93c744e212513b488720916858b2ba300e"} err="failed to get container status \"3d970d61c00173ef83a2bd892378ea93c744e212513b488720916858b2ba300e\": rpc error: code = NotFound desc = could not find container \"3d970d61c00173ef83a2bd892378ea93c744e212513b488720916858b2ba300e\": container with ID starting with 3d970d61c00173ef83a2bd892378ea93c744e212513b488720916858b2ba300e not found: ID does not exist" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.168873 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x"] Oct 02 03:45:00 crc kubenswrapper[4775]: E1002 03:45:00.169889 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c31a9c91-09db-4d7d-a039-5508c7c0216f" containerName="extract-utilities" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.169904 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c31a9c91-09db-4d7d-a039-5508c7c0216f" containerName="extract-utilities" Oct 02 03:45:00 crc kubenswrapper[4775]: E1002 03:45:00.169923 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c31a9c91-09db-4d7d-a039-5508c7c0216f" containerName="extract-content" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.169930 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c31a9c91-09db-4d7d-a039-5508c7c0216f" containerName="extract-content" Oct 02 03:45:00 crc kubenswrapper[4775]: E1002 03:45:00.169949 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c31a9c91-09db-4d7d-a039-5508c7c0216f" containerName="registry-server" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.169955 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="c31a9c91-09db-4d7d-a039-5508c7c0216f" containerName="registry-server" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.170217 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="c31a9c91-09db-4d7d-a039-5508c7c0216f" containerName="registry-server" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.171027 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.173957 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.175496 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.201348 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x"] Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.245058 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa9a4ec4-fd61-4b68-839f-b617df1204eb-config-volume\") pod \"collect-profiles-29322945-bds7x\" (UID: \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.245151 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fb6g\" (UniqueName: \"kubernetes.io/projected/fa9a4ec4-fd61-4b68-839f-b617df1204eb-kube-api-access-8fb6g\") pod \"collect-profiles-29322945-bds7x\" (UID: \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.245190 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa9a4ec4-fd61-4b68-839f-b617df1204eb-secret-volume\") pod \"collect-profiles-29322945-bds7x\" (UID: \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.347945 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa9a4ec4-fd61-4b68-839f-b617df1204eb-config-volume\") pod \"collect-profiles-29322945-bds7x\" (UID: \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.348098 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fb6g\" (UniqueName: \"kubernetes.io/projected/fa9a4ec4-fd61-4b68-839f-b617df1204eb-kube-api-access-8fb6g\") pod \"collect-profiles-29322945-bds7x\" (UID: \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.348139 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa9a4ec4-fd61-4b68-839f-b617df1204eb-secret-volume\") pod \"collect-profiles-29322945-bds7x\" (UID: \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.349785 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa9a4ec4-fd61-4b68-839f-b617df1204eb-config-volume\") pod \"collect-profiles-29322945-bds7x\" (UID: \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.356601 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa9a4ec4-fd61-4b68-839f-b617df1204eb-secret-volume\") pod \"collect-profiles-29322945-bds7x\" (UID: \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.378129 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fb6g\" (UniqueName: \"kubernetes.io/projected/fa9a4ec4-fd61-4b68-839f-b617df1204eb-kube-api-access-8fb6g\") pod \"collect-profiles-29322945-bds7x\" (UID: \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" Oct 02 03:45:00 crc kubenswrapper[4775]: I1002 03:45:00.507562 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" Oct 02 03:45:01 crc kubenswrapper[4775]: I1002 03:45:01.014318 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x"] Oct 02 03:45:01 crc kubenswrapper[4775]: I1002 03:45:01.746020 4775 generic.go:334] "Generic (PLEG): container finished" podID="fa9a4ec4-fd61-4b68-839f-b617df1204eb" containerID="0b8651815abbac281863662f159ddfd32dd0b8b07d1c687d2a16eca3dec0622d" exitCode=0 Oct 02 03:45:01 crc kubenswrapper[4775]: I1002 03:45:01.746091 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" event={"ID":"fa9a4ec4-fd61-4b68-839f-b617df1204eb","Type":"ContainerDied","Data":"0b8651815abbac281863662f159ddfd32dd0b8b07d1c687d2a16eca3dec0622d"} Oct 02 03:45:01 crc kubenswrapper[4775]: I1002 03:45:01.746338 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" event={"ID":"fa9a4ec4-fd61-4b68-839f-b617df1204eb","Type":"ContainerStarted","Data":"f181ae011526341d1e7d3db447c84dcc6d7962d3911b84456918f6d596c7340b"} Oct 02 03:45:03 crc kubenswrapper[4775]: I1002 03:45:03.267624 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" Oct 02 03:45:03 crc kubenswrapper[4775]: I1002 03:45:03.426489 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fb6g\" (UniqueName: \"kubernetes.io/projected/fa9a4ec4-fd61-4b68-839f-b617df1204eb-kube-api-access-8fb6g\") pod \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\" (UID: \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\") " Oct 02 03:45:03 crc kubenswrapper[4775]: I1002 03:45:03.426823 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa9a4ec4-fd61-4b68-839f-b617df1204eb-secret-volume\") pod \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\" (UID: \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\") " Oct 02 03:45:03 crc kubenswrapper[4775]: I1002 03:45:03.426918 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa9a4ec4-fd61-4b68-839f-b617df1204eb-config-volume\") pod \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\" (UID: \"fa9a4ec4-fd61-4b68-839f-b617df1204eb\") " Oct 02 03:45:03 crc kubenswrapper[4775]: I1002 03:45:03.427732 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa9a4ec4-fd61-4b68-839f-b617df1204eb-config-volume" (OuterVolumeSpecName: "config-volume") pod "fa9a4ec4-fd61-4b68-839f-b617df1204eb" (UID: "fa9a4ec4-fd61-4b68-839f-b617df1204eb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:45:03 crc kubenswrapper[4775]: I1002 03:45:03.428437 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fa9a4ec4-fd61-4b68-839f-b617df1204eb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 03:45:03 crc kubenswrapper[4775]: I1002 03:45:03.432784 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa9a4ec4-fd61-4b68-839f-b617df1204eb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fa9a4ec4-fd61-4b68-839f-b617df1204eb" (UID: "fa9a4ec4-fd61-4b68-839f-b617df1204eb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:45:03 crc kubenswrapper[4775]: I1002 03:45:03.438203 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa9a4ec4-fd61-4b68-839f-b617df1204eb-kube-api-access-8fb6g" (OuterVolumeSpecName: "kube-api-access-8fb6g") pod "fa9a4ec4-fd61-4b68-839f-b617df1204eb" (UID: "fa9a4ec4-fd61-4b68-839f-b617df1204eb"). InnerVolumeSpecName "kube-api-access-8fb6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:45:03 crc kubenswrapper[4775]: I1002 03:45:03.531027 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fa9a4ec4-fd61-4b68-839f-b617df1204eb-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 03:45:03 crc kubenswrapper[4775]: I1002 03:45:03.531059 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fb6g\" (UniqueName: \"kubernetes.io/projected/fa9a4ec4-fd61-4b68-839f-b617df1204eb-kube-api-access-8fb6g\") on node \"crc\" DevicePath \"\"" Oct 02 03:45:03 crc kubenswrapper[4775]: I1002 03:45:03.783273 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" Oct 02 03:45:03 crc kubenswrapper[4775]: I1002 03:45:03.788365 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x" event={"ID":"fa9a4ec4-fd61-4b68-839f-b617df1204eb","Type":"ContainerDied","Data":"f181ae011526341d1e7d3db447c84dcc6d7962d3911b84456918f6d596c7340b"} Oct 02 03:45:03 crc kubenswrapper[4775]: I1002 03:45:03.788411 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f181ae011526341d1e7d3db447c84dcc6d7962d3911b84456918f6d596c7340b" Oct 02 03:45:04 crc kubenswrapper[4775]: I1002 03:45:04.357898 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l"] Oct 02 03:45:04 crc kubenswrapper[4775]: I1002 03:45:04.368477 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322900-bft5l"] Oct 02 03:45:05 crc kubenswrapper[4775]: I1002 03:45:05.799897 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="50401de0-705a-4de5-b211-56688c040651" path="/var/lib/kubelet/pods/50401de0-705a-4de5-b211-56688c040651/volumes" Oct 02 03:45:17 crc kubenswrapper[4775]: I1002 03:45:17.289454 4775 scope.go:117] "RemoveContainer" containerID="3fcb369bc37736ee508ccc3df41fe39721387a332a1c8dffcd55e2d473ade333" Oct 02 03:45:42 crc kubenswrapper[4775]: I1002 03:45:42.321225 4775 generic.go:334] "Generic (PLEG): container finished" podID="e15f9838-d5f3-4b6c-b968-de51d37ffaf9" containerID="fde99f848a13206969111a51ff4d46d89b750075f20eb229463e4a54ab713d5e" exitCode=0 Oct 02 03:45:42 crc kubenswrapper[4775]: I1002 03:45:42.321307 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" event={"ID":"e15f9838-d5f3-4b6c-b968-de51d37ffaf9","Type":"ContainerDied","Data":"fde99f848a13206969111a51ff4d46d89b750075f20eb229463e4a54ab713d5e"} Oct 02 03:45:43 crc kubenswrapper[4775]: I1002 03:45:43.937648 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.102667 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9gnc\" (UniqueName: \"kubernetes.io/projected/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-kube-api-access-m9gnc\") pod \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.102805 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-ceph\") pod \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.102942 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-ssh-key\") pod \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.103098 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-inventory\") pod \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\" (UID: \"e15f9838-d5f3-4b6c-b968-de51d37ffaf9\") " Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.109572 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-kube-api-access-m9gnc" (OuterVolumeSpecName: "kube-api-access-m9gnc") pod "e15f9838-d5f3-4b6c-b968-de51d37ffaf9" (UID: "e15f9838-d5f3-4b6c-b968-de51d37ffaf9"). InnerVolumeSpecName "kube-api-access-m9gnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.110191 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-ceph" (OuterVolumeSpecName: "ceph") pod "e15f9838-d5f3-4b6c-b968-de51d37ffaf9" (UID: "e15f9838-d5f3-4b6c-b968-de51d37ffaf9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.132627 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-inventory" (OuterVolumeSpecName: "inventory") pod "e15f9838-d5f3-4b6c-b968-de51d37ffaf9" (UID: "e15f9838-d5f3-4b6c-b968-de51d37ffaf9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.153010 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e15f9838-d5f3-4b6c-b968-de51d37ffaf9" (UID: "e15f9838-d5f3-4b6c-b968-de51d37ffaf9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.205746 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9gnc\" (UniqueName: \"kubernetes.io/projected/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-kube-api-access-m9gnc\") on node \"crc\" DevicePath \"\"" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.205783 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.205796 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.205810 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e15f9838-d5f3-4b6c-b968-de51d37ffaf9-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.347741 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" event={"ID":"e15f9838-d5f3-4b6c-b968-de51d37ffaf9","Type":"ContainerDied","Data":"9ee797e2cd2425f65e028d865efb8b49b2f97336d1aae5997a2f6c0b72dc8c72"} Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.347787 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ee797e2cd2425f65e028d865efb8b49b2f97336d1aae5997a2f6c0b72dc8c72" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.347823 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-5mg6v" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.473431 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-x58ln"] Oct 02 03:45:44 crc kubenswrapper[4775]: E1002 03:45:44.474148 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e15f9838-d5f3-4b6c-b968-de51d37ffaf9" containerName="download-cache-openstack-openstack-cell1" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.474194 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e15f9838-d5f3-4b6c-b968-de51d37ffaf9" containerName="download-cache-openstack-openstack-cell1" Oct 02 03:45:44 crc kubenswrapper[4775]: E1002 03:45:44.474252 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa9a4ec4-fd61-4b68-839f-b617df1204eb" containerName="collect-profiles" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.474263 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa9a4ec4-fd61-4b68-839f-b617df1204eb" containerName="collect-profiles" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.474536 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e15f9838-d5f3-4b6c-b968-de51d37ffaf9" containerName="download-cache-openstack-openstack-cell1" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.474588 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa9a4ec4-fd61-4b68-839f-b617df1204eb" containerName="collect-profiles" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.475802 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.479070 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.479462 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.479918 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.481748 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.501254 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-x58ln"] Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.616139 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-inventory\") pod \"configure-network-openstack-openstack-cell1-x58ln\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.616274 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-ceph\") pod \"configure-network-openstack-openstack-cell1-x58ln\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.616353 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2fc7\" (UniqueName: \"kubernetes.io/projected/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-kube-api-access-f2fc7\") pod \"configure-network-openstack-openstack-cell1-x58ln\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.616377 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-ssh-key\") pod \"configure-network-openstack-openstack-cell1-x58ln\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.718927 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-inventory\") pod \"configure-network-openstack-openstack-cell1-x58ln\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.719038 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-ceph\") pod \"configure-network-openstack-openstack-cell1-x58ln\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.719133 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2fc7\" (UniqueName: \"kubernetes.io/projected/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-kube-api-access-f2fc7\") pod \"configure-network-openstack-openstack-cell1-x58ln\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.719163 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-ssh-key\") pod \"configure-network-openstack-openstack-cell1-x58ln\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.726229 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-ssh-key\") pod \"configure-network-openstack-openstack-cell1-x58ln\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.726425 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-inventory\") pod \"configure-network-openstack-openstack-cell1-x58ln\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.727569 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-ceph\") pod \"configure-network-openstack-openstack-cell1-x58ln\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.757606 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2fc7\" (UniqueName: \"kubernetes.io/projected/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-kube-api-access-f2fc7\") pod \"configure-network-openstack-openstack-cell1-x58ln\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:45:44 crc kubenswrapper[4775]: I1002 03:45:44.797616 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:45:45 crc kubenswrapper[4775]: I1002 03:45:45.406671 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-x58ln"] Oct 02 03:45:45 crc kubenswrapper[4775]: W1002 03:45:45.414550 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3dc106c_ef5c_4394_91ee_d0ba9cf0c1da.slice/crio-8cdf147d90eefe4905be99b5c1cb561ba3686c5c128c0da31883cc0ae96f45e9 WatchSource:0}: Error finding container 8cdf147d90eefe4905be99b5c1cb561ba3686c5c128c0da31883cc0ae96f45e9: Status 404 returned error can't find the container with id 8cdf147d90eefe4905be99b5c1cb561ba3686c5c128c0da31883cc0ae96f45e9 Oct 02 03:45:46 crc kubenswrapper[4775]: I1002 03:45:46.382699 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-x58ln" event={"ID":"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da","Type":"ContainerStarted","Data":"8cdf147d90eefe4905be99b5c1cb561ba3686c5c128c0da31883cc0ae96f45e9"} Oct 02 03:45:47 crc kubenswrapper[4775]: I1002 03:45:47.397507 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-x58ln" event={"ID":"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da","Type":"ContainerStarted","Data":"af2629c1b9fceff4a99f153eee59d307f292c1e559722c5aa135aa0d5a946331"} Oct 02 03:45:47 crc kubenswrapper[4775]: I1002 03:45:47.419645 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-x58ln" podStartSLOduration=2.61974764 podStartE2EDuration="3.419619138s" podCreationTimestamp="2025-10-02 03:45:44 +0000 UTC" firstStartedPulling="2025-10-02 03:45:45.418027256 +0000 UTC m=+7482.584771306" lastFinishedPulling="2025-10-02 03:45:46.217898724 +0000 UTC m=+7483.384642804" observedRunningTime="2025-10-02 03:45:47.415904022 +0000 UTC m=+7484.582648102" watchObservedRunningTime="2025-10-02 03:45:47.419619138 +0000 UTC m=+7484.586363218" Oct 02 03:46:07 crc kubenswrapper[4775]: I1002 03:46:07.233287 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:46:07 crc kubenswrapper[4775]: I1002 03:46:07.233933 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:46:37 crc kubenswrapper[4775]: I1002 03:46:37.233930 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:46:37 crc kubenswrapper[4775]: I1002 03:46:37.234746 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:47:07 crc kubenswrapper[4775]: I1002 03:47:07.233892 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:47:07 crc kubenswrapper[4775]: I1002 03:47:07.234591 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:47:07 crc kubenswrapper[4775]: I1002 03:47:07.234652 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 03:47:07 crc kubenswrapper[4775]: I1002 03:47:07.235693 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"82e9bf60a688a73f4e67e12ca66bd36d7e6ce4ea01f01845c9edec604f41bbff"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 03:47:07 crc kubenswrapper[4775]: I1002 03:47:07.235764 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://82e9bf60a688a73f4e67e12ca66bd36d7e6ce4ea01f01845c9edec604f41bbff" gracePeriod=600 Oct 02 03:47:07 crc kubenswrapper[4775]: I1002 03:47:07.425477 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="82e9bf60a688a73f4e67e12ca66bd36d7e6ce4ea01f01845c9edec604f41bbff" exitCode=0 Oct 02 03:47:07 crc kubenswrapper[4775]: I1002 03:47:07.425554 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"82e9bf60a688a73f4e67e12ca66bd36d7e6ce4ea01f01845c9edec604f41bbff"} Oct 02 03:47:07 crc kubenswrapper[4775]: I1002 03:47:07.425834 4775 scope.go:117] "RemoveContainer" containerID="135fed1ed3e26f43380ba79796bb0fb34a37fc70c0f687bfebe63ccfad2aa8c0" Oct 02 03:47:08 crc kubenswrapper[4775]: I1002 03:47:08.439202 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4"} Oct 02 03:47:12 crc kubenswrapper[4775]: I1002 03:47:12.499534 4775 generic.go:334] "Generic (PLEG): container finished" podID="e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da" containerID="af2629c1b9fceff4a99f153eee59d307f292c1e559722c5aa135aa0d5a946331" exitCode=0 Oct 02 03:47:12 crc kubenswrapper[4775]: I1002 03:47:12.499606 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-x58ln" event={"ID":"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da","Type":"ContainerDied","Data":"af2629c1b9fceff4a99f153eee59d307f292c1e559722c5aa135aa0d5a946331"} Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.097388 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.207775 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-ceph\") pod \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.207888 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-ssh-key\") pod \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.208099 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2fc7\" (UniqueName: \"kubernetes.io/projected/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-kube-api-access-f2fc7\") pod \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.208157 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-inventory\") pod \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\" (UID: \"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da\") " Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.214032 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-ceph" (OuterVolumeSpecName: "ceph") pod "e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da" (UID: "e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.214350 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-kube-api-access-f2fc7" (OuterVolumeSpecName: "kube-api-access-f2fc7") pod "e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da" (UID: "e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da"). InnerVolumeSpecName "kube-api-access-f2fc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.242497 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-inventory" (OuterVolumeSpecName: "inventory") pod "e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da" (UID: "e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.245920 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da" (UID: "e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.310473 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2fc7\" (UniqueName: \"kubernetes.io/projected/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-kube-api-access-f2fc7\") on node \"crc\" DevicePath \"\"" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.310522 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.310541 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.310558 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.524030 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-x58ln" event={"ID":"e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da","Type":"ContainerDied","Data":"8cdf147d90eefe4905be99b5c1cb561ba3686c5c128c0da31883cc0ae96f45e9"} Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.524080 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cdf147d90eefe4905be99b5c1cb561ba3686c5c128c0da31883cc0ae96f45e9" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.524244 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-x58ln" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.615989 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-qt4nc"] Oct 02 03:47:14 crc kubenswrapper[4775]: E1002 03:47:14.616560 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da" containerName="configure-network-openstack-openstack-cell1" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.616583 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da" containerName="configure-network-openstack-openstack-cell1" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.616831 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da" containerName="configure-network-openstack-openstack-cell1" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.617842 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.623510 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.623905 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.624071 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.624263 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.639107 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-qt4nc"] Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.730364 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-ssh-key\") pod \"validate-network-openstack-openstack-cell1-qt4nc\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.730694 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-ceph\") pod \"validate-network-openstack-openstack-cell1-qt4nc\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.730756 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpxm6\" (UniqueName: \"kubernetes.io/projected/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-kube-api-access-xpxm6\") pod \"validate-network-openstack-openstack-cell1-qt4nc\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.730775 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-inventory\") pod \"validate-network-openstack-openstack-cell1-qt4nc\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.832891 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-ssh-key\") pod \"validate-network-openstack-openstack-cell1-qt4nc\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.832958 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-ceph\") pod \"validate-network-openstack-openstack-cell1-qt4nc\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.832991 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpxm6\" (UniqueName: \"kubernetes.io/projected/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-kube-api-access-xpxm6\") pod \"validate-network-openstack-openstack-cell1-qt4nc\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.833010 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-inventory\") pod \"validate-network-openstack-openstack-cell1-qt4nc\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.837116 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-ssh-key\") pod \"validate-network-openstack-openstack-cell1-qt4nc\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.837314 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-ceph\") pod \"validate-network-openstack-openstack-cell1-qt4nc\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.837405 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-inventory\") pod \"validate-network-openstack-openstack-cell1-qt4nc\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.849622 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpxm6\" (UniqueName: \"kubernetes.io/projected/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-kube-api-access-xpxm6\") pod \"validate-network-openstack-openstack-cell1-qt4nc\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:14 crc kubenswrapper[4775]: I1002 03:47:14.936983 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:16 crc kubenswrapper[4775]: I1002 03:47:15.655323 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-qt4nc"] Oct 02 03:47:16 crc kubenswrapper[4775]: I1002 03:47:15.658028 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 03:47:16 crc kubenswrapper[4775]: I1002 03:47:16.551501 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" event={"ID":"663e6cdb-d635-4083-bf0a-7b00e61dc1ff","Type":"ContainerStarted","Data":"6f5dd3d7b332489b511c730343187b188456bfb94bb841604a4b36c6e28b34e8"} Oct 02 03:47:16 crc kubenswrapper[4775]: I1002 03:47:16.552071 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" event={"ID":"663e6cdb-d635-4083-bf0a-7b00e61dc1ff","Type":"ContainerStarted","Data":"df2cd18edb6acbe4d179db48e6717a4eacaac1e38ddc1ddda05ff2b44d7bcfb8"} Oct 02 03:47:16 crc kubenswrapper[4775]: I1002 03:47:16.582335 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" podStartSLOduration=1.951382671 podStartE2EDuration="2.582313519s" podCreationTimestamp="2025-10-02 03:47:14 +0000 UTC" firstStartedPulling="2025-10-02 03:47:15.657825033 +0000 UTC m=+7572.824569073" lastFinishedPulling="2025-10-02 03:47:16.288755841 +0000 UTC m=+7573.455499921" observedRunningTime="2025-10-02 03:47:16.572750423 +0000 UTC m=+7573.739494483" watchObservedRunningTime="2025-10-02 03:47:16.582313519 +0000 UTC m=+7573.749057559" Oct 02 03:47:22 crc kubenswrapper[4775]: I1002 03:47:22.628569 4775 generic.go:334] "Generic (PLEG): container finished" podID="663e6cdb-d635-4083-bf0a-7b00e61dc1ff" containerID="6f5dd3d7b332489b511c730343187b188456bfb94bb841604a4b36c6e28b34e8" exitCode=0 Oct 02 03:47:22 crc kubenswrapper[4775]: I1002 03:47:22.628659 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" event={"ID":"663e6cdb-d635-4083-bf0a-7b00e61dc1ff","Type":"ContainerDied","Data":"6f5dd3d7b332489b511c730343187b188456bfb94bb841604a4b36c6e28b34e8"} Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.110320 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.245293 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-ceph\") pod \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.245429 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpxm6\" (UniqueName: \"kubernetes.io/projected/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-kube-api-access-xpxm6\") pod \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.245568 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-inventory\") pod \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.245819 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-ssh-key\") pod \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\" (UID: \"663e6cdb-d635-4083-bf0a-7b00e61dc1ff\") " Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.250916 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-kube-api-access-xpxm6" (OuterVolumeSpecName: "kube-api-access-xpxm6") pod "663e6cdb-d635-4083-bf0a-7b00e61dc1ff" (UID: "663e6cdb-d635-4083-bf0a-7b00e61dc1ff"). InnerVolumeSpecName "kube-api-access-xpxm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.251681 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-ceph" (OuterVolumeSpecName: "ceph") pod "663e6cdb-d635-4083-bf0a-7b00e61dc1ff" (UID: "663e6cdb-d635-4083-bf0a-7b00e61dc1ff"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.273592 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-inventory" (OuterVolumeSpecName: "inventory") pod "663e6cdb-d635-4083-bf0a-7b00e61dc1ff" (UID: "663e6cdb-d635-4083-bf0a-7b00e61dc1ff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.274498 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "663e6cdb-d635-4083-bf0a-7b00e61dc1ff" (UID: "663e6cdb-d635-4083-bf0a-7b00e61dc1ff"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.348174 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.348206 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpxm6\" (UniqueName: \"kubernetes.io/projected/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-kube-api-access-xpxm6\") on node \"crc\" DevicePath \"\"" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.348217 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.348229 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/663e6cdb-d635-4083-bf0a-7b00e61dc1ff-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.649123 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" event={"ID":"663e6cdb-d635-4083-bf0a-7b00e61dc1ff","Type":"ContainerDied","Data":"df2cd18edb6acbe4d179db48e6717a4eacaac1e38ddc1ddda05ff2b44d7bcfb8"} Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.649410 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df2cd18edb6acbe4d179db48e6717a4eacaac1e38ddc1ddda05ff2b44d7bcfb8" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.649166 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-qt4nc" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.720410 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-9jghw"] Oct 02 03:47:24 crc kubenswrapper[4775]: E1002 03:47:24.720825 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="663e6cdb-d635-4083-bf0a-7b00e61dc1ff" containerName="validate-network-openstack-openstack-cell1" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.720842 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="663e6cdb-d635-4083-bf0a-7b00e61dc1ff" containerName="validate-network-openstack-openstack-cell1" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.721045 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="663e6cdb-d635-4083-bf0a-7b00e61dc1ff" containerName="validate-network-openstack-openstack-cell1" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.721714 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.727638 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.727840 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.727980 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.728670 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.738396 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-9jghw"] Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.857828 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-ceph\") pod \"install-os-openstack-openstack-cell1-9jghw\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.858135 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-ssh-key\") pod \"install-os-openstack-openstack-cell1-9jghw\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.858451 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-inventory\") pod \"install-os-openstack-openstack-cell1-9jghw\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.858617 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvbw8\" (UniqueName: \"kubernetes.io/projected/d998f892-b126-406a-bfad-568524759afd-kube-api-access-xvbw8\") pod \"install-os-openstack-openstack-cell1-9jghw\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.961310 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-ceph\") pod \"install-os-openstack-openstack-cell1-9jghw\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.961452 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-ssh-key\") pod \"install-os-openstack-openstack-cell1-9jghw\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.961618 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-inventory\") pod \"install-os-openstack-openstack-cell1-9jghw\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.961726 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvbw8\" (UniqueName: \"kubernetes.io/projected/d998f892-b126-406a-bfad-568524759afd-kube-api-access-xvbw8\") pod \"install-os-openstack-openstack-cell1-9jghw\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.967064 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-ceph\") pod \"install-os-openstack-openstack-cell1-9jghw\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.968477 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-ssh-key\") pod \"install-os-openstack-openstack-cell1-9jghw\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.969501 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-inventory\") pod \"install-os-openstack-openstack-cell1-9jghw\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:47:24 crc kubenswrapper[4775]: I1002 03:47:24.978146 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvbw8\" (UniqueName: \"kubernetes.io/projected/d998f892-b126-406a-bfad-568524759afd-kube-api-access-xvbw8\") pod \"install-os-openstack-openstack-cell1-9jghw\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:47:25 crc kubenswrapper[4775]: I1002 03:47:25.041241 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:47:25 crc kubenswrapper[4775]: I1002 03:47:25.658247 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-9jghw"] Oct 02 03:47:26 crc kubenswrapper[4775]: I1002 03:47:26.672286 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-9jghw" event={"ID":"d998f892-b126-406a-bfad-568524759afd","Type":"ContainerStarted","Data":"af6e5f40b18f7ca533e5d23b4282927e00379da63bb20eb914cb93898c2854d7"} Oct 02 03:47:26 crc kubenswrapper[4775]: I1002 03:47:26.672900 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-9jghw" event={"ID":"d998f892-b126-406a-bfad-568524759afd","Type":"ContainerStarted","Data":"2d82a09ca259ab5c4ff7ec406370d9e0e7678d79d401e1a5308ddf42d35f8f89"} Oct 02 03:48:09 crc kubenswrapper[4775]: I1002 03:48:09.239946 4775 generic.go:334] "Generic (PLEG): container finished" podID="d998f892-b126-406a-bfad-568524759afd" containerID="af6e5f40b18f7ca533e5d23b4282927e00379da63bb20eb914cb93898c2854d7" exitCode=0 Oct 02 03:48:09 crc kubenswrapper[4775]: I1002 03:48:09.240142 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-9jghw" event={"ID":"d998f892-b126-406a-bfad-568524759afd","Type":"ContainerDied","Data":"af6e5f40b18f7ca533e5d23b4282927e00379da63bb20eb914cb93898c2854d7"} Oct 02 03:48:10 crc kubenswrapper[4775]: I1002 03:48:10.879940 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:48:10 crc kubenswrapper[4775]: I1002 03:48:10.969178 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvbw8\" (UniqueName: \"kubernetes.io/projected/d998f892-b126-406a-bfad-568524759afd-kube-api-access-xvbw8\") pod \"d998f892-b126-406a-bfad-568524759afd\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " Oct 02 03:48:10 crc kubenswrapper[4775]: I1002 03:48:10.969387 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-ceph\") pod \"d998f892-b126-406a-bfad-568524759afd\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " Oct 02 03:48:10 crc kubenswrapper[4775]: I1002 03:48:10.969649 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-ssh-key\") pod \"d998f892-b126-406a-bfad-568524759afd\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " Oct 02 03:48:10 crc kubenswrapper[4775]: I1002 03:48:10.969862 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-inventory\") pod \"d998f892-b126-406a-bfad-568524759afd\" (UID: \"d998f892-b126-406a-bfad-568524759afd\") " Oct 02 03:48:10 crc kubenswrapper[4775]: I1002 03:48:10.975887 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-ceph" (OuterVolumeSpecName: "ceph") pod "d998f892-b126-406a-bfad-568524759afd" (UID: "d998f892-b126-406a-bfad-568524759afd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:48:10 crc kubenswrapper[4775]: I1002 03:48:10.975989 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d998f892-b126-406a-bfad-568524759afd-kube-api-access-xvbw8" (OuterVolumeSpecName: "kube-api-access-xvbw8") pod "d998f892-b126-406a-bfad-568524759afd" (UID: "d998f892-b126-406a-bfad-568524759afd"). InnerVolumeSpecName "kube-api-access-xvbw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.008660 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-inventory" (OuterVolumeSpecName: "inventory") pod "d998f892-b126-406a-bfad-568524759afd" (UID: "d998f892-b126-406a-bfad-568524759afd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.008774 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d998f892-b126-406a-bfad-568524759afd" (UID: "d998f892-b126-406a-bfad-568524759afd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.073174 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.073202 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvbw8\" (UniqueName: \"kubernetes.io/projected/d998f892-b126-406a-bfad-568524759afd-kube-api-access-xvbw8\") on node \"crc\" DevicePath \"\"" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.073213 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.073221 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d998f892-b126-406a-bfad-568524759afd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.272564 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-9jghw" event={"ID":"d998f892-b126-406a-bfad-568524759afd","Type":"ContainerDied","Data":"2d82a09ca259ab5c4ff7ec406370d9e0e7678d79d401e1a5308ddf42d35f8f89"} Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.272664 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-9jghw" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.272666 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d82a09ca259ab5c4ff7ec406370d9e0e7678d79d401e1a5308ddf42d35f8f89" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.378135 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-gsmzl"] Oct 02 03:48:11 crc kubenswrapper[4775]: E1002 03:48:11.378631 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d998f892-b126-406a-bfad-568524759afd" containerName="install-os-openstack-openstack-cell1" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.378649 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="d998f892-b126-406a-bfad-568524759afd" containerName="install-os-openstack-openstack-cell1" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.378874 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="d998f892-b126-406a-bfad-568524759afd" containerName="install-os-openstack-openstack-cell1" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.380267 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.382032 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.385184 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.385464 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.387136 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.391139 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-gsmzl"] Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.479304 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-inventory\") pod \"configure-os-openstack-openstack-cell1-gsmzl\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.479370 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-ssh-key\") pod \"configure-os-openstack-openstack-cell1-gsmzl\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.479457 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-ceph\") pod \"configure-os-openstack-openstack-cell1-gsmzl\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.479697 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8lvf\" (UniqueName: \"kubernetes.io/projected/1e299556-d363-499a-9aec-6f1241060e4a-kube-api-access-t8lvf\") pod \"configure-os-openstack-openstack-cell1-gsmzl\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:48:11 crc kubenswrapper[4775]: E1002 03:48:11.502834 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd998f892_b126_406a_bfad_568524759afd.slice/crio-2d82a09ca259ab5c4ff7ec406370d9e0e7678d79d401e1a5308ddf42d35f8f89\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd998f892_b126_406a_bfad_568524759afd.slice\": RecentStats: unable to find data in memory cache]" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.585039 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-ceph\") pod \"configure-os-openstack-openstack-cell1-gsmzl\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.585220 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8lvf\" (UniqueName: \"kubernetes.io/projected/1e299556-d363-499a-9aec-6f1241060e4a-kube-api-access-t8lvf\") pod \"configure-os-openstack-openstack-cell1-gsmzl\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.585326 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-inventory\") pod \"configure-os-openstack-openstack-cell1-gsmzl\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.585410 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-ssh-key\") pod \"configure-os-openstack-openstack-cell1-gsmzl\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.594003 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-inventory\") pod \"configure-os-openstack-openstack-cell1-gsmzl\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.594083 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-ceph\") pod \"configure-os-openstack-openstack-cell1-gsmzl\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.594383 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-ssh-key\") pod \"configure-os-openstack-openstack-cell1-gsmzl\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.607062 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8lvf\" (UniqueName: \"kubernetes.io/projected/1e299556-d363-499a-9aec-6f1241060e4a-kube-api-access-t8lvf\") pod \"configure-os-openstack-openstack-cell1-gsmzl\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:48:11 crc kubenswrapper[4775]: I1002 03:48:11.702870 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:48:12 crc kubenswrapper[4775]: I1002 03:48:12.428814 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-gsmzl"] Oct 02 03:48:13 crc kubenswrapper[4775]: I1002 03:48:13.301704 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" event={"ID":"1e299556-d363-499a-9aec-6f1241060e4a","Type":"ContainerStarted","Data":"7b02d0c26a1689a33e95a508b3d68658adbc816043a4e267e37598c61d267475"} Oct 02 03:48:14 crc kubenswrapper[4775]: I1002 03:48:14.319153 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" event={"ID":"1e299556-d363-499a-9aec-6f1241060e4a","Type":"ContainerStarted","Data":"d8398d38e344c59004c39c8f50f91f2909662f354946c73e818a977901bc26a9"} Oct 02 03:48:14 crc kubenswrapper[4775]: I1002 03:48:14.343111 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" podStartSLOduration=2.692275646 podStartE2EDuration="3.343082146s" podCreationTimestamp="2025-10-02 03:48:11 +0000 UTC" firstStartedPulling="2025-10-02 03:48:12.415573337 +0000 UTC m=+7629.582317407" lastFinishedPulling="2025-10-02 03:48:13.066379857 +0000 UTC m=+7630.233123907" observedRunningTime="2025-10-02 03:48:14.339410211 +0000 UTC m=+7631.506154311" watchObservedRunningTime="2025-10-02 03:48:14.343082146 +0000 UTC m=+7631.509826216" Oct 02 03:48:59 crc kubenswrapper[4775]: I1002 03:48:59.885449 4775 generic.go:334] "Generic (PLEG): container finished" podID="1e299556-d363-499a-9aec-6f1241060e4a" containerID="d8398d38e344c59004c39c8f50f91f2909662f354946c73e818a977901bc26a9" exitCode=0 Oct 02 03:48:59 crc kubenswrapper[4775]: I1002 03:48:59.885529 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" event={"ID":"1e299556-d363-499a-9aec-6f1241060e4a","Type":"ContainerDied","Data":"d8398d38e344c59004c39c8f50f91f2909662f354946c73e818a977901bc26a9"} Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.463046 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.613911 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-ceph\") pod \"1e299556-d363-499a-9aec-6f1241060e4a\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.614293 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-ssh-key\") pod \"1e299556-d363-499a-9aec-6f1241060e4a\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.614344 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8lvf\" (UniqueName: \"kubernetes.io/projected/1e299556-d363-499a-9aec-6f1241060e4a-kube-api-access-t8lvf\") pod \"1e299556-d363-499a-9aec-6f1241060e4a\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.614544 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-inventory\") pod \"1e299556-d363-499a-9aec-6f1241060e4a\" (UID: \"1e299556-d363-499a-9aec-6f1241060e4a\") " Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.621839 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-ceph" (OuterVolumeSpecName: "ceph") pod "1e299556-d363-499a-9aec-6f1241060e4a" (UID: "1e299556-d363-499a-9aec-6f1241060e4a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.624634 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e299556-d363-499a-9aec-6f1241060e4a-kube-api-access-t8lvf" (OuterVolumeSpecName: "kube-api-access-t8lvf") pod "1e299556-d363-499a-9aec-6f1241060e4a" (UID: "1e299556-d363-499a-9aec-6f1241060e4a"). InnerVolumeSpecName "kube-api-access-t8lvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.664450 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1e299556-d363-499a-9aec-6f1241060e4a" (UID: "1e299556-d363-499a-9aec-6f1241060e4a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.665667 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-inventory" (OuterVolumeSpecName: "inventory") pod "1e299556-d363-499a-9aec-6f1241060e4a" (UID: "1e299556-d363-499a-9aec-6f1241060e4a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.716913 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.716970 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.716982 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1e299556-d363-499a-9aec-6f1241060e4a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.716993 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8lvf\" (UniqueName: \"kubernetes.io/projected/1e299556-d363-499a-9aec-6f1241060e4a-kube-api-access-t8lvf\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.909405 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.909393 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-gsmzl" event={"ID":"1e299556-d363-499a-9aec-6f1241060e4a","Type":"ContainerDied","Data":"7b02d0c26a1689a33e95a508b3d68658adbc816043a4e267e37598c61d267475"} Oct 02 03:49:01 crc kubenswrapper[4775]: I1002 03:49:01.909841 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b02d0c26a1689a33e95a508b3d68658adbc816043a4e267e37598c61d267475" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.011486 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-bzmkw"] Oct 02 03:49:02 crc kubenswrapper[4775]: E1002 03:49:02.013087 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e299556-d363-499a-9aec-6f1241060e4a" containerName="configure-os-openstack-openstack-cell1" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.013109 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e299556-d363-499a-9aec-6f1241060e4a" containerName="configure-os-openstack-openstack-cell1" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.013316 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e299556-d363-499a-9aec-6f1241060e4a" containerName="configure-os-openstack-openstack-cell1" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.014075 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.017707 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.018151 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.018196 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.018234 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.028136 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-bzmkw"] Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.085919 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-ceph\") pod \"ssh-known-hosts-openstack-bzmkw\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.188484 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpqj9\" (UniqueName: \"kubernetes.io/projected/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-kube-api-access-mpqj9\") pod \"ssh-known-hosts-openstack-bzmkw\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.188764 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-inventory-0\") pod \"ssh-known-hosts-openstack-bzmkw\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.188916 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-ceph\") pod \"ssh-known-hosts-openstack-bzmkw\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.189204 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-bzmkw\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.195894 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-ceph\") pod \"ssh-known-hosts-openstack-bzmkw\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.291636 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-inventory-0\") pod \"ssh-known-hosts-openstack-bzmkw\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.292054 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-bzmkw\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.292345 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpqj9\" (UniqueName: \"kubernetes.io/projected/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-kube-api-access-mpqj9\") pod \"ssh-known-hosts-openstack-bzmkw\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.295420 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-bzmkw\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.295757 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-inventory-0\") pod \"ssh-known-hosts-openstack-bzmkw\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.310870 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpqj9\" (UniqueName: \"kubernetes.io/projected/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-kube-api-access-mpqj9\") pod \"ssh-known-hosts-openstack-bzmkw\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.338905 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:02 crc kubenswrapper[4775]: I1002 03:49:02.958333 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-bzmkw"] Oct 02 03:49:03 crc kubenswrapper[4775]: I1002 03:49:03.931149 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-bzmkw" event={"ID":"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd","Type":"ContainerStarted","Data":"96eb236e404192aa75600ee6e1d78ce6a97e5271cca07bff26c5584788e434fe"} Oct 02 03:49:03 crc kubenswrapper[4775]: I1002 03:49:03.931646 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-bzmkw" event={"ID":"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd","Type":"ContainerStarted","Data":"f30e8a4bf0bbda2f53af895c74d11e17150607287262f835b5c5108da6ccec82"} Oct 02 03:49:03 crc kubenswrapper[4775]: I1002 03:49:03.948530 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-bzmkw" podStartSLOduration=2.363675041 podStartE2EDuration="2.948503089s" podCreationTimestamp="2025-10-02 03:49:01 +0000 UTC" firstStartedPulling="2025-10-02 03:49:02.978026586 +0000 UTC m=+7680.144770626" lastFinishedPulling="2025-10-02 03:49:03.562854594 +0000 UTC m=+7680.729598674" observedRunningTime="2025-10-02 03:49:03.944076205 +0000 UTC m=+7681.110820265" watchObservedRunningTime="2025-10-02 03:49:03.948503089 +0000 UTC m=+7681.115247139" Oct 02 03:49:07 crc kubenswrapper[4775]: I1002 03:49:07.234125 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:49:07 crc kubenswrapper[4775]: I1002 03:49:07.235030 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:49:13 crc kubenswrapper[4775]: I1002 03:49:13.050142 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-bzmkw" event={"ID":"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd","Type":"ContainerDied","Data":"96eb236e404192aa75600ee6e1d78ce6a97e5271cca07bff26c5584788e434fe"} Oct 02 03:49:13 crc kubenswrapper[4775]: I1002 03:49:13.052177 4775 generic.go:334] "Generic (PLEG): container finished" podID="cb54d64b-0c87-401c-9e60-ea2c16ca0bfd" containerID="96eb236e404192aa75600ee6e1d78ce6a97e5271cca07bff26c5584788e434fe" exitCode=0 Oct 02 03:49:14 crc kubenswrapper[4775]: I1002 03:49:14.769011 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:14 crc kubenswrapper[4775]: I1002 03:49:14.940899 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-inventory-0\") pod \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " Oct 02 03:49:14 crc kubenswrapper[4775]: I1002 03:49:14.941037 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-ssh-key-openstack-cell1\") pod \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " Oct 02 03:49:14 crc kubenswrapper[4775]: I1002 03:49:14.941080 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mpqj9\" (UniqueName: \"kubernetes.io/projected/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-kube-api-access-mpqj9\") pod \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " Oct 02 03:49:14 crc kubenswrapper[4775]: I1002 03:49:14.941124 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-ceph\") pod \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\" (UID: \"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd\") " Oct 02 03:49:14 crc kubenswrapper[4775]: I1002 03:49:14.946459 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-kube-api-access-mpqj9" (OuterVolumeSpecName: "kube-api-access-mpqj9") pod "cb54d64b-0c87-401c-9e60-ea2c16ca0bfd" (UID: "cb54d64b-0c87-401c-9e60-ea2c16ca0bfd"). InnerVolumeSpecName "kube-api-access-mpqj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:49:14 crc kubenswrapper[4775]: I1002 03:49:14.946844 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-ceph" (OuterVolumeSpecName: "ceph") pod "cb54d64b-0c87-401c-9e60-ea2c16ca0bfd" (UID: "cb54d64b-0c87-401c-9e60-ea2c16ca0bfd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:49:14 crc kubenswrapper[4775]: I1002 03:49:14.979157 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "cb54d64b-0c87-401c-9e60-ea2c16ca0bfd" (UID: "cb54d64b-0c87-401c-9e60-ea2c16ca0bfd"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:49:14 crc kubenswrapper[4775]: I1002 03:49:14.983866 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "cb54d64b-0c87-401c-9e60-ea2c16ca0bfd" (UID: "cb54d64b-0c87-401c-9e60-ea2c16ca0bfd"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.043908 4775 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.044024 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.044043 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mpqj9\" (UniqueName: \"kubernetes.io/projected/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-kube-api-access-mpqj9\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.044056 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/cb54d64b-0c87-401c-9e60-ea2c16ca0bfd-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.078835 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-bzmkw" event={"ID":"cb54d64b-0c87-401c-9e60-ea2c16ca0bfd","Type":"ContainerDied","Data":"f30e8a4bf0bbda2f53af895c74d11e17150607287262f835b5c5108da6ccec82"} Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.078890 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-bzmkw" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.078896 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f30e8a4bf0bbda2f53af895c74d11e17150607287262f835b5c5108da6ccec82" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.166031 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-679nv"] Oct 02 03:49:15 crc kubenswrapper[4775]: E1002 03:49:15.166789 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb54d64b-0c87-401c-9e60-ea2c16ca0bfd" containerName="ssh-known-hosts-openstack" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.166823 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb54d64b-0c87-401c-9e60-ea2c16ca0bfd" containerName="ssh-known-hosts-openstack" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.167323 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb54d64b-0c87-401c-9e60-ea2c16ca0bfd" containerName="ssh-known-hosts-openstack" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.168696 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.172549 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.172947 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.173031 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.173682 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.179934 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-679nv"] Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.350169 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ldxg\" (UniqueName: \"kubernetes.io/projected/79906cc8-51c5-4aaa-a552-15d2770dc621-kube-api-access-8ldxg\") pod \"run-os-openstack-openstack-cell1-679nv\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.350226 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-inventory\") pod \"run-os-openstack-openstack-cell1-679nv\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.350299 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-ssh-key\") pod \"run-os-openstack-openstack-cell1-679nv\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.350603 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-ceph\") pod \"run-os-openstack-openstack-cell1-679nv\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.453433 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ldxg\" (UniqueName: \"kubernetes.io/projected/79906cc8-51c5-4aaa-a552-15d2770dc621-kube-api-access-8ldxg\") pod \"run-os-openstack-openstack-cell1-679nv\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.453503 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-inventory\") pod \"run-os-openstack-openstack-cell1-679nv\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.453566 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-ssh-key\") pod \"run-os-openstack-openstack-cell1-679nv\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.453793 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-ceph\") pod \"run-os-openstack-openstack-cell1-679nv\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.460105 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-ceph\") pod \"run-os-openstack-openstack-cell1-679nv\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.466387 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-ssh-key\") pod \"run-os-openstack-openstack-cell1-679nv\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.467111 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-inventory\") pod \"run-os-openstack-openstack-cell1-679nv\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.471451 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ldxg\" (UniqueName: \"kubernetes.io/projected/79906cc8-51c5-4aaa-a552-15d2770dc621-kube-api-access-8ldxg\") pod \"run-os-openstack-openstack-cell1-679nv\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:15 crc kubenswrapper[4775]: I1002 03:49:15.507744 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:16 crc kubenswrapper[4775]: I1002 03:49:16.149467 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-679nv"] Oct 02 03:49:16 crc kubenswrapper[4775]: W1002 03:49:16.158717 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79906cc8_51c5_4aaa_a552_15d2770dc621.slice/crio-80ced12659c9f650e187141ff4c4fed003a269b436b8f81d8ee5c2e0fffdf927 WatchSource:0}: Error finding container 80ced12659c9f650e187141ff4c4fed003a269b436b8f81d8ee5c2e0fffdf927: Status 404 returned error can't find the container with id 80ced12659c9f650e187141ff4c4fed003a269b436b8f81d8ee5c2e0fffdf927 Oct 02 03:49:17 crc kubenswrapper[4775]: I1002 03:49:17.113223 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-679nv" event={"ID":"79906cc8-51c5-4aaa-a552-15d2770dc621","Type":"ContainerStarted","Data":"1207ff37595070fe7bba158dcb72b03fa5e4ef45a4945b4a0cea863afb2c983f"} Oct 02 03:49:17 crc kubenswrapper[4775]: I1002 03:49:17.113779 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-679nv" event={"ID":"79906cc8-51c5-4aaa-a552-15d2770dc621","Type":"ContainerStarted","Data":"80ced12659c9f650e187141ff4c4fed003a269b436b8f81d8ee5c2e0fffdf927"} Oct 02 03:49:17 crc kubenswrapper[4775]: I1002 03:49:17.151039 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-679nv" podStartSLOduration=1.654831747 podStartE2EDuration="2.151023157s" podCreationTimestamp="2025-10-02 03:49:15 +0000 UTC" firstStartedPulling="2025-10-02 03:49:16.162063046 +0000 UTC m=+7693.328807096" lastFinishedPulling="2025-10-02 03:49:16.658254456 +0000 UTC m=+7693.824998506" observedRunningTime="2025-10-02 03:49:17.139925981 +0000 UTC m=+7694.306670061" watchObservedRunningTime="2025-10-02 03:49:17.151023157 +0000 UTC m=+7694.317767207" Oct 02 03:49:25 crc kubenswrapper[4775]: I1002 03:49:25.212596 4775 generic.go:334] "Generic (PLEG): container finished" podID="79906cc8-51c5-4aaa-a552-15d2770dc621" containerID="1207ff37595070fe7bba158dcb72b03fa5e4ef45a4945b4a0cea863afb2c983f" exitCode=0 Oct 02 03:49:25 crc kubenswrapper[4775]: I1002 03:49:25.213401 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-679nv" event={"ID":"79906cc8-51c5-4aaa-a552-15d2770dc621","Type":"ContainerDied","Data":"1207ff37595070fe7bba158dcb72b03fa5e4ef45a4945b4a0cea863afb2c983f"} Oct 02 03:49:26 crc kubenswrapper[4775]: I1002 03:49:26.817236 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:26 crc kubenswrapper[4775]: I1002 03:49:26.949798 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-inventory\") pod \"79906cc8-51c5-4aaa-a552-15d2770dc621\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " Oct 02 03:49:26 crc kubenswrapper[4775]: I1002 03:49:26.950098 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-ssh-key\") pod \"79906cc8-51c5-4aaa-a552-15d2770dc621\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " Oct 02 03:49:26 crc kubenswrapper[4775]: I1002 03:49:26.950172 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ldxg\" (UniqueName: \"kubernetes.io/projected/79906cc8-51c5-4aaa-a552-15d2770dc621-kube-api-access-8ldxg\") pod \"79906cc8-51c5-4aaa-a552-15d2770dc621\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " Oct 02 03:49:26 crc kubenswrapper[4775]: I1002 03:49:26.950214 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-ceph\") pod \"79906cc8-51c5-4aaa-a552-15d2770dc621\" (UID: \"79906cc8-51c5-4aaa-a552-15d2770dc621\") " Oct 02 03:49:26 crc kubenswrapper[4775]: I1002 03:49:26.956204 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79906cc8-51c5-4aaa-a552-15d2770dc621-kube-api-access-8ldxg" (OuterVolumeSpecName: "kube-api-access-8ldxg") pod "79906cc8-51c5-4aaa-a552-15d2770dc621" (UID: "79906cc8-51c5-4aaa-a552-15d2770dc621"). InnerVolumeSpecName "kube-api-access-8ldxg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:49:26 crc kubenswrapper[4775]: I1002 03:49:26.956678 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-ceph" (OuterVolumeSpecName: "ceph") pod "79906cc8-51c5-4aaa-a552-15d2770dc621" (UID: "79906cc8-51c5-4aaa-a552-15d2770dc621"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:49:26 crc kubenswrapper[4775]: I1002 03:49:26.987392 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "79906cc8-51c5-4aaa-a552-15d2770dc621" (UID: "79906cc8-51c5-4aaa-a552-15d2770dc621"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.006846 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-inventory" (OuterVolumeSpecName: "inventory") pod "79906cc8-51c5-4aaa-a552-15d2770dc621" (UID: "79906cc8-51c5-4aaa-a552-15d2770dc621"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.053028 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.053062 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ldxg\" (UniqueName: \"kubernetes.io/projected/79906cc8-51c5-4aaa-a552-15d2770dc621-kube-api-access-8ldxg\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.053075 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.053087 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/79906cc8-51c5-4aaa-a552-15d2770dc621-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.237620 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-679nv" event={"ID":"79906cc8-51c5-4aaa-a552-15d2770dc621","Type":"ContainerDied","Data":"80ced12659c9f650e187141ff4c4fed003a269b436b8f81d8ee5c2e0fffdf927"} Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.237653 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-679nv" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.237672 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80ced12659c9f650e187141ff4c4fed003a269b436b8f81d8ee5c2e0fffdf927" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.344389 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-x4bp2"] Oct 02 03:49:27 crc kubenswrapper[4775]: E1002 03:49:27.344994 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79906cc8-51c5-4aaa-a552-15d2770dc621" containerName="run-os-openstack-openstack-cell1" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.345014 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="79906cc8-51c5-4aaa-a552-15d2770dc621" containerName="run-os-openstack-openstack-cell1" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.345317 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="79906cc8-51c5-4aaa-a552-15d2770dc621" containerName="run-os-openstack-openstack-cell1" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.346177 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.348415 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.349621 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.349847 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.350073 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.357991 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-x4bp2"] Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.460949 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-inventory\") pod \"reboot-os-openstack-openstack-cell1-x4bp2\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.461020 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-x4bp2\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.461087 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75cjz\" (UniqueName: \"kubernetes.io/projected/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-kube-api-access-75cjz\") pod \"reboot-os-openstack-openstack-cell1-x4bp2\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.461311 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-ceph\") pod \"reboot-os-openstack-openstack-cell1-x4bp2\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.563419 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-inventory\") pod \"reboot-os-openstack-openstack-cell1-x4bp2\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.563472 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-x4bp2\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.563528 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75cjz\" (UniqueName: \"kubernetes.io/projected/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-kube-api-access-75cjz\") pod \"reboot-os-openstack-openstack-cell1-x4bp2\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.563692 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-ceph\") pod \"reboot-os-openstack-openstack-cell1-x4bp2\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.567323 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-inventory\") pod \"reboot-os-openstack-openstack-cell1-x4bp2\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.567537 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-ceph\") pod \"reboot-os-openstack-openstack-cell1-x4bp2\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.570250 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-x4bp2\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.581234 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75cjz\" (UniqueName: \"kubernetes.io/projected/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-kube-api-access-75cjz\") pod \"reboot-os-openstack-openstack-cell1-x4bp2\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:27 crc kubenswrapper[4775]: I1002 03:49:27.678115 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:28 crc kubenswrapper[4775]: I1002 03:49:28.271522 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-x4bp2"] Oct 02 03:49:29 crc kubenswrapper[4775]: I1002 03:49:29.270627 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" event={"ID":"986a4be9-d34f-4204-b1cc-bb7e7ad629ec","Type":"ContainerStarted","Data":"24909c4ec9ec2ebc411c5c07db91fea55bf009303666a47d93a23d3039c3984f"} Oct 02 03:49:29 crc kubenswrapper[4775]: I1002 03:49:29.271125 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" event={"ID":"986a4be9-d34f-4204-b1cc-bb7e7ad629ec","Type":"ContainerStarted","Data":"2e355b297f0a6fa140bce206962844f1593212aef2625b60efa8fd5b8fa2f099"} Oct 02 03:49:29 crc kubenswrapper[4775]: I1002 03:49:29.296653 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" podStartSLOduration=1.77886581 podStartE2EDuration="2.296630756s" podCreationTimestamp="2025-10-02 03:49:27 +0000 UTC" firstStartedPulling="2025-10-02 03:49:28.276097881 +0000 UTC m=+7705.442841921" lastFinishedPulling="2025-10-02 03:49:28.793862827 +0000 UTC m=+7705.960606867" observedRunningTime="2025-10-02 03:49:29.291827042 +0000 UTC m=+7706.458571112" watchObservedRunningTime="2025-10-02 03:49:29.296630756 +0000 UTC m=+7706.463374836" Oct 02 03:49:37 crc kubenswrapper[4775]: I1002 03:49:37.233913 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:49:37 crc kubenswrapper[4775]: I1002 03:49:37.234498 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:49:45 crc kubenswrapper[4775]: I1002 03:49:45.487879 4775 generic.go:334] "Generic (PLEG): container finished" podID="986a4be9-d34f-4204-b1cc-bb7e7ad629ec" containerID="24909c4ec9ec2ebc411c5c07db91fea55bf009303666a47d93a23d3039c3984f" exitCode=0 Oct 02 03:49:45 crc kubenswrapper[4775]: I1002 03:49:45.488016 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" event={"ID":"986a4be9-d34f-4204-b1cc-bb7e7ad629ec","Type":"ContainerDied","Data":"24909c4ec9ec2ebc411c5c07db91fea55bf009303666a47d93a23d3039c3984f"} Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.003451 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.126593 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-ssh-key\") pod \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.126782 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-inventory\") pod \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.127025 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-ceph\") pod \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.127211 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75cjz\" (UniqueName: \"kubernetes.io/projected/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-kube-api-access-75cjz\") pod \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\" (UID: \"986a4be9-d34f-4204-b1cc-bb7e7ad629ec\") " Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.136865 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-ceph" (OuterVolumeSpecName: "ceph") pod "986a4be9-d34f-4204-b1cc-bb7e7ad629ec" (UID: "986a4be9-d34f-4204-b1cc-bb7e7ad629ec"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.137905 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-kube-api-access-75cjz" (OuterVolumeSpecName: "kube-api-access-75cjz") pod "986a4be9-d34f-4204-b1cc-bb7e7ad629ec" (UID: "986a4be9-d34f-4204-b1cc-bb7e7ad629ec"). InnerVolumeSpecName "kube-api-access-75cjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.169134 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "986a4be9-d34f-4204-b1cc-bb7e7ad629ec" (UID: "986a4be9-d34f-4204-b1cc-bb7e7ad629ec"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.185552 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-inventory" (OuterVolumeSpecName: "inventory") pod "986a4be9-d34f-4204-b1cc-bb7e7ad629ec" (UID: "986a4be9-d34f-4204-b1cc-bb7e7ad629ec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.230466 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.230515 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.230534 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75cjz\" (UniqueName: \"kubernetes.io/projected/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-kube-api-access-75cjz\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.230554 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/986a4be9-d34f-4204-b1cc-bb7e7ad629ec-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.515287 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" event={"ID":"986a4be9-d34f-4204-b1cc-bb7e7ad629ec","Type":"ContainerDied","Data":"2e355b297f0a6fa140bce206962844f1593212aef2625b60efa8fd5b8fa2f099"} Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.515359 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e355b297f0a6fa140bce206962844f1593212aef2625b60efa8fd5b8fa2f099" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.515374 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-x4bp2" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.649842 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-shjt8"] Oct 02 03:49:47 crc kubenswrapper[4775]: E1002 03:49:47.650647 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="986a4be9-d34f-4204-b1cc-bb7e7ad629ec" containerName="reboot-os-openstack-openstack-cell1" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.650677 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="986a4be9-d34f-4204-b1cc-bb7e7ad629ec" containerName="reboot-os-openstack-openstack-cell1" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.651107 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="986a4be9-d34f-4204-b1cc-bb7e7ad629ec" containerName="reboot-os-openstack-openstack-cell1" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.652493 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.657055 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.657573 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.657915 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.658166 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.660337 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-shjt8"] Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.742372 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.742464 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.742518 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.742552 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ceph\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.742583 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.742620 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.742780 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.742815 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-inventory\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.742846 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.742991 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.743137 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsgtp\" (UniqueName: \"kubernetes.io/projected/5e735a83-11ab-4b9d-b97c-a60174da0899-kube-api-access-jsgtp\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.743204 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ssh-key\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.845668 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.845762 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.845882 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsgtp\" (UniqueName: \"kubernetes.io/projected/5e735a83-11ab-4b9d-b97c-a60174da0899-kube-api-access-jsgtp\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.845939 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ssh-key\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.846080 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.849129 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.849619 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.849668 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ceph\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.849704 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.849754 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.849788 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.850069 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.850121 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-inventory\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.850441 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ssh-key\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.853471 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.853471 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.854032 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.854323 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-inventory\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.854528 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.855765 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.856289 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.856517 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ceph\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.856865 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.866117 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsgtp\" (UniqueName: \"kubernetes.io/projected/5e735a83-11ab-4b9d-b97c-a60174da0899-kube-api-access-jsgtp\") pod \"install-certs-openstack-openstack-cell1-shjt8\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:47 crc kubenswrapper[4775]: I1002 03:49:47.977316 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:49:48 crc kubenswrapper[4775]: I1002 03:49:48.399103 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-shjt8"] Oct 02 03:49:48 crc kubenswrapper[4775]: I1002 03:49:48.525364 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-shjt8" event={"ID":"5e735a83-11ab-4b9d-b97c-a60174da0899","Type":"ContainerStarted","Data":"d2adcdd2d7d09b3f57bd94cdc130dbb9b2cd220940b3df097b2d855f68c9bab3"} Oct 02 03:49:49 crc kubenswrapper[4775]: I1002 03:49:49.538315 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-shjt8" event={"ID":"5e735a83-11ab-4b9d-b97c-a60174da0899","Type":"ContainerStarted","Data":"c6cd5b478bd77764caf22a2b549d3f2189f4b77dd8dec8b7f65a5c1c33f50b05"} Oct 02 03:49:49 crc kubenswrapper[4775]: I1002 03:49:49.569117 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-shjt8" podStartSLOduration=2.111375509 podStartE2EDuration="2.569101105s" podCreationTimestamp="2025-10-02 03:49:47 +0000 UTC" firstStartedPulling="2025-10-02 03:49:48.399180774 +0000 UTC m=+7725.565924804" lastFinishedPulling="2025-10-02 03:49:48.85690635 +0000 UTC m=+7726.023650400" observedRunningTime="2025-10-02 03:49:49.561380706 +0000 UTC m=+7726.728124746" watchObservedRunningTime="2025-10-02 03:49:49.569101105 +0000 UTC m=+7726.735845135" Oct 02 03:50:01 crc kubenswrapper[4775]: I1002 03:50:01.109161 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wqwth"] Oct 02 03:50:01 crc kubenswrapper[4775]: I1002 03:50:01.116136 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:01 crc kubenswrapper[4775]: I1002 03:50:01.122759 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wqwth"] Oct 02 03:50:01 crc kubenswrapper[4775]: I1002 03:50:01.280696 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q54mk\" (UniqueName: \"kubernetes.io/projected/6d9775f7-550a-4c3c-89b4-d0ddace4b981-kube-api-access-q54mk\") pod \"certified-operators-wqwth\" (UID: \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\") " pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:01 crc kubenswrapper[4775]: I1002 03:50:01.280766 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9775f7-550a-4c3c-89b4-d0ddace4b981-utilities\") pod \"certified-operators-wqwth\" (UID: \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\") " pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:01 crc kubenswrapper[4775]: I1002 03:50:01.281018 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9775f7-550a-4c3c-89b4-d0ddace4b981-catalog-content\") pod \"certified-operators-wqwth\" (UID: \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\") " pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:01 crc kubenswrapper[4775]: I1002 03:50:01.383292 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9775f7-550a-4c3c-89b4-d0ddace4b981-catalog-content\") pod \"certified-operators-wqwth\" (UID: \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\") " pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:01 crc kubenswrapper[4775]: I1002 03:50:01.383501 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q54mk\" (UniqueName: \"kubernetes.io/projected/6d9775f7-550a-4c3c-89b4-d0ddace4b981-kube-api-access-q54mk\") pod \"certified-operators-wqwth\" (UID: \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\") " pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:01 crc kubenswrapper[4775]: I1002 03:50:01.383532 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9775f7-550a-4c3c-89b4-d0ddace4b981-utilities\") pod \"certified-operators-wqwth\" (UID: \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\") " pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:01 crc kubenswrapper[4775]: I1002 03:50:01.384110 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9775f7-550a-4c3c-89b4-d0ddace4b981-utilities\") pod \"certified-operators-wqwth\" (UID: \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\") " pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:01 crc kubenswrapper[4775]: I1002 03:50:01.384115 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9775f7-550a-4c3c-89b4-d0ddace4b981-catalog-content\") pod \"certified-operators-wqwth\" (UID: \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\") " pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:01 crc kubenswrapper[4775]: I1002 03:50:01.402994 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q54mk\" (UniqueName: \"kubernetes.io/projected/6d9775f7-550a-4c3c-89b4-d0ddace4b981-kube-api-access-q54mk\") pod \"certified-operators-wqwth\" (UID: \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\") " pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:01 crc kubenswrapper[4775]: I1002 03:50:01.488580 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:02 crc kubenswrapper[4775]: I1002 03:50:02.070779 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wqwth"] Oct 02 03:50:02 crc kubenswrapper[4775]: I1002 03:50:02.715026 4775 generic.go:334] "Generic (PLEG): container finished" podID="6d9775f7-550a-4c3c-89b4-d0ddace4b981" containerID="3596dd33bdffe5c781ce72c2634bce08ffa03b3d1279bd40ba9540b2fabd78a4" exitCode=0 Oct 02 03:50:02 crc kubenswrapper[4775]: I1002 03:50:02.715123 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqwth" event={"ID":"6d9775f7-550a-4c3c-89b4-d0ddace4b981","Type":"ContainerDied","Data":"3596dd33bdffe5c781ce72c2634bce08ffa03b3d1279bd40ba9540b2fabd78a4"} Oct 02 03:50:02 crc kubenswrapper[4775]: I1002 03:50:02.715493 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqwth" event={"ID":"6d9775f7-550a-4c3c-89b4-d0ddace4b981","Type":"ContainerStarted","Data":"4b3c1c75553c6089ca62e39ead383d94308d82a88e3365240901bb6bc10fac90"} Oct 02 03:50:03 crc kubenswrapper[4775]: I1002 03:50:03.731608 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqwth" event={"ID":"6d9775f7-550a-4c3c-89b4-d0ddace4b981","Type":"ContainerStarted","Data":"e4d649dd86199b52dbc7c094a0ba1620cb5e2bf140c13cbc6803dbd7b90345d7"} Oct 02 03:50:05 crc kubenswrapper[4775]: I1002 03:50:05.835345 4775 generic.go:334] "Generic (PLEG): container finished" podID="6d9775f7-550a-4c3c-89b4-d0ddace4b981" containerID="e4d649dd86199b52dbc7c094a0ba1620cb5e2bf140c13cbc6803dbd7b90345d7" exitCode=0 Oct 02 03:50:05 crc kubenswrapper[4775]: I1002 03:50:05.835451 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqwth" event={"ID":"6d9775f7-550a-4c3c-89b4-d0ddace4b981","Type":"ContainerDied","Data":"e4d649dd86199b52dbc7c094a0ba1620cb5e2bf140c13cbc6803dbd7b90345d7"} Oct 02 03:50:06 crc kubenswrapper[4775]: I1002 03:50:06.847449 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqwth" event={"ID":"6d9775f7-550a-4c3c-89b4-d0ddace4b981","Type":"ContainerStarted","Data":"478a8c74516f1ba5862e1a1dbf552902edd4546bd7dfe709bd123019d11c6eaa"} Oct 02 03:50:06 crc kubenswrapper[4775]: I1002 03:50:06.877776 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wqwth" podStartSLOduration=2.225951178 podStartE2EDuration="5.877759041s" podCreationTimestamp="2025-10-02 03:50:01 +0000 UTC" firstStartedPulling="2025-10-02 03:50:02.716476796 +0000 UTC m=+7739.883220856" lastFinishedPulling="2025-10-02 03:50:06.368284649 +0000 UTC m=+7743.535028719" observedRunningTime="2025-10-02 03:50:06.870151535 +0000 UTC m=+7744.036895585" watchObservedRunningTime="2025-10-02 03:50:06.877759041 +0000 UTC m=+7744.044503081" Oct 02 03:50:07 crc kubenswrapper[4775]: I1002 03:50:07.239363 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:50:07 crc kubenswrapper[4775]: I1002 03:50:07.239439 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:50:07 crc kubenswrapper[4775]: I1002 03:50:07.239501 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 03:50:07 crc kubenswrapper[4775]: I1002 03:50:07.240993 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 03:50:07 crc kubenswrapper[4775]: I1002 03:50:07.241132 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" gracePeriod=600 Oct 02 03:50:07 crc kubenswrapper[4775]: E1002 03:50:07.386407 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:50:07 crc kubenswrapper[4775]: I1002 03:50:07.862024 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" exitCode=0 Oct 02 03:50:07 crc kubenswrapper[4775]: I1002 03:50:07.862071 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4"} Oct 02 03:50:07 crc kubenswrapper[4775]: I1002 03:50:07.862108 4775 scope.go:117] "RemoveContainer" containerID="82e9bf60a688a73f4e67e12ca66bd36d7e6ce4ea01f01845c9edec604f41bbff" Oct 02 03:50:07 crc kubenswrapper[4775]: I1002 03:50:07.862639 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:50:07 crc kubenswrapper[4775]: E1002 03:50:07.863033 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:50:09 crc kubenswrapper[4775]: I1002 03:50:09.888759 4775 generic.go:334] "Generic (PLEG): container finished" podID="5e735a83-11ab-4b9d-b97c-a60174da0899" containerID="c6cd5b478bd77764caf22a2b549d3f2189f4b77dd8dec8b7f65a5c1c33f50b05" exitCode=0 Oct 02 03:50:09 crc kubenswrapper[4775]: I1002 03:50:09.888841 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-shjt8" event={"ID":"5e735a83-11ab-4b9d-b97c-a60174da0899","Type":"ContainerDied","Data":"c6cd5b478bd77764caf22a2b549d3f2189f4b77dd8dec8b7f65a5c1c33f50b05"} Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.489528 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.489934 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.530094 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.559266 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.592338 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsgtp\" (UniqueName: \"kubernetes.io/projected/5e735a83-11ab-4b9d-b97c-a60174da0899-kube-api-access-jsgtp\") pod \"5e735a83-11ab-4b9d-b97c-a60174da0899\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.592413 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-dhcp-combined-ca-bundle\") pod \"5e735a83-11ab-4b9d-b97c-a60174da0899\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.592440 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-inventory\") pod \"5e735a83-11ab-4b9d-b97c-a60174da0899\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.592469 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-nova-combined-ca-bundle\") pod \"5e735a83-11ab-4b9d-b97c-a60174da0899\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.592495 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ceph\") pod \"5e735a83-11ab-4b9d-b97c-a60174da0899\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.592518 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ovn-combined-ca-bundle\") pod \"5e735a83-11ab-4b9d-b97c-a60174da0899\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.592548 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-bootstrap-combined-ca-bundle\") pod \"5e735a83-11ab-4b9d-b97c-a60174da0899\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.592595 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ssh-key\") pod \"5e735a83-11ab-4b9d-b97c-a60174da0899\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.592623 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-metadata-combined-ca-bundle\") pod \"5e735a83-11ab-4b9d-b97c-a60174da0899\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.592660 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-libvirt-combined-ca-bundle\") pod \"5e735a83-11ab-4b9d-b97c-a60174da0899\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.592704 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-telemetry-combined-ca-bundle\") pod \"5e735a83-11ab-4b9d-b97c-a60174da0899\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.592737 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-sriov-combined-ca-bundle\") pod \"5e735a83-11ab-4b9d-b97c-a60174da0899\" (UID: \"5e735a83-11ab-4b9d-b97c-a60174da0899\") " Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.602226 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "5e735a83-11ab-4b9d-b97c-a60174da0899" (UID: "5e735a83-11ab-4b9d-b97c-a60174da0899"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.602307 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "5e735a83-11ab-4b9d-b97c-a60174da0899" (UID: "5e735a83-11ab-4b9d-b97c-a60174da0899"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.602344 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e735a83-11ab-4b9d-b97c-a60174da0899-kube-api-access-jsgtp" (OuterVolumeSpecName: "kube-api-access-jsgtp") pod "5e735a83-11ab-4b9d-b97c-a60174da0899" (UID: "5e735a83-11ab-4b9d-b97c-a60174da0899"). InnerVolumeSpecName "kube-api-access-jsgtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.602351 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "5e735a83-11ab-4b9d-b97c-a60174da0899" (UID: "5e735a83-11ab-4b9d-b97c-a60174da0899"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.602386 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "5e735a83-11ab-4b9d-b97c-a60174da0899" (UID: "5e735a83-11ab-4b9d-b97c-a60174da0899"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.602872 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "5e735a83-11ab-4b9d-b97c-a60174da0899" (UID: "5e735a83-11ab-4b9d-b97c-a60174da0899"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.604458 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "5e735a83-11ab-4b9d-b97c-a60174da0899" (UID: "5e735a83-11ab-4b9d-b97c-a60174da0899"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.605427 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "5e735a83-11ab-4b9d-b97c-a60174da0899" (UID: "5e735a83-11ab-4b9d-b97c-a60174da0899"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.606748 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ceph" (OuterVolumeSpecName: "ceph") pod "5e735a83-11ab-4b9d-b97c-a60174da0899" (UID: "5e735a83-11ab-4b9d-b97c-a60174da0899"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.608560 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "5e735a83-11ab-4b9d-b97c-a60174da0899" (UID: "5e735a83-11ab-4b9d-b97c-a60174da0899"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.634042 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-inventory" (OuterVolumeSpecName: "inventory") pod "5e735a83-11ab-4b9d-b97c-a60174da0899" (UID: "5e735a83-11ab-4b9d-b97c-a60174da0899"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.645477 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5e735a83-11ab-4b9d-b97c-a60174da0899" (UID: "5e735a83-11ab-4b9d-b97c-a60174da0899"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.697251 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.697295 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.697312 4775 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.697324 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.697335 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.697347 4775 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.697360 4775 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.697372 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.697385 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsgtp\" (UniqueName: \"kubernetes.io/projected/5e735a83-11ab-4b9d-b97c-a60174da0899-kube-api-access-jsgtp\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.697398 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.697413 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.697427 4775 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e735a83-11ab-4b9d-b97c-a60174da0899-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.920585 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-shjt8" event={"ID":"5e735a83-11ab-4b9d-b97c-a60174da0899","Type":"ContainerDied","Data":"d2adcdd2d7d09b3f57bd94cdc130dbb9b2cd220940b3df097b2d855f68c9bab3"} Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.920652 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2adcdd2d7d09b3f57bd94cdc130dbb9b2cd220940b3df097b2d855f68c9bab3" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.920598 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-shjt8" Oct 02 03:50:11 crc kubenswrapper[4775]: I1002 03:50:11.988213 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.146140 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-r82b9"] Oct 02 03:50:12 crc kubenswrapper[4775]: E1002 03:50:12.147068 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e735a83-11ab-4b9d-b97c-a60174da0899" containerName="install-certs-openstack-openstack-cell1" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.147094 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e735a83-11ab-4b9d-b97c-a60174da0899" containerName="install-certs-openstack-openstack-cell1" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.147344 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e735a83-11ab-4b9d-b97c-a60174da0899" containerName="install-certs-openstack-openstack-cell1" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.148320 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.150496 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.152769 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.153585 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.153718 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.167402 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-r82b9"] Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.212099 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-ceph\") pod \"ceph-client-openstack-openstack-cell1-r82b9\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.212182 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-r82b9\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.212288 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5gcm\" (UniqueName: \"kubernetes.io/projected/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-kube-api-access-n5gcm\") pod \"ceph-client-openstack-openstack-cell1-r82b9\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.212368 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-inventory\") pod \"ceph-client-openstack-openstack-cell1-r82b9\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.313437 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5gcm\" (UniqueName: \"kubernetes.io/projected/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-kube-api-access-n5gcm\") pod \"ceph-client-openstack-openstack-cell1-r82b9\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.313551 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-inventory\") pod \"ceph-client-openstack-openstack-cell1-r82b9\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.313790 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-ceph\") pod \"ceph-client-openstack-openstack-cell1-r82b9\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.313850 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-r82b9\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.321927 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-inventory\") pod \"ceph-client-openstack-openstack-cell1-r82b9\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.322476 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-r82b9\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.323007 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-ceph\") pod \"ceph-client-openstack-openstack-cell1-r82b9\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.333637 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5gcm\" (UniqueName: \"kubernetes.io/projected/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-kube-api-access-n5gcm\") pod \"ceph-client-openstack-openstack-cell1-r82b9\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.503984 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.909944 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-r82b9"] Oct 02 03:50:12 crc kubenswrapper[4775]: I1002 03:50:12.935490 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" event={"ID":"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58","Type":"ContainerStarted","Data":"0283dcab98491c3ad90858c1ca042d400e5ae427a3bcc510540e4c650324f947"} Oct 02 03:50:13 crc kubenswrapper[4775]: I1002 03:50:13.690689 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wqwth"] Oct 02 03:50:13 crc kubenswrapper[4775]: I1002 03:50:13.951019 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" event={"ID":"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58","Type":"ContainerStarted","Data":"09d0e26b0f89571b277daa043db978b80d9f970bb84ae5e2bbb8be99a863885c"} Oct 02 03:50:13 crc kubenswrapper[4775]: I1002 03:50:13.951193 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wqwth" podUID="6d9775f7-550a-4c3c-89b4-d0ddace4b981" containerName="registry-server" containerID="cri-o://478a8c74516f1ba5862e1a1dbf552902edd4546bd7dfe709bd123019d11c6eaa" gracePeriod=2 Oct 02 03:50:13 crc kubenswrapper[4775]: I1002 03:50:13.997545 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" podStartSLOduration=1.521940252 podStartE2EDuration="1.997443897s" podCreationTimestamp="2025-10-02 03:50:12 +0000 UTC" firstStartedPulling="2025-10-02 03:50:12.895848859 +0000 UTC m=+7750.062592919" lastFinishedPulling="2025-10-02 03:50:13.371352514 +0000 UTC m=+7750.538096564" observedRunningTime="2025-10-02 03:50:13.985445208 +0000 UTC m=+7751.152189278" watchObservedRunningTime="2025-10-02 03:50:13.997443897 +0000 UTC m=+7751.164187977" Oct 02 03:50:14 crc kubenswrapper[4775]: I1002 03:50:14.962457 4775 generic.go:334] "Generic (PLEG): container finished" podID="6d9775f7-550a-4c3c-89b4-d0ddace4b981" containerID="478a8c74516f1ba5862e1a1dbf552902edd4546bd7dfe709bd123019d11c6eaa" exitCode=0 Oct 02 03:50:14 crc kubenswrapper[4775]: I1002 03:50:14.962541 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqwth" event={"ID":"6d9775f7-550a-4c3c-89b4-d0ddace4b981","Type":"ContainerDied","Data":"478a8c74516f1ba5862e1a1dbf552902edd4546bd7dfe709bd123019d11c6eaa"} Oct 02 03:50:14 crc kubenswrapper[4775]: I1002 03:50:14.963249 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wqwth" event={"ID":"6d9775f7-550a-4c3c-89b4-d0ddace4b981","Type":"ContainerDied","Data":"4b3c1c75553c6089ca62e39ead383d94308d82a88e3365240901bb6bc10fac90"} Oct 02 03:50:14 crc kubenswrapper[4775]: I1002 03:50:14.963283 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b3c1c75553c6089ca62e39ead383d94308d82a88e3365240901bb6bc10fac90" Oct 02 03:50:14 crc kubenswrapper[4775]: I1002 03:50:14.981422 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:15 crc kubenswrapper[4775]: I1002 03:50:15.181604 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9775f7-550a-4c3c-89b4-d0ddace4b981-catalog-content\") pod \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\" (UID: \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\") " Oct 02 03:50:15 crc kubenswrapper[4775]: I1002 03:50:15.181762 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q54mk\" (UniqueName: \"kubernetes.io/projected/6d9775f7-550a-4c3c-89b4-d0ddace4b981-kube-api-access-q54mk\") pod \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\" (UID: \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\") " Oct 02 03:50:15 crc kubenswrapper[4775]: I1002 03:50:15.182116 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9775f7-550a-4c3c-89b4-d0ddace4b981-utilities\") pod \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\" (UID: \"6d9775f7-550a-4c3c-89b4-d0ddace4b981\") " Oct 02 03:50:15 crc kubenswrapper[4775]: I1002 03:50:15.183793 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d9775f7-550a-4c3c-89b4-d0ddace4b981-utilities" (OuterVolumeSpecName: "utilities") pod "6d9775f7-550a-4c3c-89b4-d0ddace4b981" (UID: "6d9775f7-550a-4c3c-89b4-d0ddace4b981"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:50:15 crc kubenswrapper[4775]: I1002 03:50:15.195303 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d9775f7-550a-4c3c-89b4-d0ddace4b981-kube-api-access-q54mk" (OuterVolumeSpecName: "kube-api-access-q54mk") pod "6d9775f7-550a-4c3c-89b4-d0ddace4b981" (UID: "6d9775f7-550a-4c3c-89b4-d0ddace4b981"). InnerVolumeSpecName "kube-api-access-q54mk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:50:15 crc kubenswrapper[4775]: I1002 03:50:15.251180 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d9775f7-550a-4c3c-89b4-d0ddace4b981-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6d9775f7-550a-4c3c-89b4-d0ddace4b981" (UID: "6d9775f7-550a-4c3c-89b4-d0ddace4b981"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:50:15 crc kubenswrapper[4775]: I1002 03:50:15.285681 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q54mk\" (UniqueName: \"kubernetes.io/projected/6d9775f7-550a-4c3c-89b4-d0ddace4b981-kube-api-access-q54mk\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:15 crc kubenswrapper[4775]: I1002 03:50:15.285747 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6d9775f7-550a-4c3c-89b4-d0ddace4b981-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:15 crc kubenswrapper[4775]: I1002 03:50:15.285774 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6d9775f7-550a-4c3c-89b4-d0ddace4b981-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:15 crc kubenswrapper[4775]: I1002 03:50:15.975698 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wqwth" Oct 02 03:50:16 crc kubenswrapper[4775]: I1002 03:50:16.013556 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wqwth"] Oct 02 03:50:16 crc kubenswrapper[4775]: I1002 03:50:16.021377 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wqwth"] Oct 02 03:50:17 crc kubenswrapper[4775]: I1002 03:50:17.792562 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d9775f7-550a-4c3c-89b4-d0ddace4b981" path="/var/lib/kubelet/pods/6d9775f7-550a-4c3c-89b4-d0ddace4b981/volumes" Oct 02 03:50:19 crc kubenswrapper[4775]: I1002 03:50:19.016670 4775 generic.go:334] "Generic (PLEG): container finished" podID="bb77ca06-d802-41b4-98f3-0cf0c6b5bb58" containerID="09d0e26b0f89571b277daa043db978b80d9f970bb84ae5e2bbb8be99a863885c" exitCode=0 Oct 02 03:50:19 crc kubenswrapper[4775]: I1002 03:50:19.016715 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" event={"ID":"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58","Type":"ContainerDied","Data":"09d0e26b0f89571b277daa043db978b80d9f970bb84ae5e2bbb8be99a863885c"} Oct 02 03:50:19 crc kubenswrapper[4775]: I1002 03:50:19.765561 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:50:19 crc kubenswrapper[4775]: E1002 03:50:19.766330 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:50:20 crc kubenswrapper[4775]: I1002 03:50:20.520024 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:20 crc kubenswrapper[4775]: I1002 03:50:20.628439 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5gcm\" (UniqueName: \"kubernetes.io/projected/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-kube-api-access-n5gcm\") pod \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " Oct 02 03:50:20 crc kubenswrapper[4775]: I1002 03:50:20.628551 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-ceph\") pod \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " Oct 02 03:50:20 crc kubenswrapper[4775]: I1002 03:50:20.628686 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-inventory\") pod \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " Oct 02 03:50:20 crc kubenswrapper[4775]: I1002 03:50:20.628713 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-ssh-key\") pod \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\" (UID: \"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58\") " Oct 02 03:50:20 crc kubenswrapper[4775]: I1002 03:50:20.634894 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-ceph" (OuterVolumeSpecName: "ceph") pod "bb77ca06-d802-41b4-98f3-0cf0c6b5bb58" (UID: "bb77ca06-d802-41b4-98f3-0cf0c6b5bb58"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:50:20 crc kubenswrapper[4775]: I1002 03:50:20.637979 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-kube-api-access-n5gcm" (OuterVolumeSpecName: "kube-api-access-n5gcm") pod "bb77ca06-d802-41b4-98f3-0cf0c6b5bb58" (UID: "bb77ca06-d802-41b4-98f3-0cf0c6b5bb58"). InnerVolumeSpecName "kube-api-access-n5gcm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:50:20 crc kubenswrapper[4775]: I1002 03:50:20.668133 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bb77ca06-d802-41b4-98f3-0cf0c6b5bb58" (UID: "bb77ca06-d802-41b4-98f3-0cf0c6b5bb58"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:50:20 crc kubenswrapper[4775]: I1002 03:50:20.681949 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-inventory" (OuterVolumeSpecName: "inventory") pod "bb77ca06-d802-41b4-98f3-0cf0c6b5bb58" (UID: "bb77ca06-d802-41b4-98f3-0cf0c6b5bb58"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:50:20 crc kubenswrapper[4775]: I1002 03:50:20.731997 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5gcm\" (UniqueName: \"kubernetes.io/projected/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-kube-api-access-n5gcm\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:20 crc kubenswrapper[4775]: I1002 03:50:20.732049 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:20 crc kubenswrapper[4775]: I1002 03:50:20.732068 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:20 crc kubenswrapper[4775]: I1002 03:50:20.732087 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb77ca06-d802-41b4-98f3-0cf0c6b5bb58-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.048935 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" event={"ID":"bb77ca06-d802-41b4-98f3-0cf0c6b5bb58","Type":"ContainerDied","Data":"0283dcab98491c3ad90858c1ca042d400e5ae427a3bcc510540e4c650324f947"} Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.049449 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0283dcab98491c3ad90858c1ca042d400e5ae427a3bcc510540e4c650324f947" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.049534 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-r82b9" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.161846 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-dv978"] Oct 02 03:50:21 crc kubenswrapper[4775]: E1002 03:50:21.163521 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d9775f7-550a-4c3c-89b4-d0ddace4b981" containerName="extract-utilities" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.163558 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d9775f7-550a-4c3c-89b4-d0ddace4b981" containerName="extract-utilities" Oct 02 03:50:21 crc kubenswrapper[4775]: E1002 03:50:21.163612 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d9775f7-550a-4c3c-89b4-d0ddace4b981" containerName="extract-content" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.163627 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d9775f7-550a-4c3c-89b4-d0ddace4b981" containerName="extract-content" Oct 02 03:50:21 crc kubenswrapper[4775]: E1002 03:50:21.163722 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb77ca06-d802-41b4-98f3-0cf0c6b5bb58" containerName="ceph-client-openstack-openstack-cell1" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.163737 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb77ca06-d802-41b4-98f3-0cf0c6b5bb58" containerName="ceph-client-openstack-openstack-cell1" Oct 02 03:50:21 crc kubenswrapper[4775]: E1002 03:50:21.163776 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d9775f7-550a-4c3c-89b4-d0ddace4b981" containerName="registry-server" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.163790 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d9775f7-550a-4c3c-89b4-d0ddace4b981" containerName="registry-server" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.164678 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb77ca06-d802-41b4-98f3-0cf0c6b5bb58" containerName="ceph-client-openstack-openstack-cell1" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.164760 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d9775f7-550a-4c3c-89b4-d0ddace4b981" containerName="registry-server" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.167608 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.170449 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.171462 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.171641 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.171811 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.172376 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.201791 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-dv978"] Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.244948 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ceph\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.245024 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-inventory\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.245071 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8654436c-452a-43f6-8932-55b0a85d60d2-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.245253 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.245404 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48z87\" (UniqueName: \"kubernetes.io/projected/8654436c-452a-43f6-8932-55b0a85d60d2-kube-api-access-48z87\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.245594 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ssh-key\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.347614 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8654436c-452a-43f6-8932-55b0a85d60d2-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.347729 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.347815 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48z87\" (UniqueName: \"kubernetes.io/projected/8654436c-452a-43f6-8932-55b0a85d60d2-kube-api-access-48z87\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.347939 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ssh-key\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.348297 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ceph\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.348355 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-inventory\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.348821 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8654436c-452a-43f6-8932-55b0a85d60d2-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.352300 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ssh-key\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.353069 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.353084 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ceph\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.361315 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-inventory\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.366816 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48z87\" (UniqueName: \"kubernetes.io/projected/8654436c-452a-43f6-8932-55b0a85d60d2-kube-api-access-48z87\") pod \"ovn-openstack-openstack-cell1-dv978\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:21 crc kubenswrapper[4775]: I1002 03:50:21.505828 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:50:22 crc kubenswrapper[4775]: I1002 03:50:22.096941 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-dv978"] Oct 02 03:50:22 crc kubenswrapper[4775]: W1002 03:50:22.104384 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8654436c_452a_43f6_8932_55b0a85d60d2.slice/crio-9ac69cfab85d8cf7c30b589eaa81ce00e76b60a9a727bf67d7af0b09fec861e0 WatchSource:0}: Error finding container 9ac69cfab85d8cf7c30b589eaa81ce00e76b60a9a727bf67d7af0b09fec861e0: Status 404 returned error can't find the container with id 9ac69cfab85d8cf7c30b589eaa81ce00e76b60a9a727bf67d7af0b09fec861e0 Oct 02 03:50:23 crc kubenswrapper[4775]: I1002 03:50:23.077049 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-dv978" event={"ID":"8654436c-452a-43f6-8932-55b0a85d60d2","Type":"ContainerStarted","Data":"94929080a2c050986163b2c8ed2abcfa5f8f8bff694f3d2e9acd0a927116bf59"} Oct 02 03:50:23 crc kubenswrapper[4775]: I1002 03:50:23.077509 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-dv978" event={"ID":"8654436c-452a-43f6-8932-55b0a85d60d2","Type":"ContainerStarted","Data":"9ac69cfab85d8cf7c30b589eaa81ce00e76b60a9a727bf67d7af0b09fec861e0"} Oct 02 03:50:23 crc kubenswrapper[4775]: I1002 03:50:23.118704 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-dv978" podStartSLOduration=1.5380189739999999 podStartE2EDuration="2.118669354s" podCreationTimestamp="2025-10-02 03:50:21 +0000 UTC" firstStartedPulling="2025-10-02 03:50:22.10751815 +0000 UTC m=+7759.274262200" lastFinishedPulling="2025-10-02 03:50:22.6881685 +0000 UTC m=+7759.854912580" observedRunningTime="2025-10-02 03:50:23.103081481 +0000 UTC m=+7760.269825561" watchObservedRunningTime="2025-10-02 03:50:23.118669354 +0000 UTC m=+7760.285413434" Oct 02 03:50:30 crc kubenswrapper[4775]: I1002 03:50:30.765672 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:50:30 crc kubenswrapper[4775]: E1002 03:50:30.766594 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:50:43 crc kubenswrapper[4775]: I1002 03:50:43.782048 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:50:43 crc kubenswrapper[4775]: E1002 03:50:43.782942 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:50:56 crc kubenswrapper[4775]: I1002 03:50:56.765876 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:50:56 crc kubenswrapper[4775]: E1002 03:50:56.766778 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:51:10 crc kubenswrapper[4775]: I1002 03:51:10.765870 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:51:10 crc kubenswrapper[4775]: E1002 03:51:10.766841 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:51:24 crc kubenswrapper[4775]: I1002 03:51:24.766128 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:51:24 crc kubenswrapper[4775]: E1002 03:51:24.767233 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:51:31 crc kubenswrapper[4775]: I1002 03:51:31.997998 4775 generic.go:334] "Generic (PLEG): container finished" podID="8654436c-452a-43f6-8932-55b0a85d60d2" containerID="94929080a2c050986163b2c8ed2abcfa5f8f8bff694f3d2e9acd0a927116bf59" exitCode=0 Oct 02 03:51:31 crc kubenswrapper[4775]: I1002 03:51:31.998667 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-dv978" event={"ID":"8654436c-452a-43f6-8932-55b0a85d60d2","Type":"ContainerDied","Data":"94929080a2c050986163b2c8ed2abcfa5f8f8bff694f3d2e9acd0a927116bf59"} Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.504873 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.630247 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8654436c-452a-43f6-8932-55b0a85d60d2-ovncontroller-config-0\") pod \"8654436c-452a-43f6-8932-55b0a85d60d2\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.630614 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ceph\") pod \"8654436c-452a-43f6-8932-55b0a85d60d2\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.630682 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ssh-key\") pod \"8654436c-452a-43f6-8932-55b0a85d60d2\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.630788 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-inventory\") pod \"8654436c-452a-43f6-8932-55b0a85d60d2\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.630846 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48z87\" (UniqueName: \"kubernetes.io/projected/8654436c-452a-43f6-8932-55b0a85d60d2-kube-api-access-48z87\") pod \"8654436c-452a-43f6-8932-55b0a85d60d2\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.630917 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ovn-combined-ca-bundle\") pod \"8654436c-452a-43f6-8932-55b0a85d60d2\" (UID: \"8654436c-452a-43f6-8932-55b0a85d60d2\") " Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.640875 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "8654436c-452a-43f6-8932-55b0a85d60d2" (UID: "8654436c-452a-43f6-8932-55b0a85d60d2"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.641544 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ceph" (OuterVolumeSpecName: "ceph") pod "8654436c-452a-43f6-8932-55b0a85d60d2" (UID: "8654436c-452a-43f6-8932-55b0a85d60d2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.645393 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8654436c-452a-43f6-8932-55b0a85d60d2-kube-api-access-48z87" (OuterVolumeSpecName: "kube-api-access-48z87") pod "8654436c-452a-43f6-8932-55b0a85d60d2" (UID: "8654436c-452a-43f6-8932-55b0a85d60d2"). InnerVolumeSpecName "kube-api-access-48z87". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.670562 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8654436c-452a-43f6-8932-55b0a85d60d2" (UID: "8654436c-452a-43f6-8932-55b0a85d60d2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.683782 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8654436c-452a-43f6-8932-55b0a85d60d2-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "8654436c-452a-43f6-8932-55b0a85d60d2" (UID: "8654436c-452a-43f6-8932-55b0a85d60d2"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.686269 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-inventory" (OuterVolumeSpecName: "inventory") pod "8654436c-452a-43f6-8932-55b0a85d60d2" (UID: "8654436c-452a-43f6-8932-55b0a85d60d2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.733837 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.733884 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48z87\" (UniqueName: \"kubernetes.io/projected/8654436c-452a-43f6-8932-55b0a85d60d2-kube-api-access-48z87\") on node \"crc\" DevicePath \"\"" Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.733898 4775 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.733913 4775 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/8654436c-452a-43f6-8932-55b0a85d60d2-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.733925 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:51:33 crc kubenswrapper[4775]: I1002 03:51:33.733937 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8654436c-452a-43f6-8932-55b0a85d60d2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.026381 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-dv978" event={"ID":"8654436c-452a-43f6-8932-55b0a85d60d2","Type":"ContainerDied","Data":"9ac69cfab85d8cf7c30b589eaa81ce00e76b60a9a727bf67d7af0b09fec861e0"} Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.026442 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ac69cfab85d8cf7c30b589eaa81ce00e76b60a9a727bf67d7af0b09fec861e0" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.026524 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-dv978" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.150067 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-2zw6m"] Oct 02 03:51:34 crc kubenswrapper[4775]: E1002 03:51:34.150572 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8654436c-452a-43f6-8932-55b0a85d60d2" containerName="ovn-openstack-openstack-cell1" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.150593 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8654436c-452a-43f6-8932-55b0a85d60d2" containerName="ovn-openstack-openstack-cell1" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.150876 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8654436c-452a-43f6-8932-55b0a85d60d2" containerName="ovn-openstack-openstack-cell1" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.151750 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.162945 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.163004 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.163372 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.170836 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.171294 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.179319 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.207025 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-2zw6m"] Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.249216 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.249450 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.249577 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.249744 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.249815 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl4bt\" (UniqueName: \"kubernetes.io/projected/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-kube-api-access-jl4bt\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.249899 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.250081 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.351909 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.352025 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.352143 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.352177 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.352261 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.352397 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.352430 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl4bt\" (UniqueName: \"kubernetes.io/projected/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-kube-api-access-jl4bt\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.357035 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.358402 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.359728 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.364493 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.365318 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.373552 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.375783 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl4bt\" (UniqueName: \"kubernetes.io/projected/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-kube-api-access-jl4bt\") pod \"neutron-metadata-openstack-openstack-cell1-2zw6m\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:34 crc kubenswrapper[4775]: I1002 03:51:34.496171 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:51:35 crc kubenswrapper[4775]: I1002 03:51:35.057386 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-2zw6m"] Oct 02 03:51:36 crc kubenswrapper[4775]: I1002 03:51:36.056013 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" event={"ID":"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5","Type":"ContainerStarted","Data":"003349250becc4b77eb1f43cec22167d8f7bb47fa08bcf7ed5db785756bead5d"} Oct 02 03:51:36 crc kubenswrapper[4775]: I1002 03:51:36.056688 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" event={"ID":"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5","Type":"ContainerStarted","Data":"611af54d0aa8c4594351133a4f24f6cb046b7dbec479b499e056bae59e236f9a"} Oct 02 03:51:36 crc kubenswrapper[4775]: I1002 03:51:36.086258 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" podStartSLOduration=1.643442936 podStartE2EDuration="2.086237657s" podCreationTimestamp="2025-10-02 03:51:34 +0000 UTC" firstStartedPulling="2025-10-02 03:51:35.057721825 +0000 UTC m=+7832.224465865" lastFinishedPulling="2025-10-02 03:51:35.500516496 +0000 UTC m=+7832.667260586" observedRunningTime="2025-10-02 03:51:36.07435923 +0000 UTC m=+7833.241103300" watchObservedRunningTime="2025-10-02 03:51:36.086237657 +0000 UTC m=+7833.252981707" Oct 02 03:51:36 crc kubenswrapper[4775]: I1002 03:51:36.765849 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:51:36 crc kubenswrapper[4775]: E1002 03:51:36.766390 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:51:50 crc kubenswrapper[4775]: I1002 03:51:50.766084 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:51:50 crc kubenswrapper[4775]: E1002 03:51:50.767278 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:52:02 crc kubenswrapper[4775]: I1002 03:52:02.990318 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rdffm"] Oct 02 03:52:02 crc kubenswrapper[4775]: I1002 03:52:02.995779 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:03 crc kubenswrapper[4775]: I1002 03:52:03.009161 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdffm"] Oct 02 03:52:03 crc kubenswrapper[4775]: I1002 03:52:03.058309 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-utilities\") pod \"redhat-marketplace-rdffm\" (UID: \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\") " pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:03 crc kubenswrapper[4775]: I1002 03:52:03.058438 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-catalog-content\") pod \"redhat-marketplace-rdffm\" (UID: \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\") " pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:03 crc kubenswrapper[4775]: I1002 03:52:03.058485 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw9sx\" (UniqueName: \"kubernetes.io/projected/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-kube-api-access-rw9sx\") pod \"redhat-marketplace-rdffm\" (UID: \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\") " pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:03 crc kubenswrapper[4775]: I1002 03:52:03.162046 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-utilities\") pod \"redhat-marketplace-rdffm\" (UID: \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\") " pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:03 crc kubenswrapper[4775]: I1002 03:52:03.162620 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-catalog-content\") pod \"redhat-marketplace-rdffm\" (UID: \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\") " pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:03 crc kubenswrapper[4775]: I1002 03:52:03.163090 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw9sx\" (UniqueName: \"kubernetes.io/projected/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-kube-api-access-rw9sx\") pod \"redhat-marketplace-rdffm\" (UID: \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\") " pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:03 crc kubenswrapper[4775]: I1002 03:52:03.163230 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-utilities\") pod \"redhat-marketplace-rdffm\" (UID: \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\") " pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:03 crc kubenswrapper[4775]: I1002 03:52:03.163243 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-catalog-content\") pod \"redhat-marketplace-rdffm\" (UID: \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\") " pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:03 crc kubenswrapper[4775]: I1002 03:52:03.208607 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw9sx\" (UniqueName: \"kubernetes.io/projected/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-kube-api-access-rw9sx\") pod \"redhat-marketplace-rdffm\" (UID: \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\") " pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:03 crc kubenswrapper[4775]: I1002 03:52:03.337658 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:03 crc kubenswrapper[4775]: I1002 03:52:03.779341 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:52:03 crc kubenswrapper[4775]: E1002 03:52:03.780660 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:52:03 crc kubenswrapper[4775]: I1002 03:52:03.851536 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdffm"] Oct 02 03:52:04 crc kubenswrapper[4775]: I1002 03:52:04.432216 4775 generic.go:334] "Generic (PLEG): container finished" podID="3af7002c-d551-42ba-a6f5-3c9c297ca5bd" containerID="7559d0e979f193fd6f5ae171d1c20f7ceb590d164dc615f1540698df85235d8c" exitCode=0 Oct 02 03:52:04 crc kubenswrapper[4775]: I1002 03:52:04.432559 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdffm" event={"ID":"3af7002c-d551-42ba-a6f5-3c9c297ca5bd","Type":"ContainerDied","Data":"7559d0e979f193fd6f5ae171d1c20f7ceb590d164dc615f1540698df85235d8c"} Oct 02 03:52:04 crc kubenswrapper[4775]: I1002 03:52:04.432591 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdffm" event={"ID":"3af7002c-d551-42ba-a6f5-3c9c297ca5bd","Type":"ContainerStarted","Data":"73890e93125e14e134f29d08351bd25768bf8eaeea01a7eb96cd1813a07e33d4"} Oct 02 03:52:05 crc kubenswrapper[4775]: I1002 03:52:05.447399 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdffm" event={"ID":"3af7002c-d551-42ba-a6f5-3c9c297ca5bd","Type":"ContainerStarted","Data":"0b4fc1da6b5c144c9ede276412c672f2b69710f9cf94db567e75c928af66c787"} Oct 02 03:52:06 crc kubenswrapper[4775]: I1002 03:52:06.466367 4775 generic.go:334] "Generic (PLEG): container finished" podID="3af7002c-d551-42ba-a6f5-3c9c297ca5bd" containerID="0b4fc1da6b5c144c9ede276412c672f2b69710f9cf94db567e75c928af66c787" exitCode=0 Oct 02 03:52:06 crc kubenswrapper[4775]: I1002 03:52:06.466535 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdffm" event={"ID":"3af7002c-d551-42ba-a6f5-3c9c297ca5bd","Type":"ContainerDied","Data":"0b4fc1da6b5c144c9ede276412c672f2b69710f9cf94db567e75c928af66c787"} Oct 02 03:52:07 crc kubenswrapper[4775]: I1002 03:52:07.487866 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdffm" event={"ID":"3af7002c-d551-42ba-a6f5-3c9c297ca5bd","Type":"ContainerStarted","Data":"0053f72cfdf4615335555d035c29b057c4a3c1372e5fa89b06a14de01dc5646c"} Oct 02 03:52:07 crc kubenswrapper[4775]: I1002 03:52:07.519587 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rdffm" podStartSLOduration=2.8334035010000003 podStartE2EDuration="5.519559764s" podCreationTimestamp="2025-10-02 03:52:02 +0000 UTC" firstStartedPulling="2025-10-02 03:52:04.434665158 +0000 UTC m=+7861.601409198" lastFinishedPulling="2025-10-02 03:52:07.120821381 +0000 UTC m=+7864.287565461" observedRunningTime="2025-10-02 03:52:07.511780054 +0000 UTC m=+7864.678524134" watchObservedRunningTime="2025-10-02 03:52:07.519559764 +0000 UTC m=+7864.686303844" Oct 02 03:52:13 crc kubenswrapper[4775]: I1002 03:52:13.338174 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:13 crc kubenswrapper[4775]: I1002 03:52:13.339109 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:13 crc kubenswrapper[4775]: I1002 03:52:13.431828 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:13 crc kubenswrapper[4775]: I1002 03:52:13.649238 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:14 crc kubenswrapper[4775]: I1002 03:52:14.020775 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdffm"] Oct 02 03:52:14 crc kubenswrapper[4775]: I1002 03:52:14.766368 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:52:14 crc kubenswrapper[4775]: E1002 03:52:14.766716 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:52:15 crc kubenswrapper[4775]: I1002 03:52:15.584330 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rdffm" podUID="3af7002c-d551-42ba-a6f5-3c9c297ca5bd" containerName="registry-server" containerID="cri-o://0053f72cfdf4615335555d035c29b057c4a3c1372e5fa89b06a14de01dc5646c" gracePeriod=2 Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.117580 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.201628 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw9sx\" (UniqueName: \"kubernetes.io/projected/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-kube-api-access-rw9sx\") pod \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\" (UID: \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\") " Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.202002 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-utilities\") pod \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\" (UID: \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\") " Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.202068 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-catalog-content\") pod \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\" (UID: \"3af7002c-d551-42ba-a6f5-3c9c297ca5bd\") " Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.202947 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-utilities" (OuterVolumeSpecName: "utilities") pod "3af7002c-d551-42ba-a6f5-3c9c297ca5bd" (UID: "3af7002c-d551-42ba-a6f5-3c9c297ca5bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.208261 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-kube-api-access-rw9sx" (OuterVolumeSpecName: "kube-api-access-rw9sx") pod "3af7002c-d551-42ba-a6f5-3c9c297ca5bd" (UID: "3af7002c-d551-42ba-a6f5-3c9c297ca5bd"). InnerVolumeSpecName "kube-api-access-rw9sx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.215277 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3af7002c-d551-42ba-a6f5-3c9c297ca5bd" (UID: "3af7002c-d551-42ba-a6f5-3c9c297ca5bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.304435 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.304475 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.304490 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw9sx\" (UniqueName: \"kubernetes.io/projected/3af7002c-d551-42ba-a6f5-3c9c297ca5bd-kube-api-access-rw9sx\") on node \"crc\" DevicePath \"\"" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.622400 4775 generic.go:334] "Generic (PLEG): container finished" podID="3af7002c-d551-42ba-a6f5-3c9c297ca5bd" containerID="0053f72cfdf4615335555d035c29b057c4a3c1372e5fa89b06a14de01dc5646c" exitCode=0 Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.622443 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdffm" event={"ID":"3af7002c-d551-42ba-a6f5-3c9c297ca5bd","Type":"ContainerDied","Data":"0053f72cfdf4615335555d035c29b057c4a3c1372e5fa89b06a14de01dc5646c"} Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.622469 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rdffm" event={"ID":"3af7002c-d551-42ba-a6f5-3c9c297ca5bd","Type":"ContainerDied","Data":"73890e93125e14e134f29d08351bd25768bf8eaeea01a7eb96cd1813a07e33d4"} Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.622486 4775 scope.go:117] "RemoveContainer" containerID="0053f72cfdf4615335555d035c29b057c4a3c1372e5fa89b06a14de01dc5646c" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.622619 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rdffm" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.670089 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdffm"] Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.680051 4775 scope.go:117] "RemoveContainer" containerID="0b4fc1da6b5c144c9ede276412c672f2b69710f9cf94db567e75c928af66c787" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.680894 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rdffm"] Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.706330 4775 scope.go:117] "RemoveContainer" containerID="7559d0e979f193fd6f5ae171d1c20f7ceb590d164dc615f1540698df85235d8c" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.756623 4775 scope.go:117] "RemoveContainer" containerID="0053f72cfdf4615335555d035c29b057c4a3c1372e5fa89b06a14de01dc5646c" Oct 02 03:52:16 crc kubenswrapper[4775]: E1002 03:52:16.757269 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0053f72cfdf4615335555d035c29b057c4a3c1372e5fa89b06a14de01dc5646c\": container with ID starting with 0053f72cfdf4615335555d035c29b057c4a3c1372e5fa89b06a14de01dc5646c not found: ID does not exist" containerID="0053f72cfdf4615335555d035c29b057c4a3c1372e5fa89b06a14de01dc5646c" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.757344 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0053f72cfdf4615335555d035c29b057c4a3c1372e5fa89b06a14de01dc5646c"} err="failed to get container status \"0053f72cfdf4615335555d035c29b057c4a3c1372e5fa89b06a14de01dc5646c\": rpc error: code = NotFound desc = could not find container \"0053f72cfdf4615335555d035c29b057c4a3c1372e5fa89b06a14de01dc5646c\": container with ID starting with 0053f72cfdf4615335555d035c29b057c4a3c1372e5fa89b06a14de01dc5646c not found: ID does not exist" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.757391 4775 scope.go:117] "RemoveContainer" containerID="0b4fc1da6b5c144c9ede276412c672f2b69710f9cf94db567e75c928af66c787" Oct 02 03:52:16 crc kubenswrapper[4775]: E1002 03:52:16.757868 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b4fc1da6b5c144c9ede276412c672f2b69710f9cf94db567e75c928af66c787\": container with ID starting with 0b4fc1da6b5c144c9ede276412c672f2b69710f9cf94db567e75c928af66c787 not found: ID does not exist" containerID="0b4fc1da6b5c144c9ede276412c672f2b69710f9cf94db567e75c928af66c787" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.757991 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b4fc1da6b5c144c9ede276412c672f2b69710f9cf94db567e75c928af66c787"} err="failed to get container status \"0b4fc1da6b5c144c9ede276412c672f2b69710f9cf94db567e75c928af66c787\": rpc error: code = NotFound desc = could not find container \"0b4fc1da6b5c144c9ede276412c672f2b69710f9cf94db567e75c928af66c787\": container with ID starting with 0b4fc1da6b5c144c9ede276412c672f2b69710f9cf94db567e75c928af66c787 not found: ID does not exist" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.758050 4775 scope.go:117] "RemoveContainer" containerID="7559d0e979f193fd6f5ae171d1c20f7ceb590d164dc615f1540698df85235d8c" Oct 02 03:52:16 crc kubenswrapper[4775]: E1002 03:52:16.758592 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7559d0e979f193fd6f5ae171d1c20f7ceb590d164dc615f1540698df85235d8c\": container with ID starting with 7559d0e979f193fd6f5ae171d1c20f7ceb590d164dc615f1540698df85235d8c not found: ID does not exist" containerID="7559d0e979f193fd6f5ae171d1c20f7ceb590d164dc615f1540698df85235d8c" Oct 02 03:52:16 crc kubenswrapper[4775]: I1002 03:52:16.758653 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7559d0e979f193fd6f5ae171d1c20f7ceb590d164dc615f1540698df85235d8c"} err="failed to get container status \"7559d0e979f193fd6f5ae171d1c20f7ceb590d164dc615f1540698df85235d8c\": rpc error: code = NotFound desc = could not find container \"7559d0e979f193fd6f5ae171d1c20f7ceb590d164dc615f1540698df85235d8c\": container with ID starting with 7559d0e979f193fd6f5ae171d1c20f7ceb590d164dc615f1540698df85235d8c not found: ID does not exist" Oct 02 03:52:17 crc kubenswrapper[4775]: I1002 03:52:17.798233 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3af7002c-d551-42ba-a6f5-3c9c297ca5bd" path="/var/lib/kubelet/pods/3af7002c-d551-42ba-a6f5-3c9c297ca5bd/volumes" Oct 02 03:52:26 crc kubenswrapper[4775]: I1002 03:52:26.766479 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:52:26 crc kubenswrapper[4775]: E1002 03:52:26.767549 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:52:30 crc kubenswrapper[4775]: I1002 03:52:30.813919 4775 generic.go:334] "Generic (PLEG): container finished" podID="93ccc1b8-b113-45a6-a69f-4c9b9e6052c5" containerID="003349250becc4b77eb1f43cec22167d8f7bb47fa08bcf7ed5db785756bead5d" exitCode=0 Oct 02 03:52:30 crc kubenswrapper[4775]: I1002 03:52:30.814027 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" event={"ID":"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5","Type":"ContainerDied","Data":"003349250becc4b77eb1f43cec22167d8f7bb47fa08bcf7ed5db785756bead5d"} Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.348374 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.406538 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-ceph\") pod \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.406604 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-nova-metadata-neutron-config-0\") pod \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.406718 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-neutron-metadata-combined-ca-bundle\") pod \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.406840 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-neutron-ovn-metadata-agent-neutron-config-0\") pod \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.407016 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-ssh-key\") pod \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.407045 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl4bt\" (UniqueName: \"kubernetes.io/projected/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-kube-api-access-jl4bt\") pod \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.407095 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-inventory\") pod \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\" (UID: \"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5\") " Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.421160 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-ceph" (OuterVolumeSpecName: "ceph") pod "93ccc1b8-b113-45a6-a69f-4c9b9e6052c5" (UID: "93ccc1b8-b113-45a6-a69f-4c9b9e6052c5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.425927 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-kube-api-access-jl4bt" (OuterVolumeSpecName: "kube-api-access-jl4bt") pod "93ccc1b8-b113-45a6-a69f-4c9b9e6052c5" (UID: "93ccc1b8-b113-45a6-a69f-4c9b9e6052c5"). InnerVolumeSpecName "kube-api-access-jl4bt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.427158 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "93ccc1b8-b113-45a6-a69f-4c9b9e6052c5" (UID: "93ccc1b8-b113-45a6-a69f-4c9b9e6052c5"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.449766 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "93ccc1b8-b113-45a6-a69f-4c9b9e6052c5" (UID: "93ccc1b8-b113-45a6-a69f-4c9b9e6052c5"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.456163 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "93ccc1b8-b113-45a6-a69f-4c9b9e6052c5" (UID: "93ccc1b8-b113-45a6-a69f-4c9b9e6052c5"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.463667 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-inventory" (OuterVolumeSpecName: "inventory") pod "93ccc1b8-b113-45a6-a69f-4c9b9e6052c5" (UID: "93ccc1b8-b113-45a6-a69f-4c9b9e6052c5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.465432 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "93ccc1b8-b113-45a6-a69f-4c9b9e6052c5" (UID: "93ccc1b8-b113-45a6-a69f-4c9b9e6052c5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.509885 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.509964 4775 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.509983 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.509996 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.510010 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.510022 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl4bt\" (UniqueName: \"kubernetes.io/projected/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-kube-api-access-jl4bt\") on node \"crc\" DevicePath \"\"" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.510036 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93ccc1b8-b113-45a6-a69f-4c9b9e6052c5-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.838016 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" event={"ID":"93ccc1b8-b113-45a6-a69f-4c9b9e6052c5","Type":"ContainerDied","Data":"611af54d0aa8c4594351133a4f24f6cb046b7dbec479b499e056bae59e236f9a"} Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.838057 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="611af54d0aa8c4594351133a4f24f6cb046b7dbec479b499e056bae59e236f9a" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.838110 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-2zw6m" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.949179 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-csddm"] Oct 02 03:52:32 crc kubenswrapper[4775]: E1002 03:52:32.950230 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93ccc1b8-b113-45a6-a69f-4c9b9e6052c5" containerName="neutron-metadata-openstack-openstack-cell1" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.950282 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="93ccc1b8-b113-45a6-a69f-4c9b9e6052c5" containerName="neutron-metadata-openstack-openstack-cell1" Oct 02 03:52:32 crc kubenswrapper[4775]: E1002 03:52:32.950319 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3af7002c-d551-42ba-a6f5-3c9c297ca5bd" containerName="extract-content" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.950336 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3af7002c-d551-42ba-a6f5-3c9c297ca5bd" containerName="extract-content" Oct 02 03:52:32 crc kubenswrapper[4775]: E1002 03:52:32.950433 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3af7002c-d551-42ba-a6f5-3c9c297ca5bd" containerName="extract-utilities" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.950458 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3af7002c-d551-42ba-a6f5-3c9c297ca5bd" containerName="extract-utilities" Oct 02 03:52:32 crc kubenswrapper[4775]: E1002 03:52:32.950512 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3af7002c-d551-42ba-a6f5-3c9c297ca5bd" containerName="registry-server" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.950531 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3af7002c-d551-42ba-a6f5-3c9c297ca5bd" containerName="registry-server" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.950920 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3af7002c-d551-42ba-a6f5-3c9c297ca5bd" containerName="registry-server" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.950998 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="93ccc1b8-b113-45a6-a69f-4c9b9e6052c5" containerName="neutron-metadata-openstack-openstack-cell1" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.952215 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.954297 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.954499 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.954503 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.956095 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.959244 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:52:32 crc kubenswrapper[4775]: I1002 03:52:32.964902 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-csddm"] Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.023135 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-ceph\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.023247 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-inventory\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.023353 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.023423 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.023450 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmcbq\" (UniqueName: \"kubernetes.io/projected/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-kube-api-access-rmcbq\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.023480 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-ssh-key\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.125016 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-ssh-key\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.125110 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-ceph\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.125181 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-inventory\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.125285 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.125356 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.125384 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmcbq\" (UniqueName: \"kubernetes.io/projected/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-kube-api-access-rmcbq\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.129218 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-ssh-key\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.129570 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-inventory\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.131703 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-ceph\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.133609 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.134160 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.145661 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmcbq\" (UniqueName: \"kubernetes.io/projected/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-kube-api-access-rmcbq\") pod \"libvirt-openstack-openstack-cell1-csddm\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.280099 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.875202 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-csddm"] Oct 02 03:52:33 crc kubenswrapper[4775]: I1002 03:52:33.883188 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 03:52:34 crc kubenswrapper[4775]: I1002 03:52:34.877109 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-csddm" event={"ID":"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc","Type":"ContainerStarted","Data":"7ea3688a0fa4179deed8340efc705f1165206a33b6f24ee1a1da01f2f3a876ed"} Oct 02 03:52:34 crc kubenswrapper[4775]: I1002 03:52:34.877722 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-csddm" event={"ID":"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc","Type":"ContainerStarted","Data":"cd7ce78f22747e6f98e8b2ca9574ed9cc1aa9ec8c53c7de1f370f6adfd55073b"} Oct 02 03:52:34 crc kubenswrapper[4775]: I1002 03:52:34.907761 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-csddm" podStartSLOduration=2.496906032 podStartE2EDuration="2.907734177s" podCreationTimestamp="2025-10-02 03:52:32 +0000 UTC" firstStartedPulling="2025-10-02 03:52:33.882905281 +0000 UTC m=+7891.049649321" lastFinishedPulling="2025-10-02 03:52:34.293733426 +0000 UTC m=+7891.460477466" observedRunningTime="2025-10-02 03:52:34.898466527 +0000 UTC m=+7892.065210667" watchObservedRunningTime="2025-10-02 03:52:34.907734177 +0000 UTC m=+7892.074478257" Oct 02 03:52:39 crc kubenswrapper[4775]: I1002 03:52:39.769207 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:52:39 crc kubenswrapper[4775]: E1002 03:52:39.769782 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:52:54 crc kubenswrapper[4775]: I1002 03:52:54.765321 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:52:54 crc kubenswrapper[4775]: E1002 03:52:54.765914 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:53:05 crc kubenswrapper[4775]: I1002 03:53:05.766269 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:53:05 crc kubenswrapper[4775]: E1002 03:53:05.767346 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:53:19 crc kubenswrapper[4775]: I1002 03:53:19.766190 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:53:19 crc kubenswrapper[4775]: E1002 03:53:19.767408 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:53:32 crc kubenswrapper[4775]: I1002 03:53:32.765579 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:53:32 crc kubenswrapper[4775]: E1002 03:53:32.766658 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:53:43 crc kubenswrapper[4775]: I1002 03:53:43.539985 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-twqfm"] Oct 02 03:53:43 crc kubenswrapper[4775]: I1002 03:53:43.545773 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:43 crc kubenswrapper[4775]: I1002 03:53:43.563196 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-twqfm"] Oct 02 03:53:43 crc kubenswrapper[4775]: I1002 03:53:43.711499 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4352459a-1e74-4756-b26b-f1f9298e1875-catalog-content\") pod \"community-operators-twqfm\" (UID: \"4352459a-1e74-4756-b26b-f1f9298e1875\") " pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:43 crc kubenswrapper[4775]: I1002 03:53:43.711577 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kndn9\" (UniqueName: \"kubernetes.io/projected/4352459a-1e74-4756-b26b-f1f9298e1875-kube-api-access-kndn9\") pod \"community-operators-twqfm\" (UID: \"4352459a-1e74-4756-b26b-f1f9298e1875\") " pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:43 crc kubenswrapper[4775]: I1002 03:53:43.711881 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4352459a-1e74-4756-b26b-f1f9298e1875-utilities\") pod \"community-operators-twqfm\" (UID: \"4352459a-1e74-4756-b26b-f1f9298e1875\") " pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:43 crc kubenswrapper[4775]: I1002 03:53:43.779654 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:53:43 crc kubenswrapper[4775]: E1002 03:53:43.780256 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:53:43 crc kubenswrapper[4775]: I1002 03:53:43.814328 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4352459a-1e74-4756-b26b-f1f9298e1875-catalog-content\") pod \"community-operators-twqfm\" (UID: \"4352459a-1e74-4756-b26b-f1f9298e1875\") " pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:43 crc kubenswrapper[4775]: I1002 03:53:43.814397 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kndn9\" (UniqueName: \"kubernetes.io/projected/4352459a-1e74-4756-b26b-f1f9298e1875-kube-api-access-kndn9\") pod \"community-operators-twqfm\" (UID: \"4352459a-1e74-4756-b26b-f1f9298e1875\") " pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:43 crc kubenswrapper[4775]: I1002 03:53:43.814617 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4352459a-1e74-4756-b26b-f1f9298e1875-utilities\") pod \"community-operators-twqfm\" (UID: \"4352459a-1e74-4756-b26b-f1f9298e1875\") " pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:43 crc kubenswrapper[4775]: I1002 03:53:43.815209 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4352459a-1e74-4756-b26b-f1f9298e1875-catalog-content\") pod \"community-operators-twqfm\" (UID: \"4352459a-1e74-4756-b26b-f1f9298e1875\") " pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:43 crc kubenswrapper[4775]: I1002 03:53:43.815296 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4352459a-1e74-4756-b26b-f1f9298e1875-utilities\") pod \"community-operators-twqfm\" (UID: \"4352459a-1e74-4756-b26b-f1f9298e1875\") " pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:43 crc kubenswrapper[4775]: I1002 03:53:43.846246 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kndn9\" (UniqueName: \"kubernetes.io/projected/4352459a-1e74-4756-b26b-f1f9298e1875-kube-api-access-kndn9\") pod \"community-operators-twqfm\" (UID: \"4352459a-1e74-4756-b26b-f1f9298e1875\") " pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:43 crc kubenswrapper[4775]: I1002 03:53:43.889979 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:44 crc kubenswrapper[4775]: I1002 03:53:44.452244 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-twqfm"] Oct 02 03:53:44 crc kubenswrapper[4775]: I1002 03:53:44.858271 4775 generic.go:334] "Generic (PLEG): container finished" podID="4352459a-1e74-4756-b26b-f1f9298e1875" containerID="a9cc55861fb2d1c3a9fab534555ed3e5453f7b58afdd60632412cd1644a5b4e6" exitCode=0 Oct 02 03:53:44 crc kubenswrapper[4775]: I1002 03:53:44.858560 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twqfm" event={"ID":"4352459a-1e74-4756-b26b-f1f9298e1875","Type":"ContainerDied","Data":"a9cc55861fb2d1c3a9fab534555ed3e5453f7b58afdd60632412cd1644a5b4e6"} Oct 02 03:53:44 crc kubenswrapper[4775]: I1002 03:53:44.858589 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twqfm" event={"ID":"4352459a-1e74-4756-b26b-f1f9298e1875","Type":"ContainerStarted","Data":"a76b8734cbadb1edfb324dfd6d4ebd7ec99b16c49bef377d59321565a9930a61"} Oct 02 03:53:45 crc kubenswrapper[4775]: I1002 03:53:45.876264 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twqfm" event={"ID":"4352459a-1e74-4756-b26b-f1f9298e1875","Type":"ContainerStarted","Data":"72a63c3beba8dc1703b040baf1047681a2b25fcc66aeec2ca5aec4d44d0a0be7"} Oct 02 03:53:47 crc kubenswrapper[4775]: I1002 03:53:47.910786 4775 generic.go:334] "Generic (PLEG): container finished" podID="4352459a-1e74-4756-b26b-f1f9298e1875" containerID="72a63c3beba8dc1703b040baf1047681a2b25fcc66aeec2ca5aec4d44d0a0be7" exitCode=0 Oct 02 03:53:47 crc kubenswrapper[4775]: I1002 03:53:47.910910 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twqfm" event={"ID":"4352459a-1e74-4756-b26b-f1f9298e1875","Type":"ContainerDied","Data":"72a63c3beba8dc1703b040baf1047681a2b25fcc66aeec2ca5aec4d44d0a0be7"} Oct 02 03:53:48 crc kubenswrapper[4775]: I1002 03:53:48.925129 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twqfm" event={"ID":"4352459a-1e74-4756-b26b-f1f9298e1875","Type":"ContainerStarted","Data":"e742dd7897a88056ca9570d6ec838133928d21230a443f560d397196fd4aa3d6"} Oct 02 03:53:48 crc kubenswrapper[4775]: I1002 03:53:48.948848 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-twqfm" podStartSLOduration=2.487006386 podStartE2EDuration="5.948832144s" podCreationTimestamp="2025-10-02 03:53:43 +0000 UTC" firstStartedPulling="2025-10-02 03:53:44.861260252 +0000 UTC m=+7962.028004302" lastFinishedPulling="2025-10-02 03:53:48.32308599 +0000 UTC m=+7965.489830060" observedRunningTime="2025-10-02 03:53:48.943661521 +0000 UTC m=+7966.110405561" watchObservedRunningTime="2025-10-02 03:53:48.948832144 +0000 UTC m=+7966.115576184" Oct 02 03:53:53 crc kubenswrapper[4775]: I1002 03:53:53.890359 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:53 crc kubenswrapper[4775]: I1002 03:53:53.891098 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:53 crc kubenswrapper[4775]: I1002 03:53:53.994144 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:54 crc kubenswrapper[4775]: I1002 03:53:54.097066 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:54 crc kubenswrapper[4775]: I1002 03:53:54.691530 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-twqfm"] Oct 02 03:53:56 crc kubenswrapper[4775]: I1002 03:53:56.037456 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-twqfm" podUID="4352459a-1e74-4756-b26b-f1f9298e1875" containerName="registry-server" containerID="cri-o://e742dd7897a88056ca9570d6ec838133928d21230a443f560d397196fd4aa3d6" gracePeriod=2 Oct 02 03:53:56 crc kubenswrapper[4775]: I1002 03:53:56.676609 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:56 crc kubenswrapper[4775]: I1002 03:53:56.766023 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:53:56 crc kubenswrapper[4775]: E1002 03:53:56.766410 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:53:56 crc kubenswrapper[4775]: I1002 03:53:56.785540 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4352459a-1e74-4756-b26b-f1f9298e1875-utilities\") pod \"4352459a-1e74-4756-b26b-f1f9298e1875\" (UID: \"4352459a-1e74-4756-b26b-f1f9298e1875\") " Oct 02 03:53:56 crc kubenswrapper[4775]: I1002 03:53:56.785688 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kndn9\" (UniqueName: \"kubernetes.io/projected/4352459a-1e74-4756-b26b-f1f9298e1875-kube-api-access-kndn9\") pod \"4352459a-1e74-4756-b26b-f1f9298e1875\" (UID: \"4352459a-1e74-4756-b26b-f1f9298e1875\") " Oct 02 03:53:56 crc kubenswrapper[4775]: I1002 03:53:56.785971 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4352459a-1e74-4756-b26b-f1f9298e1875-catalog-content\") pod \"4352459a-1e74-4756-b26b-f1f9298e1875\" (UID: \"4352459a-1e74-4756-b26b-f1f9298e1875\") " Oct 02 03:53:56 crc kubenswrapper[4775]: I1002 03:53:56.786388 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4352459a-1e74-4756-b26b-f1f9298e1875-utilities" (OuterVolumeSpecName: "utilities") pod "4352459a-1e74-4756-b26b-f1f9298e1875" (UID: "4352459a-1e74-4756-b26b-f1f9298e1875"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:53:56 crc kubenswrapper[4775]: I1002 03:53:56.786641 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4352459a-1e74-4756-b26b-f1f9298e1875-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:53:56 crc kubenswrapper[4775]: I1002 03:53:56.791351 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4352459a-1e74-4756-b26b-f1f9298e1875-kube-api-access-kndn9" (OuterVolumeSpecName: "kube-api-access-kndn9") pod "4352459a-1e74-4756-b26b-f1f9298e1875" (UID: "4352459a-1e74-4756-b26b-f1f9298e1875"). InnerVolumeSpecName "kube-api-access-kndn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:53:56 crc kubenswrapper[4775]: I1002 03:53:56.843833 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4352459a-1e74-4756-b26b-f1f9298e1875-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4352459a-1e74-4756-b26b-f1f9298e1875" (UID: "4352459a-1e74-4756-b26b-f1f9298e1875"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:53:56 crc kubenswrapper[4775]: I1002 03:53:56.889186 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4352459a-1e74-4756-b26b-f1f9298e1875-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:53:56 crc kubenswrapper[4775]: I1002 03:53:56.889225 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kndn9\" (UniqueName: \"kubernetes.io/projected/4352459a-1e74-4756-b26b-f1f9298e1875-kube-api-access-kndn9\") on node \"crc\" DevicePath \"\"" Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.055215 4775 generic.go:334] "Generic (PLEG): container finished" podID="4352459a-1e74-4756-b26b-f1f9298e1875" containerID="e742dd7897a88056ca9570d6ec838133928d21230a443f560d397196fd4aa3d6" exitCode=0 Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.055288 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twqfm" event={"ID":"4352459a-1e74-4756-b26b-f1f9298e1875","Type":"ContainerDied","Data":"e742dd7897a88056ca9570d6ec838133928d21230a443f560d397196fd4aa3d6"} Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.055292 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-twqfm" Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.055341 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-twqfm" event={"ID":"4352459a-1e74-4756-b26b-f1f9298e1875","Type":"ContainerDied","Data":"a76b8734cbadb1edfb324dfd6d4ebd7ec99b16c49bef377d59321565a9930a61"} Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.055378 4775 scope.go:117] "RemoveContainer" containerID="e742dd7897a88056ca9570d6ec838133928d21230a443f560d397196fd4aa3d6" Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.085367 4775 scope.go:117] "RemoveContainer" containerID="72a63c3beba8dc1703b040baf1047681a2b25fcc66aeec2ca5aec4d44d0a0be7" Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.113810 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-twqfm"] Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.132377 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-twqfm"] Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.138150 4775 scope.go:117] "RemoveContainer" containerID="a9cc55861fb2d1c3a9fab534555ed3e5453f7b58afdd60632412cd1644a5b4e6" Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.187458 4775 scope.go:117] "RemoveContainer" containerID="e742dd7897a88056ca9570d6ec838133928d21230a443f560d397196fd4aa3d6" Oct 02 03:53:57 crc kubenswrapper[4775]: E1002 03:53:57.188047 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e742dd7897a88056ca9570d6ec838133928d21230a443f560d397196fd4aa3d6\": container with ID starting with e742dd7897a88056ca9570d6ec838133928d21230a443f560d397196fd4aa3d6 not found: ID does not exist" containerID="e742dd7897a88056ca9570d6ec838133928d21230a443f560d397196fd4aa3d6" Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.188122 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e742dd7897a88056ca9570d6ec838133928d21230a443f560d397196fd4aa3d6"} err="failed to get container status \"e742dd7897a88056ca9570d6ec838133928d21230a443f560d397196fd4aa3d6\": rpc error: code = NotFound desc = could not find container \"e742dd7897a88056ca9570d6ec838133928d21230a443f560d397196fd4aa3d6\": container with ID starting with e742dd7897a88056ca9570d6ec838133928d21230a443f560d397196fd4aa3d6 not found: ID does not exist" Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.188167 4775 scope.go:117] "RemoveContainer" containerID="72a63c3beba8dc1703b040baf1047681a2b25fcc66aeec2ca5aec4d44d0a0be7" Oct 02 03:53:57 crc kubenswrapper[4775]: E1002 03:53:57.188669 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72a63c3beba8dc1703b040baf1047681a2b25fcc66aeec2ca5aec4d44d0a0be7\": container with ID starting with 72a63c3beba8dc1703b040baf1047681a2b25fcc66aeec2ca5aec4d44d0a0be7 not found: ID does not exist" containerID="72a63c3beba8dc1703b040baf1047681a2b25fcc66aeec2ca5aec4d44d0a0be7" Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.188712 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72a63c3beba8dc1703b040baf1047681a2b25fcc66aeec2ca5aec4d44d0a0be7"} err="failed to get container status \"72a63c3beba8dc1703b040baf1047681a2b25fcc66aeec2ca5aec4d44d0a0be7\": rpc error: code = NotFound desc = could not find container \"72a63c3beba8dc1703b040baf1047681a2b25fcc66aeec2ca5aec4d44d0a0be7\": container with ID starting with 72a63c3beba8dc1703b040baf1047681a2b25fcc66aeec2ca5aec4d44d0a0be7 not found: ID does not exist" Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.188739 4775 scope.go:117] "RemoveContainer" containerID="a9cc55861fb2d1c3a9fab534555ed3e5453f7b58afdd60632412cd1644a5b4e6" Oct 02 03:53:57 crc kubenswrapper[4775]: E1002 03:53:57.189150 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9cc55861fb2d1c3a9fab534555ed3e5453f7b58afdd60632412cd1644a5b4e6\": container with ID starting with a9cc55861fb2d1c3a9fab534555ed3e5453f7b58afdd60632412cd1644a5b4e6 not found: ID does not exist" containerID="a9cc55861fb2d1c3a9fab534555ed3e5453f7b58afdd60632412cd1644a5b4e6" Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.189188 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9cc55861fb2d1c3a9fab534555ed3e5453f7b58afdd60632412cd1644a5b4e6"} err="failed to get container status \"a9cc55861fb2d1c3a9fab534555ed3e5453f7b58afdd60632412cd1644a5b4e6\": rpc error: code = NotFound desc = could not find container \"a9cc55861fb2d1c3a9fab534555ed3e5453f7b58afdd60632412cd1644a5b4e6\": container with ID starting with a9cc55861fb2d1c3a9fab534555ed3e5453f7b58afdd60632412cd1644a5b4e6 not found: ID does not exist" Oct 02 03:53:57 crc kubenswrapper[4775]: I1002 03:53:57.791797 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4352459a-1e74-4756-b26b-f1f9298e1875" path="/var/lib/kubelet/pods/4352459a-1e74-4756-b26b-f1f9298e1875/volumes" Oct 02 03:54:09 crc kubenswrapper[4775]: I1002 03:54:09.766024 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:54:09 crc kubenswrapper[4775]: E1002 03:54:09.766933 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:54:21 crc kubenswrapper[4775]: I1002 03:54:21.766445 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:54:21 crc kubenswrapper[4775]: E1002 03:54:21.767600 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:54:36 crc kubenswrapper[4775]: I1002 03:54:36.766084 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:54:36 crc kubenswrapper[4775]: E1002 03:54:36.766879 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:54:49 crc kubenswrapper[4775]: I1002 03:54:49.769131 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:54:49 crc kubenswrapper[4775]: E1002 03:54:49.770000 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:55:01 crc kubenswrapper[4775]: I1002 03:55:01.766063 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:55:01 crc kubenswrapper[4775]: E1002 03:55:01.767119 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 03:55:16 crc kubenswrapper[4775]: I1002 03:55:16.765445 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 03:55:17 crc kubenswrapper[4775]: I1002 03:55:17.102592 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"bcdfff962e86b5bff70329682f81b7b093a8fcf67a837119fd8bdcdc3b8b7084"} Oct 02 03:55:51 crc kubenswrapper[4775]: I1002 03:55:51.935646 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tgc9m"] Oct 02 03:55:51 crc kubenswrapper[4775]: E1002 03:55:51.936487 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4352459a-1e74-4756-b26b-f1f9298e1875" containerName="extract-utilities" Oct 02 03:55:51 crc kubenswrapper[4775]: I1002 03:55:51.936499 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4352459a-1e74-4756-b26b-f1f9298e1875" containerName="extract-utilities" Oct 02 03:55:51 crc kubenswrapper[4775]: E1002 03:55:51.936527 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4352459a-1e74-4756-b26b-f1f9298e1875" containerName="extract-content" Oct 02 03:55:51 crc kubenswrapper[4775]: I1002 03:55:51.936533 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4352459a-1e74-4756-b26b-f1f9298e1875" containerName="extract-content" Oct 02 03:55:51 crc kubenswrapper[4775]: E1002 03:55:51.936564 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4352459a-1e74-4756-b26b-f1f9298e1875" containerName="registry-server" Oct 02 03:55:51 crc kubenswrapper[4775]: I1002 03:55:51.936570 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4352459a-1e74-4756-b26b-f1f9298e1875" containerName="registry-server" Oct 02 03:55:51 crc kubenswrapper[4775]: I1002 03:55:51.936757 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4352459a-1e74-4756-b26b-f1f9298e1875" containerName="registry-server" Oct 02 03:55:51 crc kubenswrapper[4775]: I1002 03:55:51.938218 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:55:51 crc kubenswrapper[4775]: I1002 03:55:51.946294 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tgc9m"] Oct 02 03:55:52 crc kubenswrapper[4775]: I1002 03:55:52.001006 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc71e48-7568-4825-a798-5a5ba4638058-utilities\") pod \"redhat-operators-tgc9m\" (UID: \"9bc71e48-7568-4825-a798-5a5ba4638058\") " pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:55:52 crc kubenswrapper[4775]: I1002 03:55:52.001165 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26mhw\" (UniqueName: \"kubernetes.io/projected/9bc71e48-7568-4825-a798-5a5ba4638058-kube-api-access-26mhw\") pod \"redhat-operators-tgc9m\" (UID: \"9bc71e48-7568-4825-a798-5a5ba4638058\") " pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:55:52 crc kubenswrapper[4775]: I1002 03:55:52.001206 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc71e48-7568-4825-a798-5a5ba4638058-catalog-content\") pod \"redhat-operators-tgc9m\" (UID: \"9bc71e48-7568-4825-a798-5a5ba4638058\") " pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:55:52 crc kubenswrapper[4775]: I1002 03:55:52.103608 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26mhw\" (UniqueName: \"kubernetes.io/projected/9bc71e48-7568-4825-a798-5a5ba4638058-kube-api-access-26mhw\") pod \"redhat-operators-tgc9m\" (UID: \"9bc71e48-7568-4825-a798-5a5ba4638058\") " pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:55:52 crc kubenswrapper[4775]: I1002 03:55:52.104093 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc71e48-7568-4825-a798-5a5ba4638058-catalog-content\") pod \"redhat-operators-tgc9m\" (UID: \"9bc71e48-7568-4825-a798-5a5ba4638058\") " pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:55:52 crc kubenswrapper[4775]: I1002 03:55:52.104408 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc71e48-7568-4825-a798-5a5ba4638058-utilities\") pod \"redhat-operators-tgc9m\" (UID: \"9bc71e48-7568-4825-a798-5a5ba4638058\") " pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:55:52 crc kubenswrapper[4775]: I1002 03:55:52.104623 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc71e48-7568-4825-a798-5a5ba4638058-catalog-content\") pod \"redhat-operators-tgc9m\" (UID: \"9bc71e48-7568-4825-a798-5a5ba4638058\") " pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:55:52 crc kubenswrapper[4775]: I1002 03:55:52.104765 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc71e48-7568-4825-a798-5a5ba4638058-utilities\") pod \"redhat-operators-tgc9m\" (UID: \"9bc71e48-7568-4825-a798-5a5ba4638058\") " pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:55:52 crc kubenswrapper[4775]: I1002 03:55:52.122613 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26mhw\" (UniqueName: \"kubernetes.io/projected/9bc71e48-7568-4825-a798-5a5ba4638058-kube-api-access-26mhw\") pod \"redhat-operators-tgc9m\" (UID: \"9bc71e48-7568-4825-a798-5a5ba4638058\") " pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:55:52 crc kubenswrapper[4775]: I1002 03:55:52.269477 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:55:52 crc kubenswrapper[4775]: I1002 03:55:52.769064 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tgc9m"] Oct 02 03:55:53 crc kubenswrapper[4775]: I1002 03:55:53.634612 4775 generic.go:334] "Generic (PLEG): container finished" podID="9bc71e48-7568-4825-a798-5a5ba4638058" containerID="644271f0b2ec2038781537588e415d14f852ef869b2a7bb1f39132456807e35e" exitCode=0 Oct 02 03:55:53 crc kubenswrapper[4775]: I1002 03:55:53.634698 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgc9m" event={"ID":"9bc71e48-7568-4825-a798-5a5ba4638058","Type":"ContainerDied","Data":"644271f0b2ec2038781537588e415d14f852ef869b2a7bb1f39132456807e35e"} Oct 02 03:55:53 crc kubenswrapper[4775]: I1002 03:55:53.635073 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgc9m" event={"ID":"9bc71e48-7568-4825-a798-5a5ba4638058","Type":"ContainerStarted","Data":"b013f99e1a5a0e998d1a59ffae357ddd128325f87b11dee0d6028bf8c5359f6a"} Oct 02 03:55:55 crc kubenswrapper[4775]: I1002 03:55:55.666821 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgc9m" event={"ID":"9bc71e48-7568-4825-a798-5a5ba4638058","Type":"ContainerStarted","Data":"c9aaab2e55cf08db169e5414d2a8dffd55a784a1b205c4b909e0ffc76e5f06a3"} Oct 02 03:55:56 crc kubenswrapper[4775]: I1002 03:55:56.684257 4775 generic.go:334] "Generic (PLEG): container finished" podID="9bc71e48-7568-4825-a798-5a5ba4638058" containerID="c9aaab2e55cf08db169e5414d2a8dffd55a784a1b205c4b909e0ffc76e5f06a3" exitCode=0 Oct 02 03:55:56 crc kubenswrapper[4775]: I1002 03:55:56.684395 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgc9m" event={"ID":"9bc71e48-7568-4825-a798-5a5ba4638058","Type":"ContainerDied","Data":"c9aaab2e55cf08db169e5414d2a8dffd55a784a1b205c4b909e0ffc76e5f06a3"} Oct 02 03:55:57 crc kubenswrapper[4775]: I1002 03:55:57.702332 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgc9m" event={"ID":"9bc71e48-7568-4825-a798-5a5ba4638058","Type":"ContainerStarted","Data":"2494f33d085627f07fdbcb52d0da5e3173d8bfb4b4dcab23881ee1386aebd5e2"} Oct 02 03:55:57 crc kubenswrapper[4775]: I1002 03:55:57.733658 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tgc9m" podStartSLOduration=3.294338484 podStartE2EDuration="6.73363067s" podCreationTimestamp="2025-10-02 03:55:51 +0000 UTC" firstStartedPulling="2025-10-02 03:55:53.637476247 +0000 UTC m=+8090.804220287" lastFinishedPulling="2025-10-02 03:55:57.076768403 +0000 UTC m=+8094.243512473" observedRunningTime="2025-10-02 03:55:57.72625322 +0000 UTC m=+8094.892997260" watchObservedRunningTime="2025-10-02 03:55:57.73363067 +0000 UTC m=+8094.900374740" Oct 02 03:56:02 crc kubenswrapper[4775]: I1002 03:56:02.270450 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:56:02 crc kubenswrapper[4775]: I1002 03:56:02.270864 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:56:03 crc kubenswrapper[4775]: I1002 03:56:03.332570 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tgc9m" podUID="9bc71e48-7568-4825-a798-5a5ba4638058" containerName="registry-server" probeResult="failure" output=< Oct 02 03:56:03 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 03:56:03 crc kubenswrapper[4775]: > Oct 02 03:56:12 crc kubenswrapper[4775]: I1002 03:56:12.346443 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:56:12 crc kubenswrapper[4775]: I1002 03:56:12.403384 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:56:12 crc kubenswrapper[4775]: I1002 03:56:12.609612 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tgc9m"] Oct 02 03:56:13 crc kubenswrapper[4775]: I1002 03:56:13.914813 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tgc9m" podUID="9bc71e48-7568-4825-a798-5a5ba4638058" containerName="registry-server" containerID="cri-o://2494f33d085627f07fdbcb52d0da5e3173d8bfb4b4dcab23881ee1386aebd5e2" gracePeriod=2 Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.457620 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.588876 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26mhw\" (UniqueName: \"kubernetes.io/projected/9bc71e48-7568-4825-a798-5a5ba4638058-kube-api-access-26mhw\") pod \"9bc71e48-7568-4825-a798-5a5ba4638058\" (UID: \"9bc71e48-7568-4825-a798-5a5ba4638058\") " Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.588969 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc71e48-7568-4825-a798-5a5ba4638058-utilities\") pod \"9bc71e48-7568-4825-a798-5a5ba4638058\" (UID: \"9bc71e48-7568-4825-a798-5a5ba4638058\") " Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.589074 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc71e48-7568-4825-a798-5a5ba4638058-catalog-content\") pod \"9bc71e48-7568-4825-a798-5a5ba4638058\" (UID: \"9bc71e48-7568-4825-a798-5a5ba4638058\") " Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.591243 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bc71e48-7568-4825-a798-5a5ba4638058-utilities" (OuterVolumeSpecName: "utilities") pod "9bc71e48-7568-4825-a798-5a5ba4638058" (UID: "9bc71e48-7568-4825-a798-5a5ba4638058"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.597940 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bc71e48-7568-4825-a798-5a5ba4638058-kube-api-access-26mhw" (OuterVolumeSpecName: "kube-api-access-26mhw") pod "9bc71e48-7568-4825-a798-5a5ba4638058" (UID: "9bc71e48-7568-4825-a798-5a5ba4638058"). InnerVolumeSpecName "kube-api-access-26mhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.679056 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9bc71e48-7568-4825-a798-5a5ba4638058-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9bc71e48-7568-4825-a798-5a5ba4638058" (UID: "9bc71e48-7568-4825-a798-5a5ba4638058"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.692315 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26mhw\" (UniqueName: \"kubernetes.io/projected/9bc71e48-7568-4825-a798-5a5ba4638058-kube-api-access-26mhw\") on node \"crc\" DevicePath \"\"" Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.692357 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9bc71e48-7568-4825-a798-5a5ba4638058-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.692379 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9bc71e48-7568-4825-a798-5a5ba4638058-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.937140 4775 generic.go:334] "Generic (PLEG): container finished" podID="9bc71e48-7568-4825-a798-5a5ba4638058" containerID="2494f33d085627f07fdbcb52d0da5e3173d8bfb4b4dcab23881ee1386aebd5e2" exitCode=0 Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.937383 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgc9m" event={"ID":"9bc71e48-7568-4825-a798-5a5ba4638058","Type":"ContainerDied","Data":"2494f33d085627f07fdbcb52d0da5e3173d8bfb4b4dcab23881ee1386aebd5e2"} Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.938187 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgc9m" event={"ID":"9bc71e48-7568-4825-a798-5a5ba4638058","Type":"ContainerDied","Data":"b013f99e1a5a0e998d1a59ffae357ddd128325f87b11dee0d6028bf8c5359f6a"} Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.938231 4775 scope.go:117] "RemoveContainer" containerID="2494f33d085627f07fdbcb52d0da5e3173d8bfb4b4dcab23881ee1386aebd5e2" Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.937505 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgc9m" Oct 02 03:56:14 crc kubenswrapper[4775]: I1002 03:56:14.997328 4775 scope.go:117] "RemoveContainer" containerID="c9aaab2e55cf08db169e5414d2a8dffd55a784a1b205c4b909e0ffc76e5f06a3" Oct 02 03:56:15 crc kubenswrapper[4775]: I1002 03:56:15.023499 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tgc9m"] Oct 02 03:56:15 crc kubenswrapper[4775]: I1002 03:56:15.033442 4775 scope.go:117] "RemoveContainer" containerID="644271f0b2ec2038781537588e415d14f852ef869b2a7bb1f39132456807e35e" Oct 02 03:56:15 crc kubenswrapper[4775]: I1002 03:56:15.035678 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tgc9m"] Oct 02 03:56:15 crc kubenswrapper[4775]: I1002 03:56:15.085015 4775 scope.go:117] "RemoveContainer" containerID="2494f33d085627f07fdbcb52d0da5e3173d8bfb4b4dcab23881ee1386aebd5e2" Oct 02 03:56:15 crc kubenswrapper[4775]: E1002 03:56:15.085549 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2494f33d085627f07fdbcb52d0da5e3173d8bfb4b4dcab23881ee1386aebd5e2\": container with ID starting with 2494f33d085627f07fdbcb52d0da5e3173d8bfb4b4dcab23881ee1386aebd5e2 not found: ID does not exist" containerID="2494f33d085627f07fdbcb52d0da5e3173d8bfb4b4dcab23881ee1386aebd5e2" Oct 02 03:56:15 crc kubenswrapper[4775]: I1002 03:56:15.085589 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2494f33d085627f07fdbcb52d0da5e3173d8bfb4b4dcab23881ee1386aebd5e2"} err="failed to get container status \"2494f33d085627f07fdbcb52d0da5e3173d8bfb4b4dcab23881ee1386aebd5e2\": rpc error: code = NotFound desc = could not find container \"2494f33d085627f07fdbcb52d0da5e3173d8bfb4b4dcab23881ee1386aebd5e2\": container with ID starting with 2494f33d085627f07fdbcb52d0da5e3173d8bfb4b4dcab23881ee1386aebd5e2 not found: ID does not exist" Oct 02 03:56:15 crc kubenswrapper[4775]: I1002 03:56:15.085616 4775 scope.go:117] "RemoveContainer" containerID="c9aaab2e55cf08db169e5414d2a8dffd55a784a1b205c4b909e0ffc76e5f06a3" Oct 02 03:56:15 crc kubenswrapper[4775]: E1002 03:56:15.086136 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9aaab2e55cf08db169e5414d2a8dffd55a784a1b205c4b909e0ffc76e5f06a3\": container with ID starting with c9aaab2e55cf08db169e5414d2a8dffd55a784a1b205c4b909e0ffc76e5f06a3 not found: ID does not exist" containerID="c9aaab2e55cf08db169e5414d2a8dffd55a784a1b205c4b909e0ffc76e5f06a3" Oct 02 03:56:15 crc kubenswrapper[4775]: I1002 03:56:15.086165 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9aaab2e55cf08db169e5414d2a8dffd55a784a1b205c4b909e0ffc76e5f06a3"} err="failed to get container status \"c9aaab2e55cf08db169e5414d2a8dffd55a784a1b205c4b909e0ffc76e5f06a3\": rpc error: code = NotFound desc = could not find container \"c9aaab2e55cf08db169e5414d2a8dffd55a784a1b205c4b909e0ffc76e5f06a3\": container with ID starting with c9aaab2e55cf08db169e5414d2a8dffd55a784a1b205c4b909e0ffc76e5f06a3 not found: ID does not exist" Oct 02 03:56:15 crc kubenswrapper[4775]: I1002 03:56:15.086182 4775 scope.go:117] "RemoveContainer" containerID="644271f0b2ec2038781537588e415d14f852ef869b2a7bb1f39132456807e35e" Oct 02 03:56:15 crc kubenswrapper[4775]: E1002 03:56:15.086517 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"644271f0b2ec2038781537588e415d14f852ef869b2a7bb1f39132456807e35e\": container with ID starting with 644271f0b2ec2038781537588e415d14f852ef869b2a7bb1f39132456807e35e not found: ID does not exist" containerID="644271f0b2ec2038781537588e415d14f852ef869b2a7bb1f39132456807e35e" Oct 02 03:56:15 crc kubenswrapper[4775]: I1002 03:56:15.086548 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"644271f0b2ec2038781537588e415d14f852ef869b2a7bb1f39132456807e35e"} err="failed to get container status \"644271f0b2ec2038781537588e415d14f852ef869b2a7bb1f39132456807e35e\": rpc error: code = NotFound desc = could not find container \"644271f0b2ec2038781537588e415d14f852ef869b2a7bb1f39132456807e35e\": container with ID starting with 644271f0b2ec2038781537588e415d14f852ef869b2a7bb1f39132456807e35e not found: ID does not exist" Oct 02 03:56:15 crc kubenswrapper[4775]: I1002 03:56:15.801444 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bc71e48-7568-4825-a798-5a5ba4638058" path="/var/lib/kubelet/pods/9bc71e48-7568-4825-a798-5a5ba4638058/volumes" Oct 02 03:56:17 crc kubenswrapper[4775]: I1002 03:56:17.690041 4775 scope.go:117] "RemoveContainer" containerID="e4d649dd86199b52dbc7c094a0ba1620cb5e2bf140c13cbc6803dbd7b90345d7" Oct 02 03:56:17 crc kubenswrapper[4775]: I1002 03:56:17.721088 4775 scope.go:117] "RemoveContainer" containerID="3596dd33bdffe5c781ce72c2634bce08ffa03b3d1279bd40ba9540b2fabd78a4" Oct 02 03:56:17 crc kubenswrapper[4775]: I1002 03:56:17.784947 4775 scope.go:117] "RemoveContainer" containerID="478a8c74516f1ba5862e1a1dbf552902edd4546bd7dfe709bd123019d11c6eaa" Oct 02 03:57:17 crc kubenswrapper[4775]: I1002 03:57:17.824233 4775 generic.go:334] "Generic (PLEG): container finished" podID="9c88413d-ac75-4d79-9d76-54e4b1c8c1fc" containerID="7ea3688a0fa4179deed8340efc705f1165206a33b6f24ee1a1da01f2f3a876ed" exitCode=0 Oct 02 03:57:17 crc kubenswrapper[4775]: I1002 03:57:17.824356 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-csddm" event={"ID":"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc","Type":"ContainerDied","Data":"7ea3688a0fa4179deed8340efc705f1165206a33b6f24ee1a1da01f2f3a876ed"} Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.452864 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.567502 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmcbq\" (UniqueName: \"kubernetes.io/projected/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-kube-api-access-rmcbq\") pod \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.567575 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-ssh-key\") pod \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.567623 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-libvirt-combined-ca-bundle\") pod \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.567835 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-libvirt-secret-0\") pod \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.567860 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-inventory\") pod \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.567918 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-ceph\") pod \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\" (UID: \"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc\") " Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.576117 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-ceph" (OuterVolumeSpecName: "ceph") pod "9c88413d-ac75-4d79-9d76-54e4b1c8c1fc" (UID: "9c88413d-ac75-4d79-9d76-54e4b1c8c1fc"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.577890 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-kube-api-access-rmcbq" (OuterVolumeSpecName: "kube-api-access-rmcbq") pod "9c88413d-ac75-4d79-9d76-54e4b1c8c1fc" (UID: "9c88413d-ac75-4d79-9d76-54e4b1c8c1fc"). InnerVolumeSpecName "kube-api-access-rmcbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.582093 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "9c88413d-ac75-4d79-9d76-54e4b1c8c1fc" (UID: "9c88413d-ac75-4d79-9d76-54e4b1c8c1fc"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.605907 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "9c88413d-ac75-4d79-9d76-54e4b1c8c1fc" (UID: "9c88413d-ac75-4d79-9d76-54e4b1c8c1fc"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.613571 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9c88413d-ac75-4d79-9d76-54e4b1c8c1fc" (UID: "9c88413d-ac75-4d79-9d76-54e4b1c8c1fc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.632098 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-inventory" (OuterVolumeSpecName: "inventory") pod "9c88413d-ac75-4d79-9d76-54e4b1c8c1fc" (UID: "9c88413d-ac75-4d79-9d76-54e4b1c8c1fc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.670299 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.670332 4775 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.670345 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.670359 4775 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.670370 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.670382 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmcbq\" (UniqueName: \"kubernetes.io/projected/9c88413d-ac75-4d79-9d76-54e4b1c8c1fc-kube-api-access-rmcbq\") on node \"crc\" DevicePath \"\"" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.865021 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-csddm" event={"ID":"9c88413d-ac75-4d79-9d76-54e4b1c8c1fc","Type":"ContainerDied","Data":"cd7ce78f22747e6f98e8b2ca9574ed9cc1aa9ec8c53c7de1f370f6adfd55073b"} Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.865375 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd7ce78f22747e6f98e8b2ca9574ed9cc1aa9ec8c53c7de1f370f6adfd55073b" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.865086 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-csddm" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.958652 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-dsmb2"] Oct 02 03:57:19 crc kubenswrapper[4775]: E1002 03:57:19.959276 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc71e48-7568-4825-a798-5a5ba4638058" containerName="registry-server" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.959299 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc71e48-7568-4825-a798-5a5ba4638058" containerName="registry-server" Oct 02 03:57:19 crc kubenswrapper[4775]: E1002 03:57:19.959322 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc71e48-7568-4825-a798-5a5ba4638058" containerName="extract-content" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.959332 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc71e48-7568-4825-a798-5a5ba4638058" containerName="extract-content" Oct 02 03:57:19 crc kubenswrapper[4775]: E1002 03:57:19.959346 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc71e48-7568-4825-a798-5a5ba4638058" containerName="extract-utilities" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.959357 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc71e48-7568-4825-a798-5a5ba4638058" containerName="extract-utilities" Oct 02 03:57:19 crc kubenswrapper[4775]: E1002 03:57:19.959391 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c88413d-ac75-4d79-9d76-54e4b1c8c1fc" containerName="libvirt-openstack-openstack-cell1" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.959399 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c88413d-ac75-4d79-9d76-54e4b1c8c1fc" containerName="libvirt-openstack-openstack-cell1" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.959641 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c88413d-ac75-4d79-9d76-54e4b1c8c1fc" containerName="libvirt-openstack-openstack-cell1" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.959670 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bc71e48-7568-4825-a798-5a5ba4638058" containerName="registry-server" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.960654 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.967287 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.967485 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.967931 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.967938 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.968124 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.968155 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.970677 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-dsmb2"] Oct 02 03:57:19 crc kubenswrapper[4775]: I1002 03:57:19.991460 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.079479 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.079538 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.079561 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-ceph\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.079581 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.079621 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.079640 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.079658 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.079713 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.079732 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9kqz\" (UniqueName: \"kubernetes.io/projected/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-kube-api-access-k9kqz\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.079787 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.079832 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-inventory\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.181432 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-inventory\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.181517 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.181542 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.181562 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-ceph\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.181580 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.181615 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.181633 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.181654 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.181705 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.181723 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9kqz\" (UniqueName: \"kubernetes.io/projected/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-kube-api-access-k9kqz\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.181774 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.183378 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.184950 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.186992 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.187120 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.187419 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.187576 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.188029 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-inventory\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.196605 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.199289 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-ceph\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.202029 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.202276 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9kqz\" (UniqueName: \"kubernetes.io/projected/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-kube-api-access-k9kqz\") pod \"nova-cell1-openstack-openstack-cell1-dsmb2\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.290615 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 03:57:20 crc kubenswrapper[4775]: I1002 03:57:20.924034 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-dsmb2"] Oct 02 03:57:20 crc kubenswrapper[4775]: W1002 03:57:20.930280 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00fb1524_9133_48e4_9fb7_63c91b3cf6fd.slice/crio-7a6745b68f0baed7231011ae24b3311062a3c4cd16195f0713867ace67dfc449 WatchSource:0}: Error finding container 7a6745b68f0baed7231011ae24b3311062a3c4cd16195f0713867ace67dfc449: Status 404 returned error can't find the container with id 7a6745b68f0baed7231011ae24b3311062a3c4cd16195f0713867ace67dfc449 Oct 02 03:57:21 crc kubenswrapper[4775]: I1002 03:57:21.886871 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" event={"ID":"00fb1524-9133-48e4-9fb7-63c91b3cf6fd","Type":"ContainerStarted","Data":"b4d942b29b5a10e85b2f3c5b780e1c467747f6b648199b4c1fd042cbadd7a96b"} Oct 02 03:57:21 crc kubenswrapper[4775]: I1002 03:57:21.887577 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" event={"ID":"00fb1524-9133-48e4-9fb7-63c91b3cf6fd","Type":"ContainerStarted","Data":"7a6745b68f0baed7231011ae24b3311062a3c4cd16195f0713867ace67dfc449"} Oct 02 03:57:21 crc kubenswrapper[4775]: I1002 03:57:21.912746 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" podStartSLOduration=2.376391557 podStartE2EDuration="2.912723443s" podCreationTimestamp="2025-10-02 03:57:19 +0000 UTC" firstStartedPulling="2025-10-02 03:57:20.933195596 +0000 UTC m=+8178.099939646" lastFinishedPulling="2025-10-02 03:57:21.469527452 +0000 UTC m=+8178.636271532" observedRunningTime="2025-10-02 03:57:21.908454463 +0000 UTC m=+8179.075198553" watchObservedRunningTime="2025-10-02 03:57:21.912723443 +0000 UTC m=+8179.079467523" Oct 02 03:57:37 crc kubenswrapper[4775]: I1002 03:57:37.234033 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:57:37 crc kubenswrapper[4775]: I1002 03:57:37.234813 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:58:07 crc kubenswrapper[4775]: I1002 03:58:07.233676 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:58:07 crc kubenswrapper[4775]: I1002 03:58:07.234420 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:58:37 crc kubenswrapper[4775]: I1002 03:58:37.233318 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 03:58:37 crc kubenswrapper[4775]: I1002 03:58:37.234140 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 03:58:37 crc kubenswrapper[4775]: I1002 03:58:37.234204 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 03:58:37 crc kubenswrapper[4775]: I1002 03:58:37.235420 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bcdfff962e86b5bff70329682f81b7b093a8fcf67a837119fd8bdcdc3b8b7084"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 03:58:37 crc kubenswrapper[4775]: I1002 03:58:37.235545 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://bcdfff962e86b5bff70329682f81b7b093a8fcf67a837119fd8bdcdc3b8b7084" gracePeriod=600 Oct 02 03:58:37 crc kubenswrapper[4775]: I1002 03:58:37.906744 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="bcdfff962e86b5bff70329682f81b7b093a8fcf67a837119fd8bdcdc3b8b7084" exitCode=0 Oct 02 03:58:37 crc kubenswrapper[4775]: I1002 03:58:37.907007 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"bcdfff962e86b5bff70329682f81b7b093a8fcf67a837119fd8bdcdc3b8b7084"} Oct 02 03:58:37 crc kubenswrapper[4775]: I1002 03:58:37.907154 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676"} Oct 02 03:58:37 crc kubenswrapper[4775]: I1002 03:58:37.907192 4775 scope.go:117] "RemoveContainer" containerID="188bcf3d4a08a467e614dfdc740760b9303b48ee56cbda8434046cad3788aed4" Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.161945 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv"] Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.165592 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.174816 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv"] Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.210979 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.211659 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.265912 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr8kb\" (UniqueName: \"kubernetes.io/projected/f272330d-5e4d-47d1-9a5d-d59256aab172-kube-api-access-cr8kb\") pod \"collect-profiles-29322960-c78nv\" (UID: \"f272330d-5e4d-47d1-9a5d-d59256aab172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.266133 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f272330d-5e4d-47d1-9a5d-d59256aab172-config-volume\") pod \"collect-profiles-29322960-c78nv\" (UID: \"f272330d-5e4d-47d1-9a5d-d59256aab172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.266272 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f272330d-5e4d-47d1-9a5d-d59256aab172-secret-volume\") pod \"collect-profiles-29322960-c78nv\" (UID: \"f272330d-5e4d-47d1-9a5d-d59256aab172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.368441 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f272330d-5e4d-47d1-9a5d-d59256aab172-secret-volume\") pod \"collect-profiles-29322960-c78nv\" (UID: \"f272330d-5e4d-47d1-9a5d-d59256aab172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.368781 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr8kb\" (UniqueName: \"kubernetes.io/projected/f272330d-5e4d-47d1-9a5d-d59256aab172-kube-api-access-cr8kb\") pod \"collect-profiles-29322960-c78nv\" (UID: \"f272330d-5e4d-47d1-9a5d-d59256aab172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.368854 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f272330d-5e4d-47d1-9a5d-d59256aab172-config-volume\") pod \"collect-profiles-29322960-c78nv\" (UID: \"f272330d-5e4d-47d1-9a5d-d59256aab172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.370244 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f272330d-5e4d-47d1-9a5d-d59256aab172-config-volume\") pod \"collect-profiles-29322960-c78nv\" (UID: \"f272330d-5e4d-47d1-9a5d-d59256aab172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.375752 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f272330d-5e4d-47d1-9a5d-d59256aab172-secret-volume\") pod \"collect-profiles-29322960-c78nv\" (UID: \"f272330d-5e4d-47d1-9a5d-d59256aab172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.390337 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr8kb\" (UniqueName: \"kubernetes.io/projected/f272330d-5e4d-47d1-9a5d-d59256aab172-kube-api-access-cr8kb\") pod \"collect-profiles-29322960-c78nv\" (UID: \"f272330d-5e4d-47d1-9a5d-d59256aab172\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" Oct 02 04:00:00 crc kubenswrapper[4775]: I1002 04:00:00.544098 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" Oct 02 04:00:01 crc kubenswrapper[4775]: I1002 04:00:01.060924 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv"] Oct 02 04:00:01 crc kubenswrapper[4775]: I1002 04:00:01.921883 4775 generic.go:334] "Generic (PLEG): container finished" podID="f272330d-5e4d-47d1-9a5d-d59256aab172" containerID="5497be0863fd4dd9c7f626f33241ceff7fce4544428bea117446fa718818ad2b" exitCode=0 Oct 02 04:00:01 crc kubenswrapper[4775]: I1002 04:00:01.921947 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" event={"ID":"f272330d-5e4d-47d1-9a5d-d59256aab172","Type":"ContainerDied","Data":"5497be0863fd4dd9c7f626f33241ceff7fce4544428bea117446fa718818ad2b"} Oct 02 04:00:01 crc kubenswrapper[4775]: I1002 04:00:01.922154 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" event={"ID":"f272330d-5e4d-47d1-9a5d-d59256aab172","Type":"ContainerStarted","Data":"d674ca4334d3785789b764417382821f6e33915587d31409a148baf0e2e0ee08"} Oct 02 04:00:03 crc kubenswrapper[4775]: I1002 04:00:03.460243 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" Oct 02 04:00:03 crc kubenswrapper[4775]: I1002 04:00:03.570672 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f272330d-5e4d-47d1-9a5d-d59256aab172-config-volume\") pod \"f272330d-5e4d-47d1-9a5d-d59256aab172\" (UID: \"f272330d-5e4d-47d1-9a5d-d59256aab172\") " Oct 02 04:00:03 crc kubenswrapper[4775]: I1002 04:00:03.570894 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f272330d-5e4d-47d1-9a5d-d59256aab172-secret-volume\") pod \"f272330d-5e4d-47d1-9a5d-d59256aab172\" (UID: \"f272330d-5e4d-47d1-9a5d-d59256aab172\") " Oct 02 04:00:03 crc kubenswrapper[4775]: I1002 04:00:03.570940 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cr8kb\" (UniqueName: \"kubernetes.io/projected/f272330d-5e4d-47d1-9a5d-d59256aab172-kube-api-access-cr8kb\") pod \"f272330d-5e4d-47d1-9a5d-d59256aab172\" (UID: \"f272330d-5e4d-47d1-9a5d-d59256aab172\") " Oct 02 04:00:03 crc kubenswrapper[4775]: I1002 04:00:03.572149 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f272330d-5e4d-47d1-9a5d-d59256aab172-config-volume" (OuterVolumeSpecName: "config-volume") pod "f272330d-5e4d-47d1-9a5d-d59256aab172" (UID: "f272330d-5e4d-47d1-9a5d-d59256aab172"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 04:00:03 crc kubenswrapper[4775]: I1002 04:00:03.582305 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f272330d-5e4d-47d1-9a5d-d59256aab172-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f272330d-5e4d-47d1-9a5d-d59256aab172" (UID: "f272330d-5e4d-47d1-9a5d-d59256aab172"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:00:03 crc kubenswrapper[4775]: I1002 04:00:03.582392 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f272330d-5e4d-47d1-9a5d-d59256aab172-kube-api-access-cr8kb" (OuterVolumeSpecName: "kube-api-access-cr8kb") pod "f272330d-5e4d-47d1-9a5d-d59256aab172" (UID: "f272330d-5e4d-47d1-9a5d-d59256aab172"). InnerVolumeSpecName "kube-api-access-cr8kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:00:03 crc kubenswrapper[4775]: I1002 04:00:03.673935 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f272330d-5e4d-47d1-9a5d-d59256aab172-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 04:00:03 crc kubenswrapper[4775]: I1002 04:00:03.673997 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f272330d-5e4d-47d1-9a5d-d59256aab172-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 04:00:03 crc kubenswrapper[4775]: I1002 04:00:03.674011 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cr8kb\" (UniqueName: \"kubernetes.io/projected/f272330d-5e4d-47d1-9a5d-d59256aab172-kube-api-access-cr8kb\") on node \"crc\" DevicePath \"\"" Oct 02 04:00:03 crc kubenswrapper[4775]: I1002 04:00:03.947738 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" event={"ID":"f272330d-5e4d-47d1-9a5d-d59256aab172","Type":"ContainerDied","Data":"d674ca4334d3785789b764417382821f6e33915587d31409a148baf0e2e0ee08"} Oct 02 04:00:03 crc kubenswrapper[4775]: I1002 04:00:03.947776 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d674ca4334d3785789b764417382821f6e33915587d31409a148baf0e2e0ee08" Oct 02 04:00:03 crc kubenswrapper[4775]: I1002 04:00:03.947776 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322960-c78nv" Oct 02 04:00:04 crc kubenswrapper[4775]: E1002 04:00:04.007837 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf272330d_5e4d_47d1_9a5d_d59256aab172.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf272330d_5e4d_47d1_9a5d_d59256aab172.slice/crio-d674ca4334d3785789b764417382821f6e33915587d31409a148baf0e2e0ee08\": RecentStats: unable to find data in memory cache]" Oct 02 04:00:04 crc kubenswrapper[4775]: I1002 04:00:04.554615 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc"] Oct 02 04:00:04 crc kubenswrapper[4775]: I1002 04:00:04.566984 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322915-sp6bc"] Oct 02 04:00:05 crc kubenswrapper[4775]: I1002 04:00:05.781985 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf9df4f5-dbb7-41f3-aff8-1b86fed890a0" path="/var/lib/kubelet/pods/bf9df4f5-dbb7-41f3-aff8-1b86fed890a0/volumes" Oct 02 04:00:17 crc kubenswrapper[4775]: I1002 04:00:17.996638 4775 scope.go:117] "RemoveContainer" containerID="82b6bf51ffc146bca85b24301f4c020584752fd5b210ba7c3f32194dc21f434a" Oct 02 04:00:37 crc kubenswrapper[4775]: I1002 04:00:37.233348 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:00:37 crc kubenswrapper[4775]: I1002 04:00:37.233922 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.183145 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29322961-tcwf8"] Oct 02 04:01:00 crc kubenswrapper[4775]: E1002 04:01:00.184883 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f272330d-5e4d-47d1-9a5d-d59256aab172" containerName="collect-profiles" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.184914 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="f272330d-5e4d-47d1-9a5d-d59256aab172" containerName="collect-profiles" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.185694 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="f272330d-5e4d-47d1-9a5d-d59256aab172" containerName="collect-profiles" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.187574 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.203662 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29322961-tcwf8"] Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.293026 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-combined-ca-bundle\") pod \"keystone-cron-29322961-tcwf8\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.293418 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-fernet-keys\") pod \"keystone-cron-29322961-tcwf8\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.293586 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sgrc\" (UniqueName: \"kubernetes.io/projected/05e4796e-9130-4382-abe0-939e36fd4f1b-kube-api-access-9sgrc\") pod \"keystone-cron-29322961-tcwf8\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.293842 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-config-data\") pod \"keystone-cron-29322961-tcwf8\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.396353 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-combined-ca-bundle\") pod \"keystone-cron-29322961-tcwf8\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.396437 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-fernet-keys\") pod \"keystone-cron-29322961-tcwf8\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.396475 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sgrc\" (UniqueName: \"kubernetes.io/projected/05e4796e-9130-4382-abe0-939e36fd4f1b-kube-api-access-9sgrc\") pod \"keystone-cron-29322961-tcwf8\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.396604 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-config-data\") pod \"keystone-cron-29322961-tcwf8\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.404273 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-config-data\") pod \"keystone-cron-29322961-tcwf8\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.405113 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-combined-ca-bundle\") pod \"keystone-cron-29322961-tcwf8\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.408919 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-fernet-keys\") pod \"keystone-cron-29322961-tcwf8\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.416845 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sgrc\" (UniqueName: \"kubernetes.io/projected/05e4796e-9130-4382-abe0-939e36fd4f1b-kube-api-access-9sgrc\") pod \"keystone-cron-29322961-tcwf8\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:00 crc kubenswrapper[4775]: I1002 04:01:00.511239 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:01 crc kubenswrapper[4775]: I1002 04:01:01.003913 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29322961-tcwf8"] Oct 02 04:01:01 crc kubenswrapper[4775]: I1002 04:01:01.674619 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322961-tcwf8" event={"ID":"05e4796e-9130-4382-abe0-939e36fd4f1b","Type":"ContainerStarted","Data":"fca714d494d6e7568f8d490202fd1f81ab9656847545685ce0d21fe2c84845db"} Oct 02 04:01:01 crc kubenswrapper[4775]: I1002 04:01:01.676473 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322961-tcwf8" event={"ID":"05e4796e-9130-4382-abe0-939e36fd4f1b","Type":"ContainerStarted","Data":"4c97b677514d17ae6d8fa73541e7f3f5d5fb2ab74fed195760961441213001c7"} Oct 02 04:01:01 crc kubenswrapper[4775]: I1002 04:01:01.694492 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29322961-tcwf8" podStartSLOduration=1.694473419 podStartE2EDuration="1.694473419s" podCreationTimestamp="2025-10-02 04:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 04:01:01.691167063 +0000 UTC m=+8398.857911113" watchObservedRunningTime="2025-10-02 04:01:01.694473419 +0000 UTC m=+8398.861217469" Oct 02 04:01:04 crc kubenswrapper[4775]: I1002 04:01:04.709628 4775 generic.go:334] "Generic (PLEG): container finished" podID="05e4796e-9130-4382-abe0-939e36fd4f1b" containerID="fca714d494d6e7568f8d490202fd1f81ab9656847545685ce0d21fe2c84845db" exitCode=0 Oct 02 04:01:04 crc kubenswrapper[4775]: I1002 04:01:04.709718 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322961-tcwf8" event={"ID":"05e4796e-9130-4382-abe0-939e36fd4f1b","Type":"ContainerDied","Data":"fca714d494d6e7568f8d490202fd1f81ab9656847545685ce0d21fe2c84845db"} Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.183188 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.340933 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sgrc\" (UniqueName: \"kubernetes.io/projected/05e4796e-9130-4382-abe0-939e36fd4f1b-kube-api-access-9sgrc\") pod \"05e4796e-9130-4382-abe0-939e36fd4f1b\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.341201 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-config-data\") pod \"05e4796e-9130-4382-abe0-939e36fd4f1b\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.341525 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-fernet-keys\") pod \"05e4796e-9130-4382-abe0-939e36fd4f1b\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.341589 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-combined-ca-bundle\") pod \"05e4796e-9130-4382-abe0-939e36fd4f1b\" (UID: \"05e4796e-9130-4382-abe0-939e36fd4f1b\") " Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.350063 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "05e4796e-9130-4382-abe0-939e36fd4f1b" (UID: "05e4796e-9130-4382-abe0-939e36fd4f1b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.354330 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05e4796e-9130-4382-abe0-939e36fd4f1b-kube-api-access-9sgrc" (OuterVolumeSpecName: "kube-api-access-9sgrc") pod "05e4796e-9130-4382-abe0-939e36fd4f1b" (UID: "05e4796e-9130-4382-abe0-939e36fd4f1b"). InnerVolumeSpecName "kube-api-access-9sgrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.400227 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05e4796e-9130-4382-abe0-939e36fd4f1b" (UID: "05e4796e-9130-4382-abe0-939e36fd4f1b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.407513 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-config-data" (OuterVolumeSpecName: "config-data") pod "05e4796e-9130-4382-abe0-939e36fd4f1b" (UID: "05e4796e-9130-4382-abe0-939e36fd4f1b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.445476 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sgrc\" (UniqueName: \"kubernetes.io/projected/05e4796e-9130-4382-abe0-939e36fd4f1b-kube-api-access-9sgrc\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.445532 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.445551 4775 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.445570 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05e4796e-9130-4382-abe0-939e36fd4f1b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.759288 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29322961-tcwf8" event={"ID":"05e4796e-9130-4382-abe0-939e36fd4f1b","Type":"ContainerDied","Data":"4c97b677514d17ae6d8fa73541e7f3f5d5fb2ab74fed195760961441213001c7"} Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.759861 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c97b677514d17ae6d8fa73541e7f3f5d5fb2ab74fed195760961441213001c7" Oct 02 04:01:06 crc kubenswrapper[4775]: I1002 04:01:06.759397 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29322961-tcwf8" Oct 02 04:01:07 crc kubenswrapper[4775]: I1002 04:01:07.233547 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:01:07 crc kubenswrapper[4775]: I1002 04:01:07.234515 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:01:10 crc kubenswrapper[4775]: I1002 04:01:10.853613 4775 generic.go:334] "Generic (PLEG): container finished" podID="00fb1524-9133-48e4-9fb7-63c91b3cf6fd" containerID="b4d942b29b5a10e85b2f3c5b780e1c467747f6b648199b4c1fd042cbadd7a96b" exitCode=0 Oct 02 04:01:10 crc kubenswrapper[4775]: I1002 04:01:10.853852 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" event={"ID":"00fb1524-9133-48e4-9fb7-63c91b3cf6fd","Type":"ContainerDied","Data":"b4d942b29b5a10e85b2f3c5b780e1c467747f6b648199b4c1fd042cbadd7a96b"} Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.460489 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.519301 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-ceph\") pod \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.519401 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9kqz\" (UniqueName: \"kubernetes.io/projected/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-kube-api-access-k9kqz\") pod \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.519463 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-compute-config-1\") pod \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.519494 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-compute-config-0\") pod \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.519539 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-migration-ssh-key-0\") pod \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.519558 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-ssh-key\") pod \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.519597 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cells-global-config-1\") pod \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.519656 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-combined-ca-bundle\") pod \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.519693 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-inventory\") pod \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.519718 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cells-global-config-0\") pod \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.519746 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-migration-ssh-key-1\") pod \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\" (UID: \"00fb1524-9133-48e4-9fb7-63c91b3cf6fd\") " Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.525882 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-kube-api-access-k9kqz" (OuterVolumeSpecName: "kube-api-access-k9kqz") pod "00fb1524-9133-48e4-9fb7-63c91b3cf6fd" (UID: "00fb1524-9133-48e4-9fb7-63c91b3cf6fd"). InnerVolumeSpecName "kube-api-access-k9kqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.535639 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "00fb1524-9133-48e4-9fb7-63c91b3cf6fd" (UID: "00fb1524-9133-48e4-9fb7-63c91b3cf6fd"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.538707 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-ceph" (OuterVolumeSpecName: "ceph") pod "00fb1524-9133-48e4-9fb7-63c91b3cf6fd" (UID: "00fb1524-9133-48e4-9fb7-63c91b3cf6fd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.550700 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "00fb1524-9133-48e4-9fb7-63c91b3cf6fd" (UID: "00fb1524-9133-48e4-9fb7-63c91b3cf6fd"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.555058 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "00fb1524-9133-48e4-9fb7-63c91b3cf6fd" (UID: "00fb1524-9133-48e4-9fb7-63c91b3cf6fd"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.558274 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "00fb1524-9133-48e4-9fb7-63c91b3cf6fd" (UID: "00fb1524-9133-48e4-9fb7-63c91b3cf6fd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.560412 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "00fb1524-9133-48e4-9fb7-63c91b3cf6fd" (UID: "00fb1524-9133-48e4-9fb7-63c91b3cf6fd"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.564544 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "00fb1524-9133-48e4-9fb7-63c91b3cf6fd" (UID: "00fb1524-9133-48e4-9fb7-63c91b3cf6fd"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.570763 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "00fb1524-9133-48e4-9fb7-63c91b3cf6fd" (UID: "00fb1524-9133-48e4-9fb7-63c91b3cf6fd"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.582667 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "00fb1524-9133-48e4-9fb7-63c91b3cf6fd" (UID: "00fb1524-9133-48e4-9fb7-63c91b3cf6fd"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.592052 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-inventory" (OuterVolumeSpecName: "inventory") pod "00fb1524-9133-48e4-9fb7-63c91b3cf6fd" (UID: "00fb1524-9133-48e4-9fb7-63c91b3cf6fd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.622927 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.622971 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.622986 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.622999 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.623012 4775 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.623023 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.623033 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9kqz\" (UniqueName: \"kubernetes.io/projected/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-kube-api-access-k9kqz\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.623044 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.623055 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.623068 4775 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.623080 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00fb1524-9133-48e4-9fb7-63c91b3cf6fd-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.886682 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" event={"ID":"00fb1524-9133-48e4-9fb7-63c91b3cf6fd","Type":"ContainerDied","Data":"7a6745b68f0baed7231011ae24b3311062a3c4cd16195f0713867ace67dfc449"} Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.886750 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a6745b68f0baed7231011ae24b3311062a3c4cd16195f0713867ace67dfc449" Oct 02 04:01:12 crc kubenswrapper[4775]: I1002 04:01:12.886893 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-dsmb2" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.120210 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-6vllg"] Oct 02 04:01:13 crc kubenswrapper[4775]: E1002 04:01:13.121278 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00fb1524-9133-48e4-9fb7-63c91b3cf6fd" containerName="nova-cell1-openstack-openstack-cell1" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.121309 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="00fb1524-9133-48e4-9fb7-63c91b3cf6fd" containerName="nova-cell1-openstack-openstack-cell1" Oct 02 04:01:13 crc kubenswrapper[4775]: E1002 04:01:13.121393 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05e4796e-9130-4382-abe0-939e36fd4f1b" containerName="keystone-cron" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.121409 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="05e4796e-9130-4382-abe0-939e36fd4f1b" containerName="keystone-cron" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.122123 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="00fb1524-9133-48e4-9fb7-63c91b3cf6fd" containerName="nova-cell1-openstack-openstack-cell1" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.122170 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="05e4796e-9130-4382-abe0-939e36fd4f1b" containerName="keystone-cron" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.124154 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.129823 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-6vllg"] Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.130043 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.130133 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.130267 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.130320 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.130469 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.235576 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-inventory\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.235724 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2btfq\" (UniqueName: \"kubernetes.io/projected/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-kube-api-access-2btfq\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.235870 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.235896 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceph\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.235929 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ssh-key\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.235988 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.236008 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.236171 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.338259 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceph\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.338369 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ssh-key\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.338480 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.338518 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.338651 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.338727 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-inventory\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.338790 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2btfq\" (UniqueName: \"kubernetes.io/projected/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-kube-api-access-2btfq\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.339421 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.343942 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.344246 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-inventory\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.345565 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.345923 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.347453 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.352518 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ssh-key\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.357088 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceph\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.358856 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2btfq\" (UniqueName: \"kubernetes.io/projected/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-kube-api-access-2btfq\") pod \"telemetry-openstack-openstack-cell1-6vllg\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:13 crc kubenswrapper[4775]: I1002 04:01:13.456619 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:01:14 crc kubenswrapper[4775]: I1002 04:01:14.109591 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-6vllg"] Oct 02 04:01:14 crc kubenswrapper[4775]: I1002 04:01:14.118707 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 04:01:14 crc kubenswrapper[4775]: I1002 04:01:14.910887 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-6vllg" event={"ID":"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d","Type":"ContainerStarted","Data":"dcccb13c70ac80a55f31d47bad24e4fb60652d7e9176e58aec454ca461170712"} Oct 02 04:01:15 crc kubenswrapper[4775]: I1002 04:01:15.924676 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-6vllg" event={"ID":"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d","Type":"ContainerStarted","Data":"2998421915a71cc4f8da7636111ca6b08427d0b676ce3f25acbea621669c5c94"} Oct 02 04:01:15 crc kubenswrapper[4775]: I1002 04:01:15.956070 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-6vllg" podStartSLOduration=2.365990343 podStartE2EDuration="2.956048365s" podCreationTimestamp="2025-10-02 04:01:13 +0000 UTC" firstStartedPulling="2025-10-02 04:01:14.118518998 +0000 UTC m=+8411.285263038" lastFinishedPulling="2025-10-02 04:01:14.70857702 +0000 UTC m=+8411.875321060" observedRunningTime="2025-10-02 04:01:15.949999448 +0000 UTC m=+8413.116743478" watchObservedRunningTime="2025-10-02 04:01:15.956048365 +0000 UTC m=+8413.122792445" Oct 02 04:01:37 crc kubenswrapper[4775]: I1002 04:01:37.233499 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:01:37 crc kubenswrapper[4775]: I1002 04:01:37.234202 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:01:37 crc kubenswrapper[4775]: I1002 04:01:37.234265 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 04:01:37 crc kubenswrapper[4775]: I1002 04:01:37.235172 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 04:01:37 crc kubenswrapper[4775]: I1002 04:01:37.235258 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" gracePeriod=600 Oct 02 04:01:37 crc kubenswrapper[4775]: E1002 04:01:37.379017 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:01:38 crc kubenswrapper[4775]: I1002 04:01:38.215809 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" exitCode=0 Oct 02 04:01:38 crc kubenswrapper[4775]: I1002 04:01:38.215924 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676"} Oct 02 04:01:38 crc kubenswrapper[4775]: I1002 04:01:38.216364 4775 scope.go:117] "RemoveContainer" containerID="bcdfff962e86b5bff70329682f81b7b093a8fcf67a837119fd8bdcdc3b8b7084" Oct 02 04:01:38 crc kubenswrapper[4775]: I1002 04:01:38.217411 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:01:38 crc kubenswrapper[4775]: E1002 04:01:38.217888 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:01:49 crc kubenswrapper[4775]: I1002 04:01:49.765629 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:01:49 crc kubenswrapper[4775]: E1002 04:01:49.766671 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:02:00 crc kubenswrapper[4775]: I1002 04:02:00.765544 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:02:00 crc kubenswrapper[4775]: E1002 04:02:00.766517 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:02:15 crc kubenswrapper[4775]: I1002 04:02:15.765820 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:02:15 crc kubenswrapper[4775]: E1002 04:02:15.766996 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:02:28 crc kubenswrapper[4775]: I1002 04:02:28.766773 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:02:28 crc kubenswrapper[4775]: E1002 04:02:28.769827 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:02:39 crc kubenswrapper[4775]: I1002 04:02:39.781588 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:02:39 crc kubenswrapper[4775]: E1002 04:02:39.782492 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:02:51 crc kubenswrapper[4775]: I1002 04:02:51.601053 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jg6pp"] Oct 02 04:02:51 crc kubenswrapper[4775]: I1002 04:02:51.617982 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:02:51 crc kubenswrapper[4775]: I1002 04:02:51.648094 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg6pp"] Oct 02 04:02:51 crc kubenswrapper[4775]: I1002 04:02:51.745699 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db76ec2a-9b26-44ca-913a-634543bf1807-catalog-content\") pod \"redhat-marketplace-jg6pp\" (UID: \"db76ec2a-9b26-44ca-913a-634543bf1807\") " pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:02:51 crc kubenswrapper[4775]: I1002 04:02:51.746066 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhw7c\" (UniqueName: \"kubernetes.io/projected/db76ec2a-9b26-44ca-913a-634543bf1807-kube-api-access-jhw7c\") pod \"redhat-marketplace-jg6pp\" (UID: \"db76ec2a-9b26-44ca-913a-634543bf1807\") " pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:02:51 crc kubenswrapper[4775]: I1002 04:02:51.746124 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db76ec2a-9b26-44ca-913a-634543bf1807-utilities\") pod \"redhat-marketplace-jg6pp\" (UID: \"db76ec2a-9b26-44ca-913a-634543bf1807\") " pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:02:51 crc kubenswrapper[4775]: I1002 04:02:51.768170 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:02:51 crc kubenswrapper[4775]: E1002 04:02:51.768398 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:02:51 crc kubenswrapper[4775]: I1002 04:02:51.848230 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db76ec2a-9b26-44ca-913a-634543bf1807-catalog-content\") pod \"redhat-marketplace-jg6pp\" (UID: \"db76ec2a-9b26-44ca-913a-634543bf1807\") " pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:02:51 crc kubenswrapper[4775]: I1002 04:02:51.848302 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhw7c\" (UniqueName: \"kubernetes.io/projected/db76ec2a-9b26-44ca-913a-634543bf1807-kube-api-access-jhw7c\") pod \"redhat-marketplace-jg6pp\" (UID: \"db76ec2a-9b26-44ca-913a-634543bf1807\") " pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:02:51 crc kubenswrapper[4775]: I1002 04:02:51.848349 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db76ec2a-9b26-44ca-913a-634543bf1807-utilities\") pod \"redhat-marketplace-jg6pp\" (UID: \"db76ec2a-9b26-44ca-913a-634543bf1807\") " pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:02:51 crc kubenswrapper[4775]: I1002 04:02:51.848901 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db76ec2a-9b26-44ca-913a-634543bf1807-utilities\") pod \"redhat-marketplace-jg6pp\" (UID: \"db76ec2a-9b26-44ca-913a-634543bf1807\") " pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:02:51 crc kubenswrapper[4775]: I1002 04:02:51.848910 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db76ec2a-9b26-44ca-913a-634543bf1807-catalog-content\") pod \"redhat-marketplace-jg6pp\" (UID: \"db76ec2a-9b26-44ca-913a-634543bf1807\") " pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:02:51 crc kubenswrapper[4775]: I1002 04:02:51.870712 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhw7c\" (UniqueName: \"kubernetes.io/projected/db76ec2a-9b26-44ca-913a-634543bf1807-kube-api-access-jhw7c\") pod \"redhat-marketplace-jg6pp\" (UID: \"db76ec2a-9b26-44ca-913a-634543bf1807\") " pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:02:51 crc kubenswrapper[4775]: I1002 04:02:51.958494 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:02:52 crc kubenswrapper[4775]: I1002 04:02:52.459804 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg6pp"] Oct 02 04:02:53 crc kubenswrapper[4775]: I1002 04:02:53.201571 4775 generic.go:334] "Generic (PLEG): container finished" podID="db76ec2a-9b26-44ca-913a-634543bf1807" containerID="564d1149540bdcbdbc3bc59b29cfb516fb24d4ddef64339b39ddb4f6db50e242" exitCode=0 Oct 02 04:02:53 crc kubenswrapper[4775]: I1002 04:02:53.201693 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg6pp" event={"ID":"db76ec2a-9b26-44ca-913a-634543bf1807","Type":"ContainerDied","Data":"564d1149540bdcbdbc3bc59b29cfb516fb24d4ddef64339b39ddb4f6db50e242"} Oct 02 04:02:53 crc kubenswrapper[4775]: I1002 04:02:53.203161 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg6pp" event={"ID":"db76ec2a-9b26-44ca-913a-634543bf1807","Type":"ContainerStarted","Data":"69dc968e68976671e4580450bd97e46515a059de344e26ec735892a1b97a8749"} Oct 02 04:02:54 crc kubenswrapper[4775]: I1002 04:02:54.217672 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg6pp" event={"ID":"db76ec2a-9b26-44ca-913a-634543bf1807","Type":"ContainerStarted","Data":"8a7c1e8d33ef61ecb9bcec18b8caf2febdc37feca760668b8f024903a00f9ac4"} Oct 02 04:02:55 crc kubenswrapper[4775]: I1002 04:02:55.232204 4775 generic.go:334] "Generic (PLEG): container finished" podID="db76ec2a-9b26-44ca-913a-634543bf1807" containerID="8a7c1e8d33ef61ecb9bcec18b8caf2febdc37feca760668b8f024903a00f9ac4" exitCode=0 Oct 02 04:02:55 crc kubenswrapper[4775]: I1002 04:02:55.232301 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg6pp" event={"ID":"db76ec2a-9b26-44ca-913a-634543bf1807","Type":"ContainerDied","Data":"8a7c1e8d33ef61ecb9bcec18b8caf2febdc37feca760668b8f024903a00f9ac4"} Oct 02 04:02:56 crc kubenswrapper[4775]: I1002 04:02:56.242540 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg6pp" event={"ID":"db76ec2a-9b26-44ca-913a-634543bf1807","Type":"ContainerStarted","Data":"03cb5be9262ef40b77be19bed6b3628029ebac807a09aff6411bd4fef17aafd8"} Oct 02 04:02:56 crc kubenswrapper[4775]: I1002 04:02:56.271376 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jg6pp" podStartSLOduration=2.714626366 podStartE2EDuration="5.271354004s" podCreationTimestamp="2025-10-02 04:02:51 +0000 UTC" firstStartedPulling="2025-10-02 04:02:53.203569352 +0000 UTC m=+8510.370313422" lastFinishedPulling="2025-10-02 04:02:55.76029701 +0000 UTC m=+8512.927041060" observedRunningTime="2025-10-02 04:02:56.264014743 +0000 UTC m=+8513.430758803" watchObservedRunningTime="2025-10-02 04:02:56.271354004 +0000 UTC m=+8513.438098044" Oct 02 04:03:01 crc kubenswrapper[4775]: I1002 04:03:01.958653 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:03:01 crc kubenswrapper[4775]: I1002 04:03:01.959343 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:03:02 crc kubenswrapper[4775]: I1002 04:03:02.056627 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:03:02 crc kubenswrapper[4775]: I1002 04:03:02.385510 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:03:02 crc kubenswrapper[4775]: I1002 04:03:02.457592 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg6pp"] Oct 02 04:03:04 crc kubenswrapper[4775]: I1002 04:03:04.338196 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jg6pp" podUID="db76ec2a-9b26-44ca-913a-634543bf1807" containerName="registry-server" containerID="cri-o://03cb5be9262ef40b77be19bed6b3628029ebac807a09aff6411bd4fef17aafd8" gracePeriod=2 Oct 02 04:03:04 crc kubenswrapper[4775]: I1002 04:03:04.893950 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:03:04 crc kubenswrapper[4775]: I1002 04:03:04.985860 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhw7c\" (UniqueName: \"kubernetes.io/projected/db76ec2a-9b26-44ca-913a-634543bf1807-kube-api-access-jhw7c\") pod \"db76ec2a-9b26-44ca-913a-634543bf1807\" (UID: \"db76ec2a-9b26-44ca-913a-634543bf1807\") " Oct 02 04:03:04 crc kubenswrapper[4775]: I1002 04:03:04.985937 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db76ec2a-9b26-44ca-913a-634543bf1807-utilities\") pod \"db76ec2a-9b26-44ca-913a-634543bf1807\" (UID: \"db76ec2a-9b26-44ca-913a-634543bf1807\") " Oct 02 04:03:04 crc kubenswrapper[4775]: I1002 04:03:04.986109 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db76ec2a-9b26-44ca-913a-634543bf1807-catalog-content\") pod \"db76ec2a-9b26-44ca-913a-634543bf1807\" (UID: \"db76ec2a-9b26-44ca-913a-634543bf1807\") " Oct 02 04:03:04 crc kubenswrapper[4775]: I1002 04:03:04.987142 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db76ec2a-9b26-44ca-913a-634543bf1807-utilities" (OuterVolumeSpecName: "utilities") pod "db76ec2a-9b26-44ca-913a-634543bf1807" (UID: "db76ec2a-9b26-44ca-913a-634543bf1807"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:03:04 crc kubenswrapper[4775]: I1002 04:03:04.997717 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db76ec2a-9b26-44ca-913a-634543bf1807-kube-api-access-jhw7c" (OuterVolumeSpecName: "kube-api-access-jhw7c") pod "db76ec2a-9b26-44ca-913a-634543bf1807" (UID: "db76ec2a-9b26-44ca-913a-634543bf1807"). InnerVolumeSpecName "kube-api-access-jhw7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.003438 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db76ec2a-9b26-44ca-913a-634543bf1807-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "db76ec2a-9b26-44ca-913a-634543bf1807" (UID: "db76ec2a-9b26-44ca-913a-634543bf1807"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.089406 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhw7c\" (UniqueName: \"kubernetes.io/projected/db76ec2a-9b26-44ca-913a-634543bf1807-kube-api-access-jhw7c\") on node \"crc\" DevicePath \"\"" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.089792 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/db76ec2a-9b26-44ca-913a-634543bf1807-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.089811 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/db76ec2a-9b26-44ca-913a-634543bf1807-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.354722 4775 generic.go:334] "Generic (PLEG): container finished" podID="db76ec2a-9b26-44ca-913a-634543bf1807" containerID="03cb5be9262ef40b77be19bed6b3628029ebac807a09aff6411bd4fef17aafd8" exitCode=0 Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.354787 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg6pp" event={"ID":"db76ec2a-9b26-44ca-913a-634543bf1807","Type":"ContainerDied","Data":"03cb5be9262ef40b77be19bed6b3628029ebac807a09aff6411bd4fef17aafd8"} Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.354815 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jg6pp" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.354828 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jg6pp" event={"ID":"db76ec2a-9b26-44ca-913a-634543bf1807","Type":"ContainerDied","Data":"69dc968e68976671e4580450bd97e46515a059de344e26ec735892a1b97a8749"} Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.354857 4775 scope.go:117] "RemoveContainer" containerID="03cb5be9262ef40b77be19bed6b3628029ebac807a09aff6411bd4fef17aafd8" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.400846 4775 scope.go:117] "RemoveContainer" containerID="8a7c1e8d33ef61ecb9bcec18b8caf2febdc37feca760668b8f024903a00f9ac4" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.413473 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg6pp"] Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.432884 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jg6pp"] Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.437906 4775 scope.go:117] "RemoveContainer" containerID="564d1149540bdcbdbc3bc59b29cfb516fb24d4ddef64339b39ddb4f6db50e242" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.489058 4775 scope.go:117] "RemoveContainer" containerID="03cb5be9262ef40b77be19bed6b3628029ebac807a09aff6411bd4fef17aafd8" Oct 02 04:03:05 crc kubenswrapper[4775]: E1002 04:03:05.489774 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03cb5be9262ef40b77be19bed6b3628029ebac807a09aff6411bd4fef17aafd8\": container with ID starting with 03cb5be9262ef40b77be19bed6b3628029ebac807a09aff6411bd4fef17aafd8 not found: ID does not exist" containerID="03cb5be9262ef40b77be19bed6b3628029ebac807a09aff6411bd4fef17aafd8" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.489837 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03cb5be9262ef40b77be19bed6b3628029ebac807a09aff6411bd4fef17aafd8"} err="failed to get container status \"03cb5be9262ef40b77be19bed6b3628029ebac807a09aff6411bd4fef17aafd8\": rpc error: code = NotFound desc = could not find container \"03cb5be9262ef40b77be19bed6b3628029ebac807a09aff6411bd4fef17aafd8\": container with ID starting with 03cb5be9262ef40b77be19bed6b3628029ebac807a09aff6411bd4fef17aafd8 not found: ID does not exist" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.489887 4775 scope.go:117] "RemoveContainer" containerID="8a7c1e8d33ef61ecb9bcec18b8caf2febdc37feca760668b8f024903a00f9ac4" Oct 02 04:03:05 crc kubenswrapper[4775]: E1002 04:03:05.490334 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a7c1e8d33ef61ecb9bcec18b8caf2febdc37feca760668b8f024903a00f9ac4\": container with ID starting with 8a7c1e8d33ef61ecb9bcec18b8caf2febdc37feca760668b8f024903a00f9ac4 not found: ID does not exist" containerID="8a7c1e8d33ef61ecb9bcec18b8caf2febdc37feca760668b8f024903a00f9ac4" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.490374 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a7c1e8d33ef61ecb9bcec18b8caf2febdc37feca760668b8f024903a00f9ac4"} err="failed to get container status \"8a7c1e8d33ef61ecb9bcec18b8caf2febdc37feca760668b8f024903a00f9ac4\": rpc error: code = NotFound desc = could not find container \"8a7c1e8d33ef61ecb9bcec18b8caf2febdc37feca760668b8f024903a00f9ac4\": container with ID starting with 8a7c1e8d33ef61ecb9bcec18b8caf2febdc37feca760668b8f024903a00f9ac4 not found: ID does not exist" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.490403 4775 scope.go:117] "RemoveContainer" containerID="564d1149540bdcbdbc3bc59b29cfb516fb24d4ddef64339b39ddb4f6db50e242" Oct 02 04:03:05 crc kubenswrapper[4775]: E1002 04:03:05.490736 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"564d1149540bdcbdbc3bc59b29cfb516fb24d4ddef64339b39ddb4f6db50e242\": container with ID starting with 564d1149540bdcbdbc3bc59b29cfb516fb24d4ddef64339b39ddb4f6db50e242 not found: ID does not exist" containerID="564d1149540bdcbdbc3bc59b29cfb516fb24d4ddef64339b39ddb4f6db50e242" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.490775 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"564d1149540bdcbdbc3bc59b29cfb516fb24d4ddef64339b39ddb4f6db50e242"} err="failed to get container status \"564d1149540bdcbdbc3bc59b29cfb516fb24d4ddef64339b39ddb4f6db50e242\": rpc error: code = NotFound desc = could not find container \"564d1149540bdcbdbc3bc59b29cfb516fb24d4ddef64339b39ddb4f6db50e242\": container with ID starting with 564d1149540bdcbdbc3bc59b29cfb516fb24d4ddef64339b39ddb4f6db50e242 not found: ID does not exist" Oct 02 04:03:05 crc kubenswrapper[4775]: I1002 04:03:05.782222 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db76ec2a-9b26-44ca-913a-634543bf1807" path="/var/lib/kubelet/pods/db76ec2a-9b26-44ca-913a-634543bf1807/volumes" Oct 02 04:03:06 crc kubenswrapper[4775]: I1002 04:03:06.765802 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:03:06 crc kubenswrapper[4775]: E1002 04:03:06.766346 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:03:21 crc kubenswrapper[4775]: I1002 04:03:21.765736 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:03:21 crc kubenswrapper[4775]: E1002 04:03:21.768040 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:03:33 crc kubenswrapper[4775]: I1002 04:03:33.776706 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:03:33 crc kubenswrapper[4775]: E1002 04:03:33.777743 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:03:45 crc kubenswrapper[4775]: I1002 04:03:45.808872 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q4xvg"] Oct 02 04:03:45 crc kubenswrapper[4775]: E1002 04:03:45.809709 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db76ec2a-9b26-44ca-913a-634543bf1807" containerName="extract-content" Oct 02 04:03:45 crc kubenswrapper[4775]: I1002 04:03:45.809722 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="db76ec2a-9b26-44ca-913a-634543bf1807" containerName="extract-content" Oct 02 04:03:45 crc kubenswrapper[4775]: E1002 04:03:45.809744 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db76ec2a-9b26-44ca-913a-634543bf1807" containerName="extract-utilities" Oct 02 04:03:45 crc kubenswrapper[4775]: I1002 04:03:45.809750 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="db76ec2a-9b26-44ca-913a-634543bf1807" containerName="extract-utilities" Oct 02 04:03:45 crc kubenswrapper[4775]: E1002 04:03:45.809780 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db76ec2a-9b26-44ca-913a-634543bf1807" containerName="registry-server" Oct 02 04:03:45 crc kubenswrapper[4775]: I1002 04:03:45.809786 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="db76ec2a-9b26-44ca-913a-634543bf1807" containerName="registry-server" Oct 02 04:03:45 crc kubenswrapper[4775]: I1002 04:03:45.810009 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="db76ec2a-9b26-44ca-913a-634543bf1807" containerName="registry-server" Oct 02 04:03:45 crc kubenswrapper[4775]: I1002 04:03:45.811475 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:45 crc kubenswrapper[4775]: I1002 04:03:45.837809 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q4xvg"] Oct 02 04:03:45 crc kubenswrapper[4775]: I1002 04:03:45.967476 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f9718b-baaa-4143-8a7e-af529970b490-catalog-content\") pod \"community-operators-q4xvg\" (UID: \"17f9718b-baaa-4143-8a7e-af529970b490\") " pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:45 crc kubenswrapper[4775]: I1002 04:03:45.967515 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f9718b-baaa-4143-8a7e-af529970b490-utilities\") pod \"community-operators-q4xvg\" (UID: \"17f9718b-baaa-4143-8a7e-af529970b490\") " pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:45 crc kubenswrapper[4775]: I1002 04:03:45.967880 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bx77\" (UniqueName: \"kubernetes.io/projected/17f9718b-baaa-4143-8a7e-af529970b490-kube-api-access-9bx77\") pod \"community-operators-q4xvg\" (UID: \"17f9718b-baaa-4143-8a7e-af529970b490\") " pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:46 crc kubenswrapper[4775]: I1002 04:03:46.069736 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bx77\" (UniqueName: \"kubernetes.io/projected/17f9718b-baaa-4143-8a7e-af529970b490-kube-api-access-9bx77\") pod \"community-operators-q4xvg\" (UID: \"17f9718b-baaa-4143-8a7e-af529970b490\") " pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:46 crc kubenswrapper[4775]: I1002 04:03:46.069937 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f9718b-baaa-4143-8a7e-af529970b490-catalog-content\") pod \"community-operators-q4xvg\" (UID: \"17f9718b-baaa-4143-8a7e-af529970b490\") " pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:46 crc kubenswrapper[4775]: I1002 04:03:46.069970 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f9718b-baaa-4143-8a7e-af529970b490-utilities\") pod \"community-operators-q4xvg\" (UID: \"17f9718b-baaa-4143-8a7e-af529970b490\") " pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:46 crc kubenswrapper[4775]: I1002 04:03:46.070442 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f9718b-baaa-4143-8a7e-af529970b490-catalog-content\") pod \"community-operators-q4xvg\" (UID: \"17f9718b-baaa-4143-8a7e-af529970b490\") " pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:46 crc kubenswrapper[4775]: I1002 04:03:46.071302 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f9718b-baaa-4143-8a7e-af529970b490-utilities\") pod \"community-operators-q4xvg\" (UID: \"17f9718b-baaa-4143-8a7e-af529970b490\") " pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:46 crc kubenswrapper[4775]: I1002 04:03:46.107598 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bx77\" (UniqueName: \"kubernetes.io/projected/17f9718b-baaa-4143-8a7e-af529970b490-kube-api-access-9bx77\") pod \"community-operators-q4xvg\" (UID: \"17f9718b-baaa-4143-8a7e-af529970b490\") " pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:46 crc kubenswrapper[4775]: I1002 04:03:46.149014 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:46 crc kubenswrapper[4775]: I1002 04:03:46.717400 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q4xvg"] Oct 02 04:03:46 crc kubenswrapper[4775]: W1002 04:03:46.723516 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17f9718b_baaa_4143_8a7e_af529970b490.slice/crio-de04b74eedb85acc6a103adda1ba3a14096a3e8939749f4228363e1406f308dd WatchSource:0}: Error finding container de04b74eedb85acc6a103adda1ba3a14096a3e8939749f4228363e1406f308dd: Status 404 returned error can't find the container with id de04b74eedb85acc6a103adda1ba3a14096a3e8939749f4228363e1406f308dd Oct 02 04:03:46 crc kubenswrapper[4775]: I1002 04:03:46.765542 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:03:46 crc kubenswrapper[4775]: E1002 04:03:46.765914 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:03:46 crc kubenswrapper[4775]: I1002 04:03:46.914564 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q4xvg" event={"ID":"17f9718b-baaa-4143-8a7e-af529970b490","Type":"ContainerStarted","Data":"de04b74eedb85acc6a103adda1ba3a14096a3e8939749f4228363e1406f308dd"} Oct 02 04:03:47 crc kubenswrapper[4775]: I1002 04:03:47.930408 4775 generic.go:334] "Generic (PLEG): container finished" podID="17f9718b-baaa-4143-8a7e-af529970b490" containerID="353911fb03bea40e1c92b570ad806e3d911d4d4fad8a708cd3d642d874730230" exitCode=0 Oct 02 04:03:47 crc kubenswrapper[4775]: I1002 04:03:47.930467 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q4xvg" event={"ID":"17f9718b-baaa-4143-8a7e-af529970b490","Type":"ContainerDied","Data":"353911fb03bea40e1c92b570ad806e3d911d4d4fad8a708cd3d642d874730230"} Oct 02 04:03:49 crc kubenswrapper[4775]: I1002 04:03:49.959414 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q4xvg" event={"ID":"17f9718b-baaa-4143-8a7e-af529970b490","Type":"ContainerStarted","Data":"616bcd666239c99e28f5a1747be42327027aa797f2f3366eda01f1bca1a1738d"} Oct 02 04:03:50 crc kubenswrapper[4775]: I1002 04:03:50.969214 4775 generic.go:334] "Generic (PLEG): container finished" podID="17f9718b-baaa-4143-8a7e-af529970b490" containerID="616bcd666239c99e28f5a1747be42327027aa797f2f3366eda01f1bca1a1738d" exitCode=0 Oct 02 04:03:50 crc kubenswrapper[4775]: I1002 04:03:50.969265 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q4xvg" event={"ID":"17f9718b-baaa-4143-8a7e-af529970b490","Type":"ContainerDied","Data":"616bcd666239c99e28f5a1747be42327027aa797f2f3366eda01f1bca1a1738d"} Oct 02 04:03:51 crc kubenswrapper[4775]: I1002 04:03:51.981610 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q4xvg" event={"ID":"17f9718b-baaa-4143-8a7e-af529970b490","Type":"ContainerStarted","Data":"c290ed04d6859e7a0ad5345bcccef6460a431e625b8a951cf0952c569b99161c"} Oct 02 04:03:52 crc kubenswrapper[4775]: I1002 04:03:52.018996 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q4xvg" podStartSLOduration=3.565555877 podStartE2EDuration="7.018946864s" podCreationTimestamp="2025-10-02 04:03:45 +0000 UTC" firstStartedPulling="2025-10-02 04:03:47.934473285 +0000 UTC m=+8565.101217355" lastFinishedPulling="2025-10-02 04:03:51.387864292 +0000 UTC m=+8568.554608342" observedRunningTime="2025-10-02 04:03:52.009473768 +0000 UTC m=+8569.176217838" watchObservedRunningTime="2025-10-02 04:03:52.018946864 +0000 UTC m=+8569.185690944" Oct 02 04:03:56 crc kubenswrapper[4775]: I1002 04:03:56.149652 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:56 crc kubenswrapper[4775]: I1002 04:03:56.150109 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:56 crc kubenswrapper[4775]: I1002 04:03:56.209521 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:57 crc kubenswrapper[4775]: I1002 04:03:57.119237 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:57 crc kubenswrapper[4775]: I1002 04:03:57.207719 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q4xvg"] Oct 02 04:03:59 crc kubenswrapper[4775]: I1002 04:03:59.065556 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-q4xvg" podUID="17f9718b-baaa-4143-8a7e-af529970b490" containerName="registry-server" containerID="cri-o://c290ed04d6859e7a0ad5345bcccef6460a431e625b8a951cf0952c569b99161c" gracePeriod=2 Oct 02 04:03:59 crc kubenswrapper[4775]: I1002 04:03:59.628041 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:03:59 crc kubenswrapper[4775]: I1002 04:03:59.682726 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f9718b-baaa-4143-8a7e-af529970b490-catalog-content\") pod \"17f9718b-baaa-4143-8a7e-af529970b490\" (UID: \"17f9718b-baaa-4143-8a7e-af529970b490\") " Oct 02 04:03:59 crc kubenswrapper[4775]: I1002 04:03:59.683283 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f9718b-baaa-4143-8a7e-af529970b490-utilities\") pod \"17f9718b-baaa-4143-8a7e-af529970b490\" (UID: \"17f9718b-baaa-4143-8a7e-af529970b490\") " Oct 02 04:03:59 crc kubenswrapper[4775]: I1002 04:03:59.683581 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bx77\" (UniqueName: \"kubernetes.io/projected/17f9718b-baaa-4143-8a7e-af529970b490-kube-api-access-9bx77\") pod \"17f9718b-baaa-4143-8a7e-af529970b490\" (UID: \"17f9718b-baaa-4143-8a7e-af529970b490\") " Oct 02 04:03:59 crc kubenswrapper[4775]: I1002 04:03:59.684328 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17f9718b-baaa-4143-8a7e-af529970b490-utilities" (OuterVolumeSpecName: "utilities") pod "17f9718b-baaa-4143-8a7e-af529970b490" (UID: "17f9718b-baaa-4143-8a7e-af529970b490"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:03:59 crc kubenswrapper[4775]: I1002 04:03:59.684747 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17f9718b-baaa-4143-8a7e-af529970b490-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 04:03:59 crc kubenswrapper[4775]: I1002 04:03:59.699760 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17f9718b-baaa-4143-8a7e-af529970b490-kube-api-access-9bx77" (OuterVolumeSpecName: "kube-api-access-9bx77") pod "17f9718b-baaa-4143-8a7e-af529970b490" (UID: "17f9718b-baaa-4143-8a7e-af529970b490"). InnerVolumeSpecName "kube-api-access-9bx77". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:03:59 crc kubenswrapper[4775]: I1002 04:03:59.765801 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:03:59 crc kubenswrapper[4775]: E1002 04:03:59.766332 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:03:59 crc kubenswrapper[4775]: I1002 04:03:59.786904 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bx77\" (UniqueName: \"kubernetes.io/projected/17f9718b-baaa-4143-8a7e-af529970b490-kube-api-access-9bx77\") on node \"crc\" DevicePath \"\"" Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.075330 4775 generic.go:334] "Generic (PLEG): container finished" podID="17f9718b-baaa-4143-8a7e-af529970b490" containerID="c290ed04d6859e7a0ad5345bcccef6460a431e625b8a951cf0952c569b99161c" exitCode=0 Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.075369 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q4xvg" event={"ID":"17f9718b-baaa-4143-8a7e-af529970b490","Type":"ContainerDied","Data":"c290ed04d6859e7a0ad5345bcccef6460a431e625b8a951cf0952c569b99161c"} Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.075395 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q4xvg" event={"ID":"17f9718b-baaa-4143-8a7e-af529970b490","Type":"ContainerDied","Data":"de04b74eedb85acc6a103adda1ba3a14096a3e8939749f4228363e1406f308dd"} Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.075413 4775 scope.go:117] "RemoveContainer" containerID="c290ed04d6859e7a0ad5345bcccef6460a431e625b8a951cf0952c569b99161c" Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.075544 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q4xvg" Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.103096 4775 scope.go:117] "RemoveContainer" containerID="616bcd666239c99e28f5a1747be42327027aa797f2f3366eda01f1bca1a1738d" Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.147679 4775 scope.go:117] "RemoveContainer" containerID="353911fb03bea40e1c92b570ad806e3d911d4d4fad8a708cd3d642d874730230" Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.181275 4775 scope.go:117] "RemoveContainer" containerID="c290ed04d6859e7a0ad5345bcccef6460a431e625b8a951cf0952c569b99161c" Oct 02 04:04:00 crc kubenswrapper[4775]: E1002 04:04:00.181822 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c290ed04d6859e7a0ad5345bcccef6460a431e625b8a951cf0952c569b99161c\": container with ID starting with c290ed04d6859e7a0ad5345bcccef6460a431e625b8a951cf0952c569b99161c not found: ID does not exist" containerID="c290ed04d6859e7a0ad5345bcccef6460a431e625b8a951cf0952c569b99161c" Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.181855 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c290ed04d6859e7a0ad5345bcccef6460a431e625b8a951cf0952c569b99161c"} err="failed to get container status \"c290ed04d6859e7a0ad5345bcccef6460a431e625b8a951cf0952c569b99161c\": rpc error: code = NotFound desc = could not find container \"c290ed04d6859e7a0ad5345bcccef6460a431e625b8a951cf0952c569b99161c\": container with ID starting with c290ed04d6859e7a0ad5345bcccef6460a431e625b8a951cf0952c569b99161c not found: ID does not exist" Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.181877 4775 scope.go:117] "RemoveContainer" containerID="616bcd666239c99e28f5a1747be42327027aa797f2f3366eda01f1bca1a1738d" Oct 02 04:04:00 crc kubenswrapper[4775]: E1002 04:04:00.183252 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"616bcd666239c99e28f5a1747be42327027aa797f2f3366eda01f1bca1a1738d\": container with ID starting with 616bcd666239c99e28f5a1747be42327027aa797f2f3366eda01f1bca1a1738d not found: ID does not exist" containerID="616bcd666239c99e28f5a1747be42327027aa797f2f3366eda01f1bca1a1738d" Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.183301 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616bcd666239c99e28f5a1747be42327027aa797f2f3366eda01f1bca1a1738d"} err="failed to get container status \"616bcd666239c99e28f5a1747be42327027aa797f2f3366eda01f1bca1a1738d\": rpc error: code = NotFound desc = could not find container \"616bcd666239c99e28f5a1747be42327027aa797f2f3366eda01f1bca1a1738d\": container with ID starting with 616bcd666239c99e28f5a1747be42327027aa797f2f3366eda01f1bca1a1738d not found: ID does not exist" Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.183337 4775 scope.go:117] "RemoveContainer" containerID="353911fb03bea40e1c92b570ad806e3d911d4d4fad8a708cd3d642d874730230" Oct 02 04:04:00 crc kubenswrapper[4775]: E1002 04:04:00.183755 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"353911fb03bea40e1c92b570ad806e3d911d4d4fad8a708cd3d642d874730230\": container with ID starting with 353911fb03bea40e1c92b570ad806e3d911d4d4fad8a708cd3d642d874730230 not found: ID does not exist" containerID="353911fb03bea40e1c92b570ad806e3d911d4d4fad8a708cd3d642d874730230" Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.183811 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"353911fb03bea40e1c92b570ad806e3d911d4d4fad8a708cd3d642d874730230"} err="failed to get container status \"353911fb03bea40e1c92b570ad806e3d911d4d4fad8a708cd3d642d874730230\": rpc error: code = NotFound desc = could not find container \"353911fb03bea40e1c92b570ad806e3d911d4d4fad8a708cd3d642d874730230\": container with ID starting with 353911fb03bea40e1c92b570ad806e3d911d4d4fad8a708cd3d642d874730230 not found: ID does not exist" Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.197003 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17f9718b-baaa-4143-8a7e-af529970b490-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17f9718b-baaa-4143-8a7e-af529970b490" (UID: "17f9718b-baaa-4143-8a7e-af529970b490"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.296718 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17f9718b-baaa-4143-8a7e-af529970b490-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.409917 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-q4xvg"] Oct 02 04:04:00 crc kubenswrapper[4775]: I1002 04:04:00.417793 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-q4xvg"] Oct 02 04:04:01 crc kubenswrapper[4775]: I1002 04:04:01.790063 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17f9718b-baaa-4143-8a7e-af529970b490" path="/var/lib/kubelet/pods/17f9718b-baaa-4143-8a7e-af529970b490/volumes" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.460178 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9jpn5"] Oct 02 04:04:05 crc kubenswrapper[4775]: E1002 04:04:05.463527 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f9718b-baaa-4143-8a7e-af529970b490" containerName="registry-server" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.463654 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f9718b-baaa-4143-8a7e-af529970b490" containerName="registry-server" Oct 02 04:04:05 crc kubenswrapper[4775]: E1002 04:04:05.463680 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f9718b-baaa-4143-8a7e-af529970b490" containerName="extract-content" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.463687 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f9718b-baaa-4143-8a7e-af529970b490" containerName="extract-content" Oct 02 04:04:05 crc kubenswrapper[4775]: E1002 04:04:05.463729 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17f9718b-baaa-4143-8a7e-af529970b490" containerName="extract-utilities" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.463738 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="17f9718b-baaa-4143-8a7e-af529970b490" containerName="extract-utilities" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.466723 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="17f9718b-baaa-4143-8a7e-af529970b490" containerName="registry-server" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.473885 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.521156 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-utilities\") pod \"certified-operators-9jpn5\" (UID: \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\") " pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.521291 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-catalog-content\") pod \"certified-operators-9jpn5\" (UID: \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\") " pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.521330 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w42tp\" (UniqueName: \"kubernetes.io/projected/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-kube-api-access-w42tp\") pod \"certified-operators-9jpn5\" (UID: \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\") " pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.522299 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9jpn5"] Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.623743 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-utilities\") pod \"certified-operators-9jpn5\" (UID: \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\") " pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.623923 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-catalog-content\") pod \"certified-operators-9jpn5\" (UID: \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\") " pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.623982 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w42tp\" (UniqueName: \"kubernetes.io/projected/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-kube-api-access-w42tp\") pod \"certified-operators-9jpn5\" (UID: \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\") " pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.624432 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-utilities\") pod \"certified-operators-9jpn5\" (UID: \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\") " pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.624468 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-catalog-content\") pod \"certified-operators-9jpn5\" (UID: \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\") " pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.659819 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w42tp\" (UniqueName: \"kubernetes.io/projected/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-kube-api-access-w42tp\") pod \"certified-operators-9jpn5\" (UID: \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\") " pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:05 crc kubenswrapper[4775]: I1002 04:04:05.825826 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:06 crc kubenswrapper[4775]: I1002 04:04:06.371076 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9jpn5"] Oct 02 04:04:07 crc kubenswrapper[4775]: I1002 04:04:07.168158 4775 generic.go:334] "Generic (PLEG): container finished" podID="62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" containerID="3d64be3c029ce5e091ea3a8984acd091426e9724d6327302aae145814c3dc38d" exitCode=0 Oct 02 04:04:07 crc kubenswrapper[4775]: I1002 04:04:07.168487 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jpn5" event={"ID":"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c","Type":"ContainerDied","Data":"3d64be3c029ce5e091ea3a8984acd091426e9724d6327302aae145814c3dc38d"} Oct 02 04:04:07 crc kubenswrapper[4775]: I1002 04:04:07.168526 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jpn5" event={"ID":"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c","Type":"ContainerStarted","Data":"8e200268ad0e5f405f05c99d8d616709d33a7169d55c923ae0441ad17461b605"} Oct 02 04:04:09 crc kubenswrapper[4775]: I1002 04:04:09.191924 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jpn5" event={"ID":"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c","Type":"ContainerStarted","Data":"508b0d2fe115b20cdeab37ff1e7e5fd8e19464c1e8a756b404a64040ebc4b866"} Oct 02 04:04:11 crc kubenswrapper[4775]: E1002 04:04:11.760070 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62a5dbfd_1260_4c70_b3b6_e0d33fa3971c.slice/crio-508b0d2fe115b20cdeab37ff1e7e5fd8e19464c1e8a756b404a64040ebc4b866.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62a5dbfd_1260_4c70_b3b6_e0d33fa3971c.slice/crio-conmon-508b0d2fe115b20cdeab37ff1e7e5fd8e19464c1e8a756b404a64040ebc4b866.scope\": RecentStats: unable to find data in memory cache]" Oct 02 04:04:12 crc kubenswrapper[4775]: I1002 04:04:12.226582 4775 generic.go:334] "Generic (PLEG): container finished" podID="62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" containerID="508b0d2fe115b20cdeab37ff1e7e5fd8e19464c1e8a756b404a64040ebc4b866" exitCode=0 Oct 02 04:04:12 crc kubenswrapper[4775]: I1002 04:04:12.226625 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jpn5" event={"ID":"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c","Type":"ContainerDied","Data":"508b0d2fe115b20cdeab37ff1e7e5fd8e19464c1e8a756b404a64040ebc4b866"} Oct 02 04:04:13 crc kubenswrapper[4775]: I1002 04:04:13.781139 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:04:13 crc kubenswrapper[4775]: E1002 04:04:13.781678 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:04:14 crc kubenswrapper[4775]: I1002 04:04:14.251622 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jpn5" event={"ID":"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c","Type":"ContainerStarted","Data":"328b7d3d9071b2a32d653609f8ba807de582be446eac79f76fe35da13b22d48f"} Oct 02 04:04:14 crc kubenswrapper[4775]: I1002 04:04:14.277128 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9jpn5" podStartSLOduration=3.486060015 podStartE2EDuration="9.27710827s" podCreationTimestamp="2025-10-02 04:04:05 +0000 UTC" firstStartedPulling="2025-10-02 04:04:07.173677718 +0000 UTC m=+8584.340421748" lastFinishedPulling="2025-10-02 04:04:12.964725963 +0000 UTC m=+8590.131470003" observedRunningTime="2025-10-02 04:04:14.273405534 +0000 UTC m=+8591.440149574" watchObservedRunningTime="2025-10-02 04:04:14.27710827 +0000 UTC m=+8591.443852340" Oct 02 04:04:15 crc kubenswrapper[4775]: I1002 04:04:15.826993 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:15 crc kubenswrapper[4775]: I1002 04:04:15.827526 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:15 crc kubenswrapper[4775]: I1002 04:04:15.905870 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:25 crc kubenswrapper[4775]: I1002 04:04:25.895557 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:25 crc kubenswrapper[4775]: I1002 04:04:25.962226 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9jpn5"] Oct 02 04:04:26 crc kubenswrapper[4775]: I1002 04:04:26.431932 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9jpn5" podUID="62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" containerName="registry-server" containerID="cri-o://328b7d3d9071b2a32d653609f8ba807de582be446eac79f76fe35da13b22d48f" gracePeriod=2 Oct 02 04:04:26 crc kubenswrapper[4775]: I1002 04:04:26.766343 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:04:26 crc kubenswrapper[4775]: E1002 04:04:26.767157 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:04:26 crc kubenswrapper[4775]: I1002 04:04:26.956465 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.102165 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w42tp\" (UniqueName: \"kubernetes.io/projected/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-kube-api-access-w42tp\") pod \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\" (UID: \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\") " Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.103779 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-catalog-content\") pod \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\" (UID: \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\") " Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.104329 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-utilities\") pod \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\" (UID: \"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c\") " Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.105838 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-utilities" (OuterVolumeSpecName: "utilities") pod "62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" (UID: "62a5dbfd-1260-4c70-b3b6-e0d33fa3971c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.107309 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.115340 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-kube-api-access-w42tp" (OuterVolumeSpecName: "kube-api-access-w42tp") pod "62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" (UID: "62a5dbfd-1260-4c70-b3b6-e0d33fa3971c"). InnerVolumeSpecName "kube-api-access-w42tp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.170277 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" (UID: "62a5dbfd-1260-4c70-b3b6-e0d33fa3971c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.209639 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w42tp\" (UniqueName: \"kubernetes.io/projected/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-kube-api-access-w42tp\") on node \"crc\" DevicePath \"\"" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.209669 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.471482 4775 generic.go:334] "Generic (PLEG): container finished" podID="62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" containerID="328b7d3d9071b2a32d653609f8ba807de582be446eac79f76fe35da13b22d48f" exitCode=0 Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.471892 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9jpn5" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.471951 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jpn5" event={"ID":"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c","Type":"ContainerDied","Data":"328b7d3d9071b2a32d653609f8ba807de582be446eac79f76fe35da13b22d48f"} Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.472082 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9jpn5" event={"ID":"62a5dbfd-1260-4c70-b3b6-e0d33fa3971c","Type":"ContainerDied","Data":"8e200268ad0e5f405f05c99d8d616709d33a7169d55c923ae0441ad17461b605"} Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.472698 4775 scope.go:117] "RemoveContainer" containerID="328b7d3d9071b2a32d653609f8ba807de582be446eac79f76fe35da13b22d48f" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.501734 4775 scope.go:117] "RemoveContainer" containerID="508b0d2fe115b20cdeab37ff1e7e5fd8e19464c1e8a756b404a64040ebc4b866" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.524087 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9jpn5"] Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.535408 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9jpn5"] Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.546837 4775 scope.go:117] "RemoveContainer" containerID="3d64be3c029ce5e091ea3a8984acd091426e9724d6327302aae145814c3dc38d" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.599832 4775 scope.go:117] "RemoveContainer" containerID="328b7d3d9071b2a32d653609f8ba807de582be446eac79f76fe35da13b22d48f" Oct 02 04:04:27 crc kubenswrapper[4775]: E1002 04:04:27.600357 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"328b7d3d9071b2a32d653609f8ba807de582be446eac79f76fe35da13b22d48f\": container with ID starting with 328b7d3d9071b2a32d653609f8ba807de582be446eac79f76fe35da13b22d48f not found: ID does not exist" containerID="328b7d3d9071b2a32d653609f8ba807de582be446eac79f76fe35da13b22d48f" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.600400 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"328b7d3d9071b2a32d653609f8ba807de582be446eac79f76fe35da13b22d48f"} err="failed to get container status \"328b7d3d9071b2a32d653609f8ba807de582be446eac79f76fe35da13b22d48f\": rpc error: code = NotFound desc = could not find container \"328b7d3d9071b2a32d653609f8ba807de582be446eac79f76fe35da13b22d48f\": container with ID starting with 328b7d3d9071b2a32d653609f8ba807de582be446eac79f76fe35da13b22d48f not found: ID does not exist" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.600425 4775 scope.go:117] "RemoveContainer" containerID="508b0d2fe115b20cdeab37ff1e7e5fd8e19464c1e8a756b404a64040ebc4b866" Oct 02 04:04:27 crc kubenswrapper[4775]: E1002 04:04:27.600845 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"508b0d2fe115b20cdeab37ff1e7e5fd8e19464c1e8a756b404a64040ebc4b866\": container with ID starting with 508b0d2fe115b20cdeab37ff1e7e5fd8e19464c1e8a756b404a64040ebc4b866 not found: ID does not exist" containerID="508b0d2fe115b20cdeab37ff1e7e5fd8e19464c1e8a756b404a64040ebc4b866" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.600874 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"508b0d2fe115b20cdeab37ff1e7e5fd8e19464c1e8a756b404a64040ebc4b866"} err="failed to get container status \"508b0d2fe115b20cdeab37ff1e7e5fd8e19464c1e8a756b404a64040ebc4b866\": rpc error: code = NotFound desc = could not find container \"508b0d2fe115b20cdeab37ff1e7e5fd8e19464c1e8a756b404a64040ebc4b866\": container with ID starting with 508b0d2fe115b20cdeab37ff1e7e5fd8e19464c1e8a756b404a64040ebc4b866 not found: ID does not exist" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.600893 4775 scope.go:117] "RemoveContainer" containerID="3d64be3c029ce5e091ea3a8984acd091426e9724d6327302aae145814c3dc38d" Oct 02 04:04:27 crc kubenswrapper[4775]: E1002 04:04:27.601144 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d64be3c029ce5e091ea3a8984acd091426e9724d6327302aae145814c3dc38d\": container with ID starting with 3d64be3c029ce5e091ea3a8984acd091426e9724d6327302aae145814c3dc38d not found: ID does not exist" containerID="3d64be3c029ce5e091ea3a8984acd091426e9724d6327302aae145814c3dc38d" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.601166 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d64be3c029ce5e091ea3a8984acd091426e9724d6327302aae145814c3dc38d"} err="failed to get container status \"3d64be3c029ce5e091ea3a8984acd091426e9724d6327302aae145814c3dc38d\": rpc error: code = NotFound desc = could not find container \"3d64be3c029ce5e091ea3a8984acd091426e9724d6327302aae145814c3dc38d\": container with ID starting with 3d64be3c029ce5e091ea3a8984acd091426e9724d6327302aae145814c3dc38d not found: ID does not exist" Oct 02 04:04:27 crc kubenswrapper[4775]: I1002 04:04:27.778737 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" path="/var/lib/kubelet/pods/62a5dbfd-1260-4c70-b3b6-e0d33fa3971c/volumes" Oct 02 04:04:37 crc kubenswrapper[4775]: I1002 04:04:37.766207 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:04:37 crc kubenswrapper[4775]: E1002 04:04:37.770440 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:04:48 crc kubenswrapper[4775]: I1002 04:04:48.767976 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:04:48 crc kubenswrapper[4775]: E1002 04:04:48.768884 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:05:00 crc kubenswrapper[4775]: I1002 04:05:00.766793 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:05:00 crc kubenswrapper[4775]: E1002 04:05:00.768590 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:05:13 crc kubenswrapper[4775]: I1002 04:05:13.772060 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:05:13 crc kubenswrapper[4775]: E1002 04:05:13.772947 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:05:28 crc kubenswrapper[4775]: I1002 04:05:28.765349 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:05:28 crc kubenswrapper[4775]: E1002 04:05:28.766039 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:05:39 crc kubenswrapper[4775]: I1002 04:05:39.434598 4775 generic.go:334] "Generic (PLEG): container finished" podID="3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d" containerID="2998421915a71cc4f8da7636111ca6b08427d0b676ce3f25acbea621669c5c94" exitCode=0 Oct 02 04:05:39 crc kubenswrapper[4775]: I1002 04:05:39.434858 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-6vllg" event={"ID":"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d","Type":"ContainerDied","Data":"2998421915a71cc4f8da7636111ca6b08427d0b676ce3f25acbea621669c5c94"} Oct 02 04:05:40 crc kubenswrapper[4775]: I1002 04:05:40.767577 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:05:40 crc kubenswrapper[4775]: E1002 04:05:40.769260 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.012382 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.053895 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2btfq\" (UniqueName: \"kubernetes.io/projected/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-kube-api-access-2btfq\") pod \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.054106 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-telemetry-combined-ca-bundle\") pod \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.054257 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceph\") pod \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.054356 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-1\") pod \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.054420 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-0\") pod \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.054675 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-2\") pod \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.054744 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-inventory\") pod \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.054863 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ssh-key\") pod \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\" (UID: \"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d\") " Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.068678 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d" (UID: "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.068718 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-kube-api-access-2btfq" (OuterVolumeSpecName: "kube-api-access-2btfq") pod "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d" (UID: "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d"). InnerVolumeSpecName "kube-api-access-2btfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.068738 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceph" (OuterVolumeSpecName: "ceph") pod "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d" (UID: "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.089303 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d" (UID: "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.098307 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-inventory" (OuterVolumeSpecName: "inventory") pod "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d" (UID: "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.107176 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d" (UID: "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.110807 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d" (UID: "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.132768 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d" (UID: "3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.156861 4775 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.156894 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.156903 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.156912 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2btfq\" (UniqueName: \"kubernetes.io/projected/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-kube-api-access-2btfq\") on node \"crc\" DevicePath \"\"" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.156921 4775 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.156929 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.156938 4775 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.156946 4775 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.463928 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-6vllg" event={"ID":"3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d","Type":"ContainerDied","Data":"dcccb13c70ac80a55f31d47bad24e4fb60652d7e9176e58aec454ca461170712"} Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.464091 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcccb13c70ac80a55f31d47bad24e4fb60652d7e9176e58aec454ca461170712" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.464044 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-6vllg" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.609606 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-qjjxk"] Oct 02 04:05:41 crc kubenswrapper[4775]: E1002 04:05:41.610020 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d" containerName="telemetry-openstack-openstack-cell1" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.610037 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d" containerName="telemetry-openstack-openstack-cell1" Oct 02 04:05:41 crc kubenswrapper[4775]: E1002 04:05:41.610058 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" containerName="extract-utilities" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.610064 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" containerName="extract-utilities" Oct 02 04:05:41 crc kubenswrapper[4775]: E1002 04:05:41.610095 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" containerName="extract-content" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.610102 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" containerName="extract-content" Oct 02 04:05:41 crc kubenswrapper[4775]: E1002 04:05:41.610117 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" containerName="registry-server" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.610124 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" containerName="registry-server" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.610303 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="62a5dbfd-1260-4c70-b3b6-e0d33fa3971c" containerName="registry-server" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.610319 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d" containerName="telemetry-openstack-openstack-cell1" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.613623 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.616253 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.616687 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.616793 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.617000 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.621349 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.626657 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-qjjxk"] Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.667275 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjmvf\" (UniqueName: \"kubernetes.io/projected/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-kube-api-access-vjmvf\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.667830 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.668011 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.668150 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.668310 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.668888 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.770463 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjmvf\" (UniqueName: \"kubernetes.io/projected/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-kube-api-access-vjmvf\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.770508 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.770555 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.770588 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.770641 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.770688 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.774469 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.775934 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.779453 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.779538 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.780554 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.790729 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjmvf\" (UniqueName: \"kubernetes.io/projected/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-kube-api-access-vjmvf\") pod \"neutron-sriov-openstack-openstack-cell1-qjjxk\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:41 crc kubenswrapper[4775]: I1002 04:05:41.947428 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:05:42 crc kubenswrapper[4775]: I1002 04:05:42.585080 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-qjjxk"] Oct 02 04:05:42 crc kubenswrapper[4775]: W1002 04:05:42.589299 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4b83b696_fe3c_4696_a9ed_f30bbd0e1ba5.slice/crio-24fcf78877affe52e30d06f7577d857d0a6df44e88167ef78d8d772223f53c17 WatchSource:0}: Error finding container 24fcf78877affe52e30d06f7577d857d0a6df44e88167ef78d8d772223f53c17: Status 404 returned error can't find the container with id 24fcf78877affe52e30d06f7577d857d0a6df44e88167ef78d8d772223f53c17 Oct 02 04:05:43 crc kubenswrapper[4775]: I1002 04:05:43.498209 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" event={"ID":"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5","Type":"ContainerStarted","Data":"987029e0803eb7f009f925903be013cb3913b2e8b661854fab4518fb51d70359"} Oct 02 04:05:43 crc kubenswrapper[4775]: I1002 04:05:43.499091 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" event={"ID":"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5","Type":"ContainerStarted","Data":"24fcf78877affe52e30d06f7577d857d0a6df44e88167ef78d8d772223f53c17"} Oct 02 04:05:43 crc kubenswrapper[4775]: I1002 04:05:43.529634 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" podStartSLOduration=2.013135193 podStartE2EDuration="2.529606488s" podCreationTimestamp="2025-10-02 04:05:41 +0000 UTC" firstStartedPulling="2025-10-02 04:05:42.592160799 +0000 UTC m=+8679.758904879" lastFinishedPulling="2025-10-02 04:05:43.108632094 +0000 UTC m=+8680.275376174" observedRunningTime="2025-10-02 04:05:43.51928026 +0000 UTC m=+8680.686024380" watchObservedRunningTime="2025-10-02 04:05:43.529606488 +0000 UTC m=+8680.696350568" Oct 02 04:05:55 crc kubenswrapper[4775]: I1002 04:05:55.766496 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:05:55 crc kubenswrapper[4775]: E1002 04:05:55.768325 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:06:05 crc kubenswrapper[4775]: I1002 04:06:05.880021 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zxqzw"] Oct 02 04:06:05 crc kubenswrapper[4775]: I1002 04:06:05.883113 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:05 crc kubenswrapper[4775]: I1002 04:06:05.892441 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zxqzw"] Oct 02 04:06:05 crc kubenswrapper[4775]: I1002 04:06:05.984486 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a53c845-c385-4b9e-aa90-753b9ab0ddce-utilities\") pod \"redhat-operators-zxqzw\" (UID: \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\") " pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:05 crc kubenswrapper[4775]: I1002 04:06:05.984614 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a53c845-c385-4b9e-aa90-753b9ab0ddce-catalog-content\") pod \"redhat-operators-zxqzw\" (UID: \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\") " pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:05 crc kubenswrapper[4775]: I1002 04:06:05.984676 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcwpf\" (UniqueName: \"kubernetes.io/projected/2a53c845-c385-4b9e-aa90-753b9ab0ddce-kube-api-access-mcwpf\") pod \"redhat-operators-zxqzw\" (UID: \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\") " pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:06 crc kubenswrapper[4775]: I1002 04:06:06.087090 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a53c845-c385-4b9e-aa90-753b9ab0ddce-catalog-content\") pod \"redhat-operators-zxqzw\" (UID: \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\") " pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:06 crc kubenswrapper[4775]: I1002 04:06:06.087425 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcwpf\" (UniqueName: \"kubernetes.io/projected/2a53c845-c385-4b9e-aa90-753b9ab0ddce-kube-api-access-mcwpf\") pod \"redhat-operators-zxqzw\" (UID: \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\") " pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:06 crc kubenswrapper[4775]: I1002 04:06:06.087621 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a53c845-c385-4b9e-aa90-753b9ab0ddce-utilities\") pod \"redhat-operators-zxqzw\" (UID: \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\") " pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:06 crc kubenswrapper[4775]: I1002 04:06:06.088033 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a53c845-c385-4b9e-aa90-753b9ab0ddce-catalog-content\") pod \"redhat-operators-zxqzw\" (UID: \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\") " pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:06 crc kubenswrapper[4775]: I1002 04:06:06.088077 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a53c845-c385-4b9e-aa90-753b9ab0ddce-utilities\") pod \"redhat-operators-zxqzw\" (UID: \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\") " pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:06 crc kubenswrapper[4775]: I1002 04:06:06.117433 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcwpf\" (UniqueName: \"kubernetes.io/projected/2a53c845-c385-4b9e-aa90-753b9ab0ddce-kube-api-access-mcwpf\") pod \"redhat-operators-zxqzw\" (UID: \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\") " pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:06 crc kubenswrapper[4775]: I1002 04:06:06.215600 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:06 crc kubenswrapper[4775]: I1002 04:06:06.670938 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zxqzw"] Oct 02 04:06:06 crc kubenswrapper[4775]: W1002 04:06:06.688983 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a53c845_c385_4b9e_aa90_753b9ab0ddce.slice/crio-67e510ff68e4ab5f2875537dbd4e337ffb261122eb90345c9ae68508d2387b1d WatchSource:0}: Error finding container 67e510ff68e4ab5f2875537dbd4e337ffb261122eb90345c9ae68508d2387b1d: Status 404 returned error can't find the container with id 67e510ff68e4ab5f2875537dbd4e337ffb261122eb90345c9ae68508d2387b1d Oct 02 04:06:06 crc kubenswrapper[4775]: I1002 04:06:06.853176 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zxqzw" event={"ID":"2a53c845-c385-4b9e-aa90-753b9ab0ddce","Type":"ContainerStarted","Data":"67e510ff68e4ab5f2875537dbd4e337ffb261122eb90345c9ae68508d2387b1d"} Oct 02 04:06:07 crc kubenswrapper[4775]: I1002 04:06:07.766160 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:06:07 crc kubenswrapper[4775]: E1002 04:06:07.766876 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:06:07 crc kubenswrapper[4775]: I1002 04:06:07.865450 4775 generic.go:334] "Generic (PLEG): container finished" podID="2a53c845-c385-4b9e-aa90-753b9ab0ddce" containerID="5b5bbc47ac62b3d88d22b8d06743cd999c6bb1dc1965ff7e439bc1a6a782fe2d" exitCode=0 Oct 02 04:06:07 crc kubenswrapper[4775]: I1002 04:06:07.865514 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zxqzw" event={"ID":"2a53c845-c385-4b9e-aa90-753b9ab0ddce","Type":"ContainerDied","Data":"5b5bbc47ac62b3d88d22b8d06743cd999c6bb1dc1965ff7e439bc1a6a782fe2d"} Oct 02 04:06:09 crc kubenswrapper[4775]: I1002 04:06:09.891399 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zxqzw" event={"ID":"2a53c845-c385-4b9e-aa90-753b9ab0ddce","Type":"ContainerStarted","Data":"a5a5c08d95c29b5c5979c901c957d93b40cd368b8bf412cce3208defebd561ca"} Oct 02 04:06:11 crc kubenswrapper[4775]: I1002 04:06:11.918615 4775 generic.go:334] "Generic (PLEG): container finished" podID="2a53c845-c385-4b9e-aa90-753b9ab0ddce" containerID="a5a5c08d95c29b5c5979c901c957d93b40cd368b8bf412cce3208defebd561ca" exitCode=0 Oct 02 04:06:11 crc kubenswrapper[4775]: I1002 04:06:11.918752 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zxqzw" event={"ID":"2a53c845-c385-4b9e-aa90-753b9ab0ddce","Type":"ContainerDied","Data":"a5a5c08d95c29b5c5979c901c957d93b40cd368b8bf412cce3208defebd561ca"} Oct 02 04:06:12 crc kubenswrapper[4775]: I1002 04:06:12.939241 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zxqzw" event={"ID":"2a53c845-c385-4b9e-aa90-753b9ab0ddce","Type":"ContainerStarted","Data":"117653445ff163c62421ee9b61ee62ce3ee9104d996f001950c70143dfca9c2c"} Oct 02 04:06:12 crc kubenswrapper[4775]: I1002 04:06:12.977633 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zxqzw" podStartSLOduration=3.517809853 podStartE2EDuration="7.977605401s" podCreationTimestamp="2025-10-02 04:06:05 +0000 UTC" firstStartedPulling="2025-10-02 04:06:07.86820699 +0000 UTC m=+8705.034951070" lastFinishedPulling="2025-10-02 04:06:12.328002538 +0000 UTC m=+8709.494746618" observedRunningTime="2025-10-02 04:06:12.963562396 +0000 UTC m=+8710.130306466" watchObservedRunningTime="2025-10-02 04:06:12.977605401 +0000 UTC m=+8710.144349481" Oct 02 04:06:16 crc kubenswrapper[4775]: I1002 04:06:16.215761 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:16 crc kubenswrapper[4775]: I1002 04:06:16.216216 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:17 crc kubenswrapper[4775]: I1002 04:06:17.257501 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zxqzw" podUID="2a53c845-c385-4b9e-aa90-753b9ab0ddce" containerName="registry-server" probeResult="failure" output=< Oct 02 04:06:17 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 04:06:17 crc kubenswrapper[4775]: > Oct 02 04:06:22 crc kubenswrapper[4775]: I1002 04:06:22.765566 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:06:22 crc kubenswrapper[4775]: E1002 04:06:22.766378 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:06:26 crc kubenswrapper[4775]: I1002 04:06:26.277902 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:26 crc kubenswrapper[4775]: I1002 04:06:26.342998 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:26 crc kubenswrapper[4775]: I1002 04:06:26.530101 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zxqzw"] Oct 02 04:06:28 crc kubenswrapper[4775]: I1002 04:06:28.125370 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zxqzw" podUID="2a53c845-c385-4b9e-aa90-753b9ab0ddce" containerName="registry-server" containerID="cri-o://117653445ff163c62421ee9b61ee62ce3ee9104d996f001950c70143dfca9c2c" gracePeriod=2 Oct 02 04:06:28 crc kubenswrapper[4775]: I1002 04:06:28.675267 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:28 crc kubenswrapper[4775]: I1002 04:06:28.748906 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcwpf\" (UniqueName: \"kubernetes.io/projected/2a53c845-c385-4b9e-aa90-753b9ab0ddce-kube-api-access-mcwpf\") pod \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\" (UID: \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\") " Oct 02 04:06:28 crc kubenswrapper[4775]: I1002 04:06:28.748976 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a53c845-c385-4b9e-aa90-753b9ab0ddce-catalog-content\") pod \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\" (UID: \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\") " Oct 02 04:06:28 crc kubenswrapper[4775]: I1002 04:06:28.749145 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a53c845-c385-4b9e-aa90-753b9ab0ddce-utilities\") pod \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\" (UID: \"2a53c845-c385-4b9e-aa90-753b9ab0ddce\") " Oct 02 04:06:28 crc kubenswrapper[4775]: I1002 04:06:28.750116 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a53c845-c385-4b9e-aa90-753b9ab0ddce-utilities" (OuterVolumeSpecName: "utilities") pod "2a53c845-c385-4b9e-aa90-753b9ab0ddce" (UID: "2a53c845-c385-4b9e-aa90-753b9ab0ddce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:06:28 crc kubenswrapper[4775]: I1002 04:06:28.759988 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a53c845-c385-4b9e-aa90-753b9ab0ddce-kube-api-access-mcwpf" (OuterVolumeSpecName: "kube-api-access-mcwpf") pod "2a53c845-c385-4b9e-aa90-753b9ab0ddce" (UID: "2a53c845-c385-4b9e-aa90-753b9ab0ddce"). InnerVolumeSpecName "kube-api-access-mcwpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:06:28 crc kubenswrapper[4775]: I1002 04:06:28.854410 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcwpf\" (UniqueName: \"kubernetes.io/projected/2a53c845-c385-4b9e-aa90-753b9ab0ddce-kube-api-access-mcwpf\") on node \"crc\" DevicePath \"\"" Oct 02 04:06:28 crc kubenswrapper[4775]: I1002 04:06:28.854452 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a53c845-c385-4b9e-aa90-753b9ab0ddce-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 04:06:28 crc kubenswrapper[4775]: I1002 04:06:28.985547 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a53c845-c385-4b9e-aa90-753b9ab0ddce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a53c845-c385-4b9e-aa90-753b9ab0ddce" (UID: "2a53c845-c385-4b9e-aa90-753b9ab0ddce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.058251 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a53c845-c385-4b9e-aa90-753b9ab0ddce-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.139317 4775 generic.go:334] "Generic (PLEG): container finished" podID="2a53c845-c385-4b9e-aa90-753b9ab0ddce" containerID="117653445ff163c62421ee9b61ee62ce3ee9104d996f001950c70143dfca9c2c" exitCode=0 Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.139363 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zxqzw" event={"ID":"2a53c845-c385-4b9e-aa90-753b9ab0ddce","Type":"ContainerDied","Data":"117653445ff163c62421ee9b61ee62ce3ee9104d996f001950c70143dfca9c2c"} Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.139378 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zxqzw" Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.139394 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zxqzw" event={"ID":"2a53c845-c385-4b9e-aa90-753b9ab0ddce","Type":"ContainerDied","Data":"67e510ff68e4ab5f2875537dbd4e337ffb261122eb90345c9ae68508d2387b1d"} Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.139417 4775 scope.go:117] "RemoveContainer" containerID="117653445ff163c62421ee9b61ee62ce3ee9104d996f001950c70143dfca9c2c" Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.181846 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zxqzw"] Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.188562 4775 scope.go:117] "RemoveContainer" containerID="a5a5c08d95c29b5c5979c901c957d93b40cd368b8bf412cce3208defebd561ca" Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.190629 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zxqzw"] Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.211608 4775 scope.go:117] "RemoveContainer" containerID="5b5bbc47ac62b3d88d22b8d06743cd999c6bb1dc1965ff7e439bc1a6a782fe2d" Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.281573 4775 scope.go:117] "RemoveContainer" containerID="117653445ff163c62421ee9b61ee62ce3ee9104d996f001950c70143dfca9c2c" Oct 02 04:06:29 crc kubenswrapper[4775]: E1002 04:06:29.282684 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"117653445ff163c62421ee9b61ee62ce3ee9104d996f001950c70143dfca9c2c\": container with ID starting with 117653445ff163c62421ee9b61ee62ce3ee9104d996f001950c70143dfca9c2c not found: ID does not exist" containerID="117653445ff163c62421ee9b61ee62ce3ee9104d996f001950c70143dfca9c2c" Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.282760 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"117653445ff163c62421ee9b61ee62ce3ee9104d996f001950c70143dfca9c2c"} err="failed to get container status \"117653445ff163c62421ee9b61ee62ce3ee9104d996f001950c70143dfca9c2c\": rpc error: code = NotFound desc = could not find container \"117653445ff163c62421ee9b61ee62ce3ee9104d996f001950c70143dfca9c2c\": container with ID starting with 117653445ff163c62421ee9b61ee62ce3ee9104d996f001950c70143dfca9c2c not found: ID does not exist" Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.282802 4775 scope.go:117] "RemoveContainer" containerID="a5a5c08d95c29b5c5979c901c957d93b40cd368b8bf412cce3208defebd561ca" Oct 02 04:06:29 crc kubenswrapper[4775]: E1002 04:06:29.283205 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5a5c08d95c29b5c5979c901c957d93b40cd368b8bf412cce3208defebd561ca\": container with ID starting with a5a5c08d95c29b5c5979c901c957d93b40cd368b8bf412cce3208defebd561ca not found: ID does not exist" containerID="a5a5c08d95c29b5c5979c901c957d93b40cd368b8bf412cce3208defebd561ca" Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.283303 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5a5c08d95c29b5c5979c901c957d93b40cd368b8bf412cce3208defebd561ca"} err="failed to get container status \"a5a5c08d95c29b5c5979c901c957d93b40cd368b8bf412cce3208defebd561ca\": rpc error: code = NotFound desc = could not find container \"a5a5c08d95c29b5c5979c901c957d93b40cd368b8bf412cce3208defebd561ca\": container with ID starting with a5a5c08d95c29b5c5979c901c957d93b40cd368b8bf412cce3208defebd561ca not found: ID does not exist" Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.283369 4775 scope.go:117] "RemoveContainer" containerID="5b5bbc47ac62b3d88d22b8d06743cd999c6bb1dc1965ff7e439bc1a6a782fe2d" Oct 02 04:06:29 crc kubenswrapper[4775]: E1002 04:06:29.283755 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b5bbc47ac62b3d88d22b8d06743cd999c6bb1dc1965ff7e439bc1a6a782fe2d\": container with ID starting with 5b5bbc47ac62b3d88d22b8d06743cd999c6bb1dc1965ff7e439bc1a6a782fe2d not found: ID does not exist" containerID="5b5bbc47ac62b3d88d22b8d06743cd999c6bb1dc1965ff7e439bc1a6a782fe2d" Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.283812 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b5bbc47ac62b3d88d22b8d06743cd999c6bb1dc1965ff7e439bc1a6a782fe2d"} err="failed to get container status \"5b5bbc47ac62b3d88d22b8d06743cd999c6bb1dc1965ff7e439bc1a6a782fe2d\": rpc error: code = NotFound desc = could not find container \"5b5bbc47ac62b3d88d22b8d06743cd999c6bb1dc1965ff7e439bc1a6a782fe2d\": container with ID starting with 5b5bbc47ac62b3d88d22b8d06743cd999c6bb1dc1965ff7e439bc1a6a782fe2d not found: ID does not exist" Oct 02 04:06:29 crc kubenswrapper[4775]: I1002 04:06:29.777853 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a53c845-c385-4b9e-aa90-753b9ab0ddce" path="/var/lib/kubelet/pods/2a53c845-c385-4b9e-aa90-753b9ab0ddce/volumes" Oct 02 04:06:35 crc kubenswrapper[4775]: I1002 04:06:35.767053 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:06:35 crc kubenswrapper[4775]: E1002 04:06:35.768142 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:06:49 crc kubenswrapper[4775]: I1002 04:06:49.766149 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:06:50 crc kubenswrapper[4775]: I1002 04:06:50.423501 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"47e4f4cf08ebdd86638e66eae2851e8881ccefb4ffb92266f0c2f1604714456c"} Oct 02 04:08:01 crc kubenswrapper[4775]: I1002 04:08:01.397912 4775 generic.go:334] "Generic (PLEG): container finished" podID="4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5" containerID="987029e0803eb7f009f925903be013cb3913b2e8b661854fab4518fb51d70359" exitCode=0 Oct 02 04:08:01 crc kubenswrapper[4775]: I1002 04:08:01.397947 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" event={"ID":"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5","Type":"ContainerDied","Data":"987029e0803eb7f009f925903be013cb3913b2e8b661854fab4518fb51d70359"} Oct 02 04:08:02 crc kubenswrapper[4775]: I1002 04:08:02.957163 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.025069 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjmvf\" (UniqueName: \"kubernetes.io/projected/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-kube-api-access-vjmvf\") pod \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.025179 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-ssh-key\") pod \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.025215 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-ceph\") pod \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.025261 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-neutron-sriov-agent-neutron-config-0\") pod \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.025314 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-neutron-sriov-combined-ca-bundle\") pod \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.025406 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-inventory\") pod \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\" (UID: \"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5\") " Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.032997 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5" (UID: "4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.033713 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-ceph" (OuterVolumeSpecName: "ceph") pod "4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5" (UID: "4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.048427 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-kube-api-access-vjmvf" (OuterVolumeSpecName: "kube-api-access-vjmvf") pod "4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5" (UID: "4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5"). InnerVolumeSpecName "kube-api-access-vjmvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.063204 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5" (UID: "4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.072565 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-inventory" (OuterVolumeSpecName: "inventory") pod "4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5" (UID: "4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.085249 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5" (UID: "4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.168285 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjmvf\" (UniqueName: \"kubernetes.io/projected/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-kube-api-access-vjmvf\") on node \"crc\" DevicePath \"\"" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.168331 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.168349 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.168366 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.168386 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.168402 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.422922 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" event={"ID":"4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5","Type":"ContainerDied","Data":"24fcf78877affe52e30d06f7577d857d0a6df44e88167ef78d8d772223f53c17"} Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.422996 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="24fcf78877affe52e30d06f7577d857d0a6df44e88167ef78d8d772223f53c17" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.423098 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-qjjxk" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.575410 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-p97fw"] Oct 02 04:08:03 crc kubenswrapper[4775]: E1002 04:08:03.576040 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a53c845-c385-4b9e-aa90-753b9ab0ddce" containerName="extract-content" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.576061 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a53c845-c385-4b9e-aa90-753b9ab0ddce" containerName="extract-content" Oct 02 04:08:03 crc kubenswrapper[4775]: E1002 04:08:03.576079 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5" containerName="neutron-sriov-openstack-openstack-cell1" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.576099 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5" containerName="neutron-sriov-openstack-openstack-cell1" Oct 02 04:08:03 crc kubenswrapper[4775]: E1002 04:08:03.576148 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a53c845-c385-4b9e-aa90-753b9ab0ddce" containerName="registry-server" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.576159 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a53c845-c385-4b9e-aa90-753b9ab0ddce" containerName="registry-server" Oct 02 04:08:03 crc kubenswrapper[4775]: E1002 04:08:03.576176 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a53c845-c385-4b9e-aa90-753b9ab0ddce" containerName="extract-utilities" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.576183 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a53c845-c385-4b9e-aa90-753b9ab0ddce" containerName="extract-utilities" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.576436 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a53c845-c385-4b9e-aa90-753b9ab0ddce" containerName="registry-server" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.576471 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5" containerName="neutron-sriov-openstack-openstack-cell1" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.577922 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.580452 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.580815 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.580874 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.581065 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.582096 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.601128 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-p97fw"] Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.682570 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.682609 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.682686 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.682790 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.682890 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbtkp\" (UniqueName: \"kubernetes.io/projected/2258fa83-dbaf-4438-a9bf-42bb3458fa98-kube-api-access-lbtkp\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.682930 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.785236 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.785707 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.785897 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.786138 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.786418 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.786721 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbtkp\" (UniqueName: \"kubernetes.io/projected/2258fa83-dbaf-4438-a9bf-42bb3458fa98-kube-api-access-lbtkp\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.790473 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.791259 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.792846 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.794029 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.798219 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.818421 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbtkp\" (UniqueName: \"kubernetes.io/projected/2258fa83-dbaf-4438-a9bf-42bb3458fa98-kube-api-access-lbtkp\") pod \"neutron-dhcp-openstack-openstack-cell1-p97fw\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.904664 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 04:08:03 crc kubenswrapper[4775]: I1002 04:08:03.911763 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:08:04 crc kubenswrapper[4775]: I1002 04:08:04.510276 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-p97fw"] Oct 02 04:08:04 crc kubenswrapper[4775]: I1002 04:08:04.520397 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 04:08:04 crc kubenswrapper[4775]: I1002 04:08:04.992247 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 04:08:05 crc kubenswrapper[4775]: I1002 04:08:05.452670 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" event={"ID":"2258fa83-dbaf-4438-a9bf-42bb3458fa98","Type":"ContainerStarted","Data":"f5d9e838cb3e0221d66cb8c20abe17252be1b2e389c48111972990e077d65171"} Oct 02 04:08:05 crc kubenswrapper[4775]: I1002 04:08:05.453226 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" event={"ID":"2258fa83-dbaf-4438-a9bf-42bb3458fa98","Type":"ContainerStarted","Data":"55b4b6f1b6a9ffe8f5c2acbba0497853aa8d44699f06dd59fca548dac4c63743"} Oct 02 04:08:05 crc kubenswrapper[4775]: I1002 04:08:05.479820 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" podStartSLOduration=2.011005832 podStartE2EDuration="2.479791237s" podCreationTimestamp="2025-10-02 04:08:03 +0000 UTC" firstStartedPulling="2025-10-02 04:08:04.520106451 +0000 UTC m=+8821.686850491" lastFinishedPulling="2025-10-02 04:08:04.988891836 +0000 UTC m=+8822.155635896" observedRunningTime="2025-10-02 04:08:05.476270695 +0000 UTC m=+8822.643014745" watchObservedRunningTime="2025-10-02 04:08:05.479791237 +0000 UTC m=+8822.646535307" Oct 02 04:09:07 crc kubenswrapper[4775]: I1002 04:09:07.233453 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:09:07 crc kubenswrapper[4775]: I1002 04:09:07.234058 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:09:37 crc kubenswrapper[4775]: I1002 04:09:37.233849 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:09:37 crc kubenswrapper[4775]: I1002 04:09:37.236183 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:10:07 crc kubenswrapper[4775]: I1002 04:10:07.234380 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:10:07 crc kubenswrapper[4775]: I1002 04:10:07.235067 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:10:07 crc kubenswrapper[4775]: I1002 04:10:07.235126 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 04:10:07 crc kubenswrapper[4775]: I1002 04:10:07.237113 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"47e4f4cf08ebdd86638e66eae2851e8881ccefb4ffb92266f0c2f1604714456c"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 04:10:07 crc kubenswrapper[4775]: I1002 04:10:07.237205 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://47e4f4cf08ebdd86638e66eae2851e8881ccefb4ffb92266f0c2f1604714456c" gracePeriod=600 Oct 02 04:10:08 crc kubenswrapper[4775]: I1002 04:10:08.108295 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="47e4f4cf08ebdd86638e66eae2851e8881ccefb4ffb92266f0c2f1604714456c" exitCode=0 Oct 02 04:10:08 crc kubenswrapper[4775]: I1002 04:10:08.108362 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"47e4f4cf08ebdd86638e66eae2851e8881ccefb4ffb92266f0c2f1604714456c"} Oct 02 04:10:08 crc kubenswrapper[4775]: I1002 04:10:08.108632 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404"} Oct 02 04:10:08 crc kubenswrapper[4775]: I1002 04:10:08.108650 4775 scope.go:117] "RemoveContainer" containerID="c42d253c4f71b32931c72005f043186fb2bf58363a0710ad593bf034570fc676" Oct 02 04:12:07 crc kubenswrapper[4775]: I1002 04:12:07.233743 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:12:07 crc kubenswrapper[4775]: I1002 04:12:07.234481 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:12:37 crc kubenswrapper[4775]: I1002 04:12:37.234638 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:12:37 crc kubenswrapper[4775]: I1002 04:12:37.235420 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:13:07 crc kubenswrapper[4775]: I1002 04:13:07.233854 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:13:07 crc kubenswrapper[4775]: I1002 04:13:07.234655 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:13:07 crc kubenswrapper[4775]: I1002 04:13:07.234707 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 04:13:07 crc kubenswrapper[4775]: I1002 04:13:07.235596 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 04:13:07 crc kubenswrapper[4775]: I1002 04:13:07.235764 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" gracePeriod=600 Oct 02 04:13:07 crc kubenswrapper[4775]: E1002 04:13:07.373626 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:13:07 crc kubenswrapper[4775]: I1002 04:13:07.495444 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" exitCode=0 Oct 02 04:13:07 crc kubenswrapper[4775]: I1002 04:13:07.495489 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404"} Oct 02 04:13:07 crc kubenswrapper[4775]: I1002 04:13:07.495522 4775 scope.go:117] "RemoveContainer" containerID="47e4f4cf08ebdd86638e66eae2851e8881ccefb4ffb92266f0c2f1604714456c" Oct 02 04:13:07 crc kubenswrapper[4775]: I1002 04:13:07.496483 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:13:07 crc kubenswrapper[4775]: E1002 04:13:07.496914 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:13:21 crc kubenswrapper[4775]: I1002 04:13:21.765940 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:13:21 crc kubenswrapper[4775]: E1002 04:13:21.766987 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:13:32 crc kubenswrapper[4775]: I1002 04:13:32.765097 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:13:32 crc kubenswrapper[4775]: E1002 04:13:32.765927 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:13:43 crc kubenswrapper[4775]: I1002 04:13:43.787576 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:13:43 crc kubenswrapper[4775]: E1002 04:13:43.788894 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:13:48 crc kubenswrapper[4775]: I1002 04:13:48.819166 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lt9gr"] Oct 02 04:13:48 crc kubenswrapper[4775]: I1002 04:13:48.822867 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:13:48 crc kubenswrapper[4775]: I1002 04:13:48.833182 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lt9gr"] Oct 02 04:13:49 crc kubenswrapper[4775]: I1002 04:13:49.003754 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14fbedd3-bb6a-45b7-94c7-05cc35c52175-utilities\") pod \"redhat-marketplace-lt9gr\" (UID: \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\") " pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:13:49 crc kubenswrapper[4775]: I1002 04:13:49.004016 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mpbz\" (UniqueName: \"kubernetes.io/projected/14fbedd3-bb6a-45b7-94c7-05cc35c52175-kube-api-access-6mpbz\") pod \"redhat-marketplace-lt9gr\" (UID: \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\") " pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:13:49 crc kubenswrapper[4775]: I1002 04:13:49.004319 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14fbedd3-bb6a-45b7-94c7-05cc35c52175-catalog-content\") pod \"redhat-marketplace-lt9gr\" (UID: \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\") " pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:13:49 crc kubenswrapper[4775]: I1002 04:13:49.107004 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14fbedd3-bb6a-45b7-94c7-05cc35c52175-utilities\") pod \"redhat-marketplace-lt9gr\" (UID: \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\") " pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:13:49 crc kubenswrapper[4775]: I1002 04:13:49.107169 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mpbz\" (UniqueName: \"kubernetes.io/projected/14fbedd3-bb6a-45b7-94c7-05cc35c52175-kube-api-access-6mpbz\") pod \"redhat-marketplace-lt9gr\" (UID: \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\") " pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:13:49 crc kubenswrapper[4775]: I1002 04:13:49.107286 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14fbedd3-bb6a-45b7-94c7-05cc35c52175-catalog-content\") pod \"redhat-marketplace-lt9gr\" (UID: \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\") " pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:13:49 crc kubenswrapper[4775]: I1002 04:13:49.107584 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14fbedd3-bb6a-45b7-94c7-05cc35c52175-utilities\") pod \"redhat-marketplace-lt9gr\" (UID: \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\") " pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:13:49 crc kubenswrapper[4775]: I1002 04:13:49.107739 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14fbedd3-bb6a-45b7-94c7-05cc35c52175-catalog-content\") pod \"redhat-marketplace-lt9gr\" (UID: \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\") " pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:13:49 crc kubenswrapper[4775]: I1002 04:13:49.125993 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mpbz\" (UniqueName: \"kubernetes.io/projected/14fbedd3-bb6a-45b7-94c7-05cc35c52175-kube-api-access-6mpbz\") pod \"redhat-marketplace-lt9gr\" (UID: \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\") " pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:13:49 crc kubenswrapper[4775]: I1002 04:13:49.149595 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:13:49 crc kubenswrapper[4775]: I1002 04:13:49.646515 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lt9gr"] Oct 02 04:13:49 crc kubenswrapper[4775]: W1002 04:13:49.669508 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod14fbedd3_bb6a_45b7_94c7_05cc35c52175.slice/crio-b87393ceda47c60651ca7369926cfe7e18922d98bd245268f03940e93be25a8a WatchSource:0}: Error finding container b87393ceda47c60651ca7369926cfe7e18922d98bd245268f03940e93be25a8a: Status 404 returned error can't find the container with id b87393ceda47c60651ca7369926cfe7e18922d98bd245268f03940e93be25a8a Oct 02 04:13:50 crc kubenswrapper[4775]: I1002 04:13:50.075797 4775 generic.go:334] "Generic (PLEG): container finished" podID="14fbedd3-bb6a-45b7-94c7-05cc35c52175" containerID="0eaaf9a2cfa98bebb5b59ce6ae84ed32502c67a441ed77e783f60023f8b44b9f" exitCode=0 Oct 02 04:13:50 crc kubenswrapper[4775]: I1002 04:13:50.075901 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lt9gr" event={"ID":"14fbedd3-bb6a-45b7-94c7-05cc35c52175","Type":"ContainerDied","Data":"0eaaf9a2cfa98bebb5b59ce6ae84ed32502c67a441ed77e783f60023f8b44b9f"} Oct 02 04:13:50 crc kubenswrapper[4775]: I1002 04:13:50.076071 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lt9gr" event={"ID":"14fbedd3-bb6a-45b7-94c7-05cc35c52175","Type":"ContainerStarted","Data":"b87393ceda47c60651ca7369926cfe7e18922d98bd245268f03940e93be25a8a"} Oct 02 04:13:50 crc kubenswrapper[4775]: I1002 04:13:50.078801 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 04:13:51 crc kubenswrapper[4775]: I1002 04:13:51.097023 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lt9gr" event={"ID":"14fbedd3-bb6a-45b7-94c7-05cc35c52175","Type":"ContainerStarted","Data":"32d7e32dd17baee267f768f9ce73330ceeeea0a8fee8ad10704f686109126c77"} Oct 02 04:13:52 crc kubenswrapper[4775]: I1002 04:13:52.115708 4775 generic.go:334] "Generic (PLEG): container finished" podID="14fbedd3-bb6a-45b7-94c7-05cc35c52175" containerID="32d7e32dd17baee267f768f9ce73330ceeeea0a8fee8ad10704f686109126c77" exitCode=0 Oct 02 04:13:52 crc kubenswrapper[4775]: I1002 04:13:52.115880 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lt9gr" event={"ID":"14fbedd3-bb6a-45b7-94c7-05cc35c52175","Type":"ContainerDied","Data":"32d7e32dd17baee267f768f9ce73330ceeeea0a8fee8ad10704f686109126c77"} Oct 02 04:13:53 crc kubenswrapper[4775]: I1002 04:13:53.133327 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lt9gr" event={"ID":"14fbedd3-bb6a-45b7-94c7-05cc35c52175","Type":"ContainerStarted","Data":"b7f81a10384fc76773864ee84ca5094c9a51b6bf7d8ee0814f8deb5a3fa1fdac"} Oct 02 04:13:53 crc kubenswrapper[4775]: I1002 04:13:53.174804 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lt9gr" podStartSLOduration=2.594663724 podStartE2EDuration="5.174779198s" podCreationTimestamp="2025-10-02 04:13:48 +0000 UTC" firstStartedPulling="2025-10-02 04:13:50.078573649 +0000 UTC m=+9167.245317689" lastFinishedPulling="2025-10-02 04:13:52.658689093 +0000 UTC m=+9169.825433163" observedRunningTime="2025-10-02 04:13:53.15752159 +0000 UTC m=+9170.324265630" watchObservedRunningTime="2025-10-02 04:13:53.174779198 +0000 UTC m=+9170.341523248" Oct 02 04:13:58 crc kubenswrapper[4775]: I1002 04:13:58.766834 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:13:58 crc kubenswrapper[4775]: E1002 04:13:58.767992 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:13:59 crc kubenswrapper[4775]: I1002 04:13:59.150576 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:13:59 crc kubenswrapper[4775]: I1002 04:13:59.150884 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:13:59 crc kubenswrapper[4775]: I1002 04:13:59.646432 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:14:00 crc kubenswrapper[4775]: I1002 04:14:00.321923 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:14:00 crc kubenswrapper[4775]: I1002 04:14:00.399848 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lt9gr"] Oct 02 04:14:02 crc kubenswrapper[4775]: I1002 04:14:02.253355 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lt9gr" podUID="14fbedd3-bb6a-45b7-94c7-05cc35c52175" containerName="registry-server" containerID="cri-o://b7f81a10384fc76773864ee84ca5094c9a51b6bf7d8ee0814f8deb5a3fa1fdac" gracePeriod=2 Oct 02 04:14:02 crc kubenswrapper[4775]: I1002 04:14:02.869093 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:14:02 crc kubenswrapper[4775]: I1002 04:14:02.955215 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14fbedd3-bb6a-45b7-94c7-05cc35c52175-utilities\") pod \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\" (UID: \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\") " Oct 02 04:14:02 crc kubenswrapper[4775]: I1002 04:14:02.955320 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mpbz\" (UniqueName: \"kubernetes.io/projected/14fbedd3-bb6a-45b7-94c7-05cc35c52175-kube-api-access-6mpbz\") pod \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\" (UID: \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\") " Oct 02 04:14:02 crc kubenswrapper[4775]: I1002 04:14:02.955367 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14fbedd3-bb6a-45b7-94c7-05cc35c52175-catalog-content\") pod \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\" (UID: \"14fbedd3-bb6a-45b7-94c7-05cc35c52175\") " Oct 02 04:14:02 crc kubenswrapper[4775]: I1002 04:14:02.957084 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14fbedd3-bb6a-45b7-94c7-05cc35c52175-utilities" (OuterVolumeSpecName: "utilities") pod "14fbedd3-bb6a-45b7-94c7-05cc35c52175" (UID: "14fbedd3-bb6a-45b7-94c7-05cc35c52175"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:14:02 crc kubenswrapper[4775]: I1002 04:14:02.963415 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14fbedd3-bb6a-45b7-94c7-05cc35c52175-kube-api-access-6mpbz" (OuterVolumeSpecName: "kube-api-access-6mpbz") pod "14fbedd3-bb6a-45b7-94c7-05cc35c52175" (UID: "14fbedd3-bb6a-45b7-94c7-05cc35c52175"). InnerVolumeSpecName "kube-api-access-6mpbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:14:02 crc kubenswrapper[4775]: I1002 04:14:02.970364 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14fbedd3-bb6a-45b7-94c7-05cc35c52175-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14fbedd3-bb6a-45b7-94c7-05cc35c52175" (UID: "14fbedd3-bb6a-45b7-94c7-05cc35c52175"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.057821 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14fbedd3-bb6a-45b7-94c7-05cc35c52175-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.057888 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mpbz\" (UniqueName: \"kubernetes.io/projected/14fbedd3-bb6a-45b7-94c7-05cc35c52175-kube-api-access-6mpbz\") on node \"crc\" DevicePath \"\"" Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.057909 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14fbedd3-bb6a-45b7-94c7-05cc35c52175-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.285312 4775 generic.go:334] "Generic (PLEG): container finished" podID="14fbedd3-bb6a-45b7-94c7-05cc35c52175" containerID="b7f81a10384fc76773864ee84ca5094c9a51b6bf7d8ee0814f8deb5a3fa1fdac" exitCode=0 Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.285358 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lt9gr" event={"ID":"14fbedd3-bb6a-45b7-94c7-05cc35c52175","Type":"ContainerDied","Data":"b7f81a10384fc76773864ee84ca5094c9a51b6bf7d8ee0814f8deb5a3fa1fdac"} Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.285370 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lt9gr" Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.285385 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lt9gr" event={"ID":"14fbedd3-bb6a-45b7-94c7-05cc35c52175","Type":"ContainerDied","Data":"b87393ceda47c60651ca7369926cfe7e18922d98bd245268f03940e93be25a8a"} Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.285429 4775 scope.go:117] "RemoveContainer" containerID="b7f81a10384fc76773864ee84ca5094c9a51b6bf7d8ee0814f8deb5a3fa1fdac" Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.328041 4775 scope.go:117] "RemoveContainer" containerID="32d7e32dd17baee267f768f9ce73330ceeeea0a8fee8ad10704f686109126c77" Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.343163 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lt9gr"] Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.357941 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lt9gr"] Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.361122 4775 scope.go:117] "RemoveContainer" containerID="0eaaf9a2cfa98bebb5b59ce6ae84ed32502c67a441ed77e783f60023f8b44b9f" Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.428772 4775 scope.go:117] "RemoveContainer" containerID="b7f81a10384fc76773864ee84ca5094c9a51b6bf7d8ee0814f8deb5a3fa1fdac" Oct 02 04:14:03 crc kubenswrapper[4775]: E1002 04:14:03.429734 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7f81a10384fc76773864ee84ca5094c9a51b6bf7d8ee0814f8deb5a3fa1fdac\": container with ID starting with b7f81a10384fc76773864ee84ca5094c9a51b6bf7d8ee0814f8deb5a3fa1fdac not found: ID does not exist" containerID="b7f81a10384fc76773864ee84ca5094c9a51b6bf7d8ee0814f8deb5a3fa1fdac" Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.429811 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f81a10384fc76773864ee84ca5094c9a51b6bf7d8ee0814f8deb5a3fa1fdac"} err="failed to get container status \"b7f81a10384fc76773864ee84ca5094c9a51b6bf7d8ee0814f8deb5a3fa1fdac\": rpc error: code = NotFound desc = could not find container \"b7f81a10384fc76773864ee84ca5094c9a51b6bf7d8ee0814f8deb5a3fa1fdac\": container with ID starting with b7f81a10384fc76773864ee84ca5094c9a51b6bf7d8ee0814f8deb5a3fa1fdac not found: ID does not exist" Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.429854 4775 scope.go:117] "RemoveContainer" containerID="32d7e32dd17baee267f768f9ce73330ceeeea0a8fee8ad10704f686109126c77" Oct 02 04:14:03 crc kubenswrapper[4775]: E1002 04:14:03.430371 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32d7e32dd17baee267f768f9ce73330ceeeea0a8fee8ad10704f686109126c77\": container with ID starting with 32d7e32dd17baee267f768f9ce73330ceeeea0a8fee8ad10704f686109126c77 not found: ID does not exist" containerID="32d7e32dd17baee267f768f9ce73330ceeeea0a8fee8ad10704f686109126c77" Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.430420 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32d7e32dd17baee267f768f9ce73330ceeeea0a8fee8ad10704f686109126c77"} err="failed to get container status \"32d7e32dd17baee267f768f9ce73330ceeeea0a8fee8ad10704f686109126c77\": rpc error: code = NotFound desc = could not find container \"32d7e32dd17baee267f768f9ce73330ceeeea0a8fee8ad10704f686109126c77\": container with ID starting with 32d7e32dd17baee267f768f9ce73330ceeeea0a8fee8ad10704f686109126c77 not found: ID does not exist" Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.430451 4775 scope.go:117] "RemoveContainer" containerID="0eaaf9a2cfa98bebb5b59ce6ae84ed32502c67a441ed77e783f60023f8b44b9f" Oct 02 04:14:03 crc kubenswrapper[4775]: E1002 04:14:03.430934 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eaaf9a2cfa98bebb5b59ce6ae84ed32502c67a441ed77e783f60023f8b44b9f\": container with ID starting with 0eaaf9a2cfa98bebb5b59ce6ae84ed32502c67a441ed77e783f60023f8b44b9f not found: ID does not exist" containerID="0eaaf9a2cfa98bebb5b59ce6ae84ed32502c67a441ed77e783f60023f8b44b9f" Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.431046 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eaaf9a2cfa98bebb5b59ce6ae84ed32502c67a441ed77e783f60023f8b44b9f"} err="failed to get container status \"0eaaf9a2cfa98bebb5b59ce6ae84ed32502c67a441ed77e783f60023f8b44b9f\": rpc error: code = NotFound desc = could not find container \"0eaaf9a2cfa98bebb5b59ce6ae84ed32502c67a441ed77e783f60023f8b44b9f\": container with ID starting with 0eaaf9a2cfa98bebb5b59ce6ae84ed32502c67a441ed77e783f60023f8b44b9f not found: ID does not exist" Oct 02 04:14:03 crc kubenswrapper[4775]: I1002 04:14:03.801064 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14fbedd3-bb6a-45b7-94c7-05cc35c52175" path="/var/lib/kubelet/pods/14fbedd3-bb6a-45b7-94c7-05cc35c52175/volumes" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.322577 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6kjv2"] Oct 02 04:14:05 crc kubenswrapper[4775]: E1002 04:14:05.323223 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14fbedd3-bb6a-45b7-94c7-05cc35c52175" containerName="extract-utilities" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.323235 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="14fbedd3-bb6a-45b7-94c7-05cc35c52175" containerName="extract-utilities" Oct 02 04:14:05 crc kubenswrapper[4775]: E1002 04:14:05.323264 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14fbedd3-bb6a-45b7-94c7-05cc35c52175" containerName="extract-content" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.323270 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="14fbedd3-bb6a-45b7-94c7-05cc35c52175" containerName="extract-content" Oct 02 04:14:05 crc kubenswrapper[4775]: E1002 04:14:05.323288 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14fbedd3-bb6a-45b7-94c7-05cc35c52175" containerName="registry-server" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.323296 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="14fbedd3-bb6a-45b7-94c7-05cc35c52175" containerName="registry-server" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.323498 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="14fbedd3-bb6a-45b7-94c7-05cc35c52175" containerName="registry-server" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.325151 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.333617 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6kjv2"] Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.436785 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j65mf\" (UniqueName: \"kubernetes.io/projected/98ffc37b-a80a-4cda-be63-2269212fe894-kube-api-access-j65mf\") pod \"community-operators-6kjv2\" (UID: \"98ffc37b-a80a-4cda-be63-2269212fe894\") " pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.437133 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ffc37b-a80a-4cda-be63-2269212fe894-utilities\") pod \"community-operators-6kjv2\" (UID: \"98ffc37b-a80a-4cda-be63-2269212fe894\") " pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.437309 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ffc37b-a80a-4cda-be63-2269212fe894-catalog-content\") pod \"community-operators-6kjv2\" (UID: \"98ffc37b-a80a-4cda-be63-2269212fe894\") " pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.539572 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j65mf\" (UniqueName: \"kubernetes.io/projected/98ffc37b-a80a-4cda-be63-2269212fe894-kube-api-access-j65mf\") pod \"community-operators-6kjv2\" (UID: \"98ffc37b-a80a-4cda-be63-2269212fe894\") " pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.539783 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ffc37b-a80a-4cda-be63-2269212fe894-utilities\") pod \"community-operators-6kjv2\" (UID: \"98ffc37b-a80a-4cda-be63-2269212fe894\") " pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.539868 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ffc37b-a80a-4cda-be63-2269212fe894-catalog-content\") pod \"community-operators-6kjv2\" (UID: \"98ffc37b-a80a-4cda-be63-2269212fe894\") " pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.540431 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ffc37b-a80a-4cda-be63-2269212fe894-utilities\") pod \"community-operators-6kjv2\" (UID: \"98ffc37b-a80a-4cda-be63-2269212fe894\") " pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.540554 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ffc37b-a80a-4cda-be63-2269212fe894-catalog-content\") pod \"community-operators-6kjv2\" (UID: \"98ffc37b-a80a-4cda-be63-2269212fe894\") " pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.563415 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j65mf\" (UniqueName: \"kubernetes.io/projected/98ffc37b-a80a-4cda-be63-2269212fe894-kube-api-access-j65mf\") pod \"community-operators-6kjv2\" (UID: \"98ffc37b-a80a-4cda-be63-2269212fe894\") " pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:05 crc kubenswrapper[4775]: I1002 04:14:05.652512 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:06 crc kubenswrapper[4775]: I1002 04:14:06.190681 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6kjv2"] Oct 02 04:14:06 crc kubenswrapper[4775]: I1002 04:14:06.331793 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kjv2" event={"ID":"98ffc37b-a80a-4cda-be63-2269212fe894","Type":"ContainerStarted","Data":"31af6efc9ad861d446a34003efc39c7aadf730880944db0382f8729eb27b6c79"} Oct 02 04:14:07 crc kubenswrapper[4775]: I1002 04:14:07.349748 4775 generic.go:334] "Generic (PLEG): container finished" podID="98ffc37b-a80a-4cda-be63-2269212fe894" containerID="af3f0cc1c00faef40362c239a8426e379fbb97ebef1fc547819a93b8cc84d06b" exitCode=0 Oct 02 04:14:07 crc kubenswrapper[4775]: I1002 04:14:07.349820 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kjv2" event={"ID":"98ffc37b-a80a-4cda-be63-2269212fe894","Type":"ContainerDied","Data":"af3f0cc1c00faef40362c239a8426e379fbb97ebef1fc547819a93b8cc84d06b"} Oct 02 04:14:09 crc kubenswrapper[4775]: I1002 04:14:09.382548 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kjv2" event={"ID":"98ffc37b-a80a-4cda-be63-2269212fe894","Type":"ContainerStarted","Data":"76850815e6533ad94e60fb9760f873e1f1004ce662b88b81b441ae0da91fbdca"} Oct 02 04:14:10 crc kubenswrapper[4775]: I1002 04:14:10.397894 4775 generic.go:334] "Generic (PLEG): container finished" podID="98ffc37b-a80a-4cda-be63-2269212fe894" containerID="76850815e6533ad94e60fb9760f873e1f1004ce662b88b81b441ae0da91fbdca" exitCode=0 Oct 02 04:14:10 crc kubenswrapper[4775]: I1002 04:14:10.397981 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kjv2" event={"ID":"98ffc37b-a80a-4cda-be63-2269212fe894","Type":"ContainerDied","Data":"76850815e6533ad94e60fb9760f873e1f1004ce662b88b81b441ae0da91fbdca"} Oct 02 04:14:11 crc kubenswrapper[4775]: I1002 04:14:11.765652 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:14:11 crc kubenswrapper[4775]: E1002 04:14:11.766716 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:14:12 crc kubenswrapper[4775]: I1002 04:14:12.425268 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kjv2" event={"ID":"98ffc37b-a80a-4cda-be63-2269212fe894","Type":"ContainerStarted","Data":"a4389f4558a7c023729563a3bd753bf0a09278cc3093f5ccb95562b3bbf90172"} Oct 02 04:14:12 crc kubenswrapper[4775]: I1002 04:14:12.454606 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6kjv2" podStartSLOduration=3.56424226 podStartE2EDuration="7.454588387s" podCreationTimestamp="2025-10-02 04:14:05 +0000 UTC" firstStartedPulling="2025-10-02 04:14:07.353393359 +0000 UTC m=+9184.520137439" lastFinishedPulling="2025-10-02 04:14:11.243739526 +0000 UTC m=+9188.410483566" observedRunningTime="2025-10-02 04:14:12.449549276 +0000 UTC m=+9189.616293326" watchObservedRunningTime="2025-10-02 04:14:12.454588387 +0000 UTC m=+9189.621332427" Oct 02 04:14:15 crc kubenswrapper[4775]: I1002 04:14:15.653281 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:15 crc kubenswrapper[4775]: I1002 04:14:15.653885 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:15 crc kubenswrapper[4775]: I1002 04:14:15.740006 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:25 crc kubenswrapper[4775]: I1002 04:14:25.741071 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:25 crc kubenswrapper[4775]: I1002 04:14:25.817238 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6kjv2"] Oct 02 04:14:26 crc kubenswrapper[4775]: I1002 04:14:26.626447 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6kjv2" podUID="98ffc37b-a80a-4cda-be63-2269212fe894" containerName="registry-server" containerID="cri-o://a4389f4558a7c023729563a3bd753bf0a09278cc3093f5ccb95562b3bbf90172" gracePeriod=2 Oct 02 04:14:26 crc kubenswrapper[4775]: I1002 04:14:26.765070 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:14:26 crc kubenswrapper[4775]: E1002 04:14:26.765488 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.624922 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.638452 4775 generic.go:334] "Generic (PLEG): container finished" podID="98ffc37b-a80a-4cda-be63-2269212fe894" containerID="a4389f4558a7c023729563a3bd753bf0a09278cc3093f5ccb95562b3bbf90172" exitCode=0 Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.638550 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kjv2" event={"ID":"98ffc37b-a80a-4cda-be63-2269212fe894","Type":"ContainerDied","Data":"a4389f4558a7c023729563a3bd753bf0a09278cc3093f5ccb95562b3bbf90172"} Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.638576 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6kjv2" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.638601 4775 scope.go:117] "RemoveContainer" containerID="a4389f4558a7c023729563a3bd753bf0a09278cc3093f5ccb95562b3bbf90172" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.638589 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6kjv2" event={"ID":"98ffc37b-a80a-4cda-be63-2269212fe894","Type":"ContainerDied","Data":"31af6efc9ad861d446a34003efc39c7aadf730880944db0382f8729eb27b6c79"} Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.681215 4775 scope.go:117] "RemoveContainer" containerID="76850815e6533ad94e60fb9760f873e1f1004ce662b88b81b441ae0da91fbdca" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.712946 4775 scope.go:117] "RemoveContainer" containerID="af3f0cc1c00faef40362c239a8426e379fbb97ebef1fc547819a93b8cc84d06b" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.755133 4775 scope.go:117] "RemoveContainer" containerID="a4389f4558a7c023729563a3bd753bf0a09278cc3093f5ccb95562b3bbf90172" Oct 02 04:14:27 crc kubenswrapper[4775]: E1002 04:14:27.755694 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4389f4558a7c023729563a3bd753bf0a09278cc3093f5ccb95562b3bbf90172\": container with ID starting with a4389f4558a7c023729563a3bd753bf0a09278cc3093f5ccb95562b3bbf90172 not found: ID does not exist" containerID="a4389f4558a7c023729563a3bd753bf0a09278cc3093f5ccb95562b3bbf90172" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.755799 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4389f4558a7c023729563a3bd753bf0a09278cc3093f5ccb95562b3bbf90172"} err="failed to get container status \"a4389f4558a7c023729563a3bd753bf0a09278cc3093f5ccb95562b3bbf90172\": rpc error: code = NotFound desc = could not find container \"a4389f4558a7c023729563a3bd753bf0a09278cc3093f5ccb95562b3bbf90172\": container with ID starting with a4389f4558a7c023729563a3bd753bf0a09278cc3093f5ccb95562b3bbf90172 not found: ID does not exist" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.755869 4775 scope.go:117] "RemoveContainer" containerID="76850815e6533ad94e60fb9760f873e1f1004ce662b88b81b441ae0da91fbdca" Oct 02 04:14:27 crc kubenswrapper[4775]: E1002 04:14:27.756421 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76850815e6533ad94e60fb9760f873e1f1004ce662b88b81b441ae0da91fbdca\": container with ID starting with 76850815e6533ad94e60fb9760f873e1f1004ce662b88b81b441ae0da91fbdca not found: ID does not exist" containerID="76850815e6533ad94e60fb9760f873e1f1004ce662b88b81b441ae0da91fbdca" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.756456 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76850815e6533ad94e60fb9760f873e1f1004ce662b88b81b441ae0da91fbdca"} err="failed to get container status \"76850815e6533ad94e60fb9760f873e1f1004ce662b88b81b441ae0da91fbdca\": rpc error: code = NotFound desc = could not find container \"76850815e6533ad94e60fb9760f873e1f1004ce662b88b81b441ae0da91fbdca\": container with ID starting with 76850815e6533ad94e60fb9760f873e1f1004ce662b88b81b441ae0da91fbdca not found: ID does not exist" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.756483 4775 scope.go:117] "RemoveContainer" containerID="af3f0cc1c00faef40362c239a8426e379fbb97ebef1fc547819a93b8cc84d06b" Oct 02 04:14:27 crc kubenswrapper[4775]: E1002 04:14:27.757272 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af3f0cc1c00faef40362c239a8426e379fbb97ebef1fc547819a93b8cc84d06b\": container with ID starting with af3f0cc1c00faef40362c239a8426e379fbb97ebef1fc547819a93b8cc84d06b not found: ID does not exist" containerID="af3f0cc1c00faef40362c239a8426e379fbb97ebef1fc547819a93b8cc84d06b" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.757304 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af3f0cc1c00faef40362c239a8426e379fbb97ebef1fc547819a93b8cc84d06b"} err="failed to get container status \"af3f0cc1c00faef40362c239a8426e379fbb97ebef1fc547819a93b8cc84d06b\": rpc error: code = NotFound desc = could not find container \"af3f0cc1c00faef40362c239a8426e379fbb97ebef1fc547819a93b8cc84d06b\": container with ID starting with af3f0cc1c00faef40362c239a8426e379fbb97ebef1fc547819a93b8cc84d06b not found: ID does not exist" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.781872 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j65mf\" (UniqueName: \"kubernetes.io/projected/98ffc37b-a80a-4cda-be63-2269212fe894-kube-api-access-j65mf\") pod \"98ffc37b-a80a-4cda-be63-2269212fe894\" (UID: \"98ffc37b-a80a-4cda-be63-2269212fe894\") " Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.782146 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ffc37b-a80a-4cda-be63-2269212fe894-utilities\") pod \"98ffc37b-a80a-4cda-be63-2269212fe894\" (UID: \"98ffc37b-a80a-4cda-be63-2269212fe894\") " Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.782256 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ffc37b-a80a-4cda-be63-2269212fe894-catalog-content\") pod \"98ffc37b-a80a-4cda-be63-2269212fe894\" (UID: \"98ffc37b-a80a-4cda-be63-2269212fe894\") " Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.784127 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98ffc37b-a80a-4cda-be63-2269212fe894-utilities" (OuterVolumeSpecName: "utilities") pod "98ffc37b-a80a-4cda-be63-2269212fe894" (UID: "98ffc37b-a80a-4cda-be63-2269212fe894"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.789207 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98ffc37b-a80a-4cda-be63-2269212fe894-kube-api-access-j65mf" (OuterVolumeSpecName: "kube-api-access-j65mf") pod "98ffc37b-a80a-4cda-be63-2269212fe894" (UID: "98ffc37b-a80a-4cda-be63-2269212fe894"). InnerVolumeSpecName "kube-api-access-j65mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.837764 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98ffc37b-a80a-4cda-be63-2269212fe894-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98ffc37b-a80a-4cda-be63-2269212fe894" (UID: "98ffc37b-a80a-4cda-be63-2269212fe894"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.885552 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98ffc37b-a80a-4cda-be63-2269212fe894-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.885585 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j65mf\" (UniqueName: \"kubernetes.io/projected/98ffc37b-a80a-4cda-be63-2269212fe894-kube-api-access-j65mf\") on node \"crc\" DevicePath \"\"" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.885596 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98ffc37b-a80a-4cda-be63-2269212fe894-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 04:14:27 crc kubenswrapper[4775]: I1002 04:14:27.980788 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6kjv2"] Oct 02 04:14:28 crc kubenswrapper[4775]: I1002 04:14:28.022400 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6kjv2"] Oct 02 04:14:29 crc kubenswrapper[4775]: I1002 04:14:29.785886 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98ffc37b-a80a-4cda-be63-2269212fe894" path="/var/lib/kubelet/pods/98ffc37b-a80a-4cda-be63-2269212fe894/volumes" Oct 02 04:14:40 crc kubenswrapper[4775]: I1002 04:14:40.765160 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:14:40 crc kubenswrapper[4775]: E1002 04:14:40.765990 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:14:48 crc kubenswrapper[4775]: I1002 04:14:48.975376 4775 generic.go:334] "Generic (PLEG): container finished" podID="2258fa83-dbaf-4438-a9bf-42bb3458fa98" containerID="f5d9e838cb3e0221d66cb8c20abe17252be1b2e389c48111972990e077d65171" exitCode=0 Oct 02 04:14:48 crc kubenswrapper[4775]: I1002 04:14:48.975510 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" event={"ID":"2258fa83-dbaf-4438-a9bf-42bb3458fa98","Type":"ContainerDied","Data":"f5d9e838cb3e0221d66cb8c20abe17252be1b2e389c48111972990e077d65171"} Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.530221 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.553139 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-neutron-dhcp-combined-ca-bundle\") pod \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.553305 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-ceph\") pod \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.553363 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-ssh-key\") pod \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.553503 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-neutron-dhcp-agent-neutron-config-0\") pod \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.553594 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbtkp\" (UniqueName: \"kubernetes.io/projected/2258fa83-dbaf-4438-a9bf-42bb3458fa98-kube-api-access-lbtkp\") pod \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.553651 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-inventory\") pod \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\" (UID: \"2258fa83-dbaf-4438-a9bf-42bb3458fa98\") " Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.560265 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-ceph" (OuterVolumeSpecName: "ceph") pod "2258fa83-dbaf-4438-a9bf-42bb3458fa98" (UID: "2258fa83-dbaf-4438-a9bf-42bb3458fa98"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.560497 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2258fa83-dbaf-4438-a9bf-42bb3458fa98-kube-api-access-lbtkp" (OuterVolumeSpecName: "kube-api-access-lbtkp") pod "2258fa83-dbaf-4438-a9bf-42bb3458fa98" (UID: "2258fa83-dbaf-4438-a9bf-42bb3458fa98"). InnerVolumeSpecName "kube-api-access-lbtkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.565467 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "2258fa83-dbaf-4438-a9bf-42bb3458fa98" (UID: "2258fa83-dbaf-4438-a9bf-42bb3458fa98"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.586535 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2258fa83-dbaf-4438-a9bf-42bb3458fa98" (UID: "2258fa83-dbaf-4438-a9bf-42bb3458fa98"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.592207 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "2258fa83-dbaf-4438-a9bf-42bb3458fa98" (UID: "2258fa83-dbaf-4438-a9bf-42bb3458fa98"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.607690 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-inventory" (OuterVolumeSpecName: "inventory") pod "2258fa83-dbaf-4438-a9bf-42bb3458fa98" (UID: "2258fa83-dbaf-4438-a9bf-42bb3458fa98"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.656567 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.656614 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.656629 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.656641 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.656656 4775 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/2258fa83-dbaf-4438-a9bf-42bb3458fa98-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 04:14:50 crc kubenswrapper[4775]: I1002 04:14:50.656669 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbtkp\" (UniqueName: \"kubernetes.io/projected/2258fa83-dbaf-4438-a9bf-42bb3458fa98-kube-api-access-lbtkp\") on node \"crc\" DevicePath \"\"" Oct 02 04:14:51 crc kubenswrapper[4775]: I1002 04:14:51.005444 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" event={"ID":"2258fa83-dbaf-4438-a9bf-42bb3458fa98","Type":"ContainerDied","Data":"55b4b6f1b6a9ffe8f5c2acbba0497853aa8d44699f06dd59fca548dac4c63743"} Oct 02 04:14:51 crc kubenswrapper[4775]: I1002 04:14:51.005509 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="55b4b6f1b6a9ffe8f5c2acbba0497853aa8d44699f06dd59fca548dac4c63743" Oct 02 04:14:51 crc kubenswrapper[4775]: I1002 04:14:51.005587 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-p97fw" Oct 02 04:14:53 crc kubenswrapper[4775]: I1002 04:14:53.783194 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:14:53 crc kubenswrapper[4775]: E1002 04:14:53.784415 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.173551 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t"] Oct 02 04:15:00 crc kubenswrapper[4775]: E1002 04:15:00.174621 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2258fa83-dbaf-4438-a9bf-42bb3458fa98" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.174637 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2258fa83-dbaf-4438-a9bf-42bb3458fa98" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 02 04:15:00 crc kubenswrapper[4775]: E1002 04:15:00.174658 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ffc37b-a80a-4cda-be63-2269212fe894" containerName="extract-content" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.174666 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ffc37b-a80a-4cda-be63-2269212fe894" containerName="extract-content" Oct 02 04:15:00 crc kubenswrapper[4775]: E1002 04:15:00.174689 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ffc37b-a80a-4cda-be63-2269212fe894" containerName="registry-server" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.174700 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ffc37b-a80a-4cda-be63-2269212fe894" containerName="registry-server" Oct 02 04:15:00 crc kubenswrapper[4775]: E1002 04:15:00.174736 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ffc37b-a80a-4cda-be63-2269212fe894" containerName="extract-utilities" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.174745 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ffc37b-a80a-4cda-be63-2269212fe894" containerName="extract-utilities" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.175030 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="98ffc37b-a80a-4cda-be63-2269212fe894" containerName="registry-server" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.175067 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2258fa83-dbaf-4438-a9bf-42bb3458fa98" containerName="neutron-dhcp-openstack-openstack-cell1" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.176018 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.180451 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.180654 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.185578 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t"] Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.279529 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8a9f798-bbde-4fed-83db-f09531e71211-secret-volume\") pod \"collect-profiles-29322975-sr54t\" (UID: \"e8a9f798-bbde-4fed-83db-f09531e71211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.279628 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8a9f798-bbde-4fed-83db-f09531e71211-config-volume\") pod \"collect-profiles-29322975-sr54t\" (UID: \"e8a9f798-bbde-4fed-83db-f09531e71211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.279659 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xbl9\" (UniqueName: \"kubernetes.io/projected/e8a9f798-bbde-4fed-83db-f09531e71211-kube-api-access-9xbl9\") pod \"collect-profiles-29322975-sr54t\" (UID: \"e8a9f798-bbde-4fed-83db-f09531e71211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.381333 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8a9f798-bbde-4fed-83db-f09531e71211-secret-volume\") pod \"collect-profiles-29322975-sr54t\" (UID: \"e8a9f798-bbde-4fed-83db-f09531e71211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.381453 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8a9f798-bbde-4fed-83db-f09531e71211-config-volume\") pod \"collect-profiles-29322975-sr54t\" (UID: \"e8a9f798-bbde-4fed-83db-f09531e71211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.381483 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xbl9\" (UniqueName: \"kubernetes.io/projected/e8a9f798-bbde-4fed-83db-f09531e71211-kube-api-access-9xbl9\") pod \"collect-profiles-29322975-sr54t\" (UID: \"e8a9f798-bbde-4fed-83db-f09531e71211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.382936 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8a9f798-bbde-4fed-83db-f09531e71211-config-volume\") pod \"collect-profiles-29322975-sr54t\" (UID: \"e8a9f798-bbde-4fed-83db-f09531e71211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.391346 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8a9f798-bbde-4fed-83db-f09531e71211-secret-volume\") pod \"collect-profiles-29322975-sr54t\" (UID: \"e8a9f798-bbde-4fed-83db-f09531e71211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.402103 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xbl9\" (UniqueName: \"kubernetes.io/projected/e8a9f798-bbde-4fed-83db-f09531e71211-kube-api-access-9xbl9\") pod \"collect-profiles-29322975-sr54t\" (UID: \"e8a9f798-bbde-4fed-83db-f09531e71211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" Oct 02 04:15:00 crc kubenswrapper[4775]: I1002 04:15:00.499038 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" Oct 02 04:15:01 crc kubenswrapper[4775]: I1002 04:15:01.004565 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t"] Oct 02 04:15:01 crc kubenswrapper[4775]: I1002 04:15:01.139844 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" event={"ID":"e8a9f798-bbde-4fed-83db-f09531e71211","Type":"ContainerStarted","Data":"dd6158dc02c40c177a97284efef52f0dd4b5e41ec5a7684ce9d1cb7586c71959"} Oct 02 04:15:01 crc kubenswrapper[4775]: E1002 04:15:01.637179 4775 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8a9f798_bbde_4fed_83db_f09531e71211.slice/crio-b3fd90efd3282bf96e9df5c7d5ccc9c3bcc17dd46ce4e466f57a683fc75f76df.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8a9f798_bbde_4fed_83db_f09531e71211.slice/crio-conmon-b3fd90efd3282bf96e9df5c7d5ccc9c3bcc17dd46ce4e466f57a683fc75f76df.scope\": RecentStats: unable to find data in memory cache]" Oct 02 04:15:02 crc kubenswrapper[4775]: I1002 04:15:02.157876 4775 generic.go:334] "Generic (PLEG): container finished" podID="e8a9f798-bbde-4fed-83db-f09531e71211" containerID="b3fd90efd3282bf96e9df5c7d5ccc9c3bcc17dd46ce4e466f57a683fc75f76df" exitCode=0 Oct 02 04:15:02 crc kubenswrapper[4775]: I1002 04:15:02.158020 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" event={"ID":"e8a9f798-bbde-4fed-83db-f09531e71211","Type":"ContainerDied","Data":"b3fd90efd3282bf96e9df5c7d5ccc9c3bcc17dd46ce4e466f57a683fc75f76df"} Oct 02 04:15:03 crc kubenswrapper[4775]: I1002 04:15:03.713729 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" Oct 02 04:15:03 crc kubenswrapper[4775]: I1002 04:15:03.866971 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8a9f798-bbde-4fed-83db-f09531e71211-secret-volume\") pod \"e8a9f798-bbde-4fed-83db-f09531e71211\" (UID: \"e8a9f798-bbde-4fed-83db-f09531e71211\") " Oct 02 04:15:03 crc kubenswrapper[4775]: I1002 04:15:03.867138 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xbl9\" (UniqueName: \"kubernetes.io/projected/e8a9f798-bbde-4fed-83db-f09531e71211-kube-api-access-9xbl9\") pod \"e8a9f798-bbde-4fed-83db-f09531e71211\" (UID: \"e8a9f798-bbde-4fed-83db-f09531e71211\") " Oct 02 04:15:03 crc kubenswrapper[4775]: I1002 04:15:03.867172 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8a9f798-bbde-4fed-83db-f09531e71211-config-volume\") pod \"e8a9f798-bbde-4fed-83db-f09531e71211\" (UID: \"e8a9f798-bbde-4fed-83db-f09531e71211\") " Oct 02 04:15:03 crc kubenswrapper[4775]: I1002 04:15:03.868199 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8a9f798-bbde-4fed-83db-f09531e71211-config-volume" (OuterVolumeSpecName: "config-volume") pod "e8a9f798-bbde-4fed-83db-f09531e71211" (UID: "e8a9f798-bbde-4fed-83db-f09531e71211"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 04:15:03 crc kubenswrapper[4775]: I1002 04:15:03.873093 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8a9f798-bbde-4fed-83db-f09531e71211-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e8a9f798-bbde-4fed-83db-f09531e71211" (UID: "e8a9f798-bbde-4fed-83db-f09531e71211"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:15:03 crc kubenswrapper[4775]: I1002 04:15:03.874620 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8a9f798-bbde-4fed-83db-f09531e71211-kube-api-access-9xbl9" (OuterVolumeSpecName: "kube-api-access-9xbl9") pod "e8a9f798-bbde-4fed-83db-f09531e71211" (UID: "e8a9f798-bbde-4fed-83db-f09531e71211"). InnerVolumeSpecName "kube-api-access-9xbl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:15:03 crc kubenswrapper[4775]: I1002 04:15:03.970192 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xbl9\" (UniqueName: \"kubernetes.io/projected/e8a9f798-bbde-4fed-83db-f09531e71211-kube-api-access-9xbl9\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:03 crc kubenswrapper[4775]: I1002 04:15:03.970235 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e8a9f798-bbde-4fed-83db-f09531e71211-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:03 crc kubenswrapper[4775]: I1002 04:15:03.970247 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e8a9f798-bbde-4fed-83db-f09531e71211-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:04 crc kubenswrapper[4775]: I1002 04:15:04.187482 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" event={"ID":"e8a9f798-bbde-4fed-83db-f09531e71211","Type":"ContainerDied","Data":"dd6158dc02c40c177a97284efef52f0dd4b5e41ec5a7684ce9d1cb7586c71959"} Oct 02 04:15:04 crc kubenswrapper[4775]: I1002 04:15:04.188071 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd6158dc02c40c177a97284efef52f0dd4b5e41ec5a7684ce9d1cb7586c71959" Oct 02 04:15:04 crc kubenswrapper[4775]: I1002 04:15:04.187575 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322975-sr54t" Oct 02 04:15:04 crc kubenswrapper[4775]: I1002 04:15:04.814124 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6"] Oct 02 04:15:04 crc kubenswrapper[4775]: I1002 04:15:04.825111 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322930-kv9d6"] Oct 02 04:15:05 crc kubenswrapper[4775]: I1002 04:15:05.781544 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f05c8229-d21d-4c03-abc5-5e5ef54121ee" path="/var/lib/kubelet/pods/f05c8229-d21d-4c03-abc5-5e5ef54121ee/volumes" Oct 02 04:15:06 crc kubenswrapper[4775]: I1002 04:15:06.765458 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:15:06 crc kubenswrapper[4775]: E1002 04:15:06.766209 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:15:18 crc kubenswrapper[4775]: I1002 04:15:18.627853 4775 scope.go:117] "RemoveContainer" containerID="89cbeaae642f9ccfa7de8f57341e73a0e28d7ac4555c74e4c752835f00e9761c" Oct 02 04:15:19 crc kubenswrapper[4775]: I1002 04:15:19.765708 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:15:19 crc kubenswrapper[4775]: E1002 04:15:19.766283 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:15:21 crc kubenswrapper[4775]: I1002 04:15:21.069493 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 04:15:21 crc kubenswrapper[4775]: I1002 04:15:21.070039 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="4e865292-9dda-49fc-acdf-3afc04304aac" containerName="nova-cell0-conductor-conductor" containerID="cri-o://6e4b2842c6945dc728e9ab20f6eae59f1424306f45f162bb0205a42c38e05645" gracePeriod=30 Oct 02 04:15:21 crc kubenswrapper[4775]: I1002 04:15:21.143088 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 04:15:21 crc kubenswrapper[4775]: I1002 04:15:21.143336 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="6fababa8-2f3f-47dc-90cd-38105370efbe" containerName="nova-cell1-conductor-conductor" containerID="cri-o://3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f" gracePeriod=30 Oct 02 04:15:21 crc kubenswrapper[4775]: E1002 04:15:21.695737 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f is running failed: container process not found" containerID="3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 04:15:21 crc kubenswrapper[4775]: E1002 04:15:21.696534 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f is running failed: container process not found" containerID="3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 04:15:21 crc kubenswrapper[4775]: E1002 04:15:21.696915 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f is running failed: container process not found" containerID="3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 04:15:21 crc kubenswrapper[4775]: E1002 04:15:21.696941 4775 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="6fababa8-2f3f-47dc-90cd-38105370efbe" containerName="nova-cell1-conductor-conductor" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.145361 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.220427 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z94ql\" (UniqueName: \"kubernetes.io/projected/6fababa8-2f3f-47dc-90cd-38105370efbe-kube-api-access-z94ql\") pod \"6fababa8-2f3f-47dc-90cd-38105370efbe\" (UID: \"6fababa8-2f3f-47dc-90cd-38105370efbe\") " Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.220506 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fababa8-2f3f-47dc-90cd-38105370efbe-combined-ca-bundle\") pod \"6fababa8-2f3f-47dc-90cd-38105370efbe\" (UID: \"6fababa8-2f3f-47dc-90cd-38105370efbe\") " Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.220609 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fababa8-2f3f-47dc-90cd-38105370efbe-config-data\") pod \"6fababa8-2f3f-47dc-90cd-38105370efbe\" (UID: \"6fababa8-2f3f-47dc-90cd-38105370efbe\") " Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.242705 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fababa8-2f3f-47dc-90cd-38105370efbe-kube-api-access-z94ql" (OuterVolumeSpecName: "kube-api-access-z94ql") pod "6fababa8-2f3f-47dc-90cd-38105370efbe" (UID: "6fababa8-2f3f-47dc-90cd-38105370efbe"). InnerVolumeSpecName "kube-api-access-z94ql". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.330674 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z94ql\" (UniqueName: \"kubernetes.io/projected/6fababa8-2f3f-47dc-90cd-38105370efbe-kube-api-access-z94ql\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.343118 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fababa8-2f3f-47dc-90cd-38105370efbe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6fababa8-2f3f-47dc-90cd-38105370efbe" (UID: "6fababa8-2f3f-47dc-90cd-38105370efbe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.366362 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fababa8-2f3f-47dc-90cd-38105370efbe-config-data" (OuterVolumeSpecName: "config-data") pod "6fababa8-2f3f-47dc-90cd-38105370efbe" (UID: "6fababa8-2f3f-47dc-90cd-38105370efbe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.402776 4775 generic.go:334] "Generic (PLEG): container finished" podID="6fababa8-2f3f-47dc-90cd-38105370efbe" containerID="3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f" exitCode=0 Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.402816 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6fababa8-2f3f-47dc-90cd-38105370efbe","Type":"ContainerDied","Data":"3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f"} Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.402840 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6fababa8-2f3f-47dc-90cd-38105370efbe","Type":"ContainerDied","Data":"cb2d5bd18f0c9a1f44fffb7031c59ce8d7066d4c9bc6109c23dd69e176b7a43a"} Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.402856 4775 scope.go:117] "RemoveContainer" containerID="3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.402968 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.431442 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fababa8-2f3f-47dc-90cd-38105370efbe-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.431477 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fababa8-2f3f-47dc-90cd-38105370efbe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.454256 4775 scope.go:117] "RemoveContainer" containerID="3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f" Oct 02 04:15:22 crc kubenswrapper[4775]: E1002 04:15:22.459286 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f\": container with ID starting with 3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f not found: ID does not exist" containerID="3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.459315 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f"} err="failed to get container status \"3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f\": rpc error: code = NotFound desc = could not find container \"3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f\": container with ID starting with 3cc2042b2b6020eef5fb6846cb79e5bb4f86476a52979fe91ce2987cb7b5191f not found: ID does not exist" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.464004 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.478846 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.496531 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 04:15:22 crc kubenswrapper[4775]: E1002 04:15:22.496947 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8a9f798-bbde-4fed-83db-f09531e71211" containerName="collect-profiles" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.496987 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8a9f798-bbde-4fed-83db-f09531e71211" containerName="collect-profiles" Oct 02 04:15:22 crc kubenswrapper[4775]: E1002 04:15:22.497051 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fababa8-2f3f-47dc-90cd-38105370efbe" containerName="nova-cell1-conductor-conductor" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.497059 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fababa8-2f3f-47dc-90cd-38105370efbe" containerName="nova-cell1-conductor-conductor" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.497241 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fababa8-2f3f-47dc-90cd-38105370efbe" containerName="nova-cell1-conductor-conductor" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.497271 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8a9f798-bbde-4fed-83db-f09531e71211" containerName="collect-profiles" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.497944 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.502726 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.543941 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx7fd\" (UniqueName: \"kubernetes.io/projected/4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84-kube-api-access-zx7fd\") pod \"nova-cell1-conductor-0\" (UID: \"4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84\") " pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.544010 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84\") " pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.544053 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84\") " pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.550803 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.646594 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx7fd\" (UniqueName: \"kubernetes.io/projected/4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84-kube-api-access-zx7fd\") pod \"nova-cell1-conductor-0\" (UID: \"4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84\") " pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.646928 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84\") " pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:22 crc kubenswrapper[4775]: I1002 04:15:22.646991 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84\") " pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:22 crc kubenswrapper[4775]: E1002 04:15:22.760294 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6e4b2842c6945dc728e9ab20f6eae59f1424306f45f162bb0205a42c38e05645" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 04:15:22 crc kubenswrapper[4775]: E1002 04:15:22.761945 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6e4b2842c6945dc728e9ab20f6eae59f1424306f45f162bb0205a42c38e05645" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 04:15:22 crc kubenswrapper[4775]: E1002 04:15:22.763655 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6e4b2842c6945dc728e9ab20f6eae59f1424306f45f162bb0205a42c38e05645" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Oct 02 04:15:22 crc kubenswrapper[4775]: E1002 04:15:22.763712 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="4e865292-9dda-49fc-acdf-3afc04304aac" containerName="nova-cell0-conductor-conductor" Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.085116 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.085483 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8986cac6-b070-4ea8-a961-24481fbb7bdd" containerName="nova-api-log" containerID="cri-o://126f1fe0b63d9c60ba34b2fc8a728348a966ea5fb359156765cd0bce9ddb9ee4" gracePeriod=30 Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.085608 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="8986cac6-b070-4ea8-a961-24481fbb7bdd" containerName="nova-api-api" containerID="cri-o://276887ff0f0d687bacfb0513089001e62180ad08e2d3bcfc486bc4e5bd888e9f" gracePeriod=30 Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.116930 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.117433 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="02de0e68-b4be-42b1-80eb-b4cb50dbd456" containerName="nova-scheduler-scheduler" containerID="cri-o://2d5246027af2edbc8f1f0bcfb9f5c32756879a301b041860ced369a8eae8dc35" gracePeriod=30 Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.131902 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.132163 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="def5a435-381a-4636-932e-8e4881989de3" containerName="nova-metadata-log" containerID="cri-o://36093f2c01943f0cca62904703c8989a7da89a2ab6946a9b8cec36ca74fb80b4" gracePeriod=30 Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.132288 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="def5a435-381a-4636-932e-8e4881989de3" containerName="nova-metadata-metadata" containerID="cri-o://4c136b0ad6b8a589bc80d338cd4a23c9b9f0075532593ecff610f11104b1f4a9" gracePeriod=30 Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.184807 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84\") " pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.184986 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84\") " pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.187414 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx7fd\" (UniqueName: \"kubernetes.io/projected/4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84-kube-api-access-zx7fd\") pod \"nova-cell1-conductor-0\" (UID: \"4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84\") " pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.415295 4775 generic.go:334] "Generic (PLEG): container finished" podID="8986cac6-b070-4ea8-a961-24481fbb7bdd" containerID="126f1fe0b63d9c60ba34b2fc8a728348a966ea5fb359156765cd0bce9ddb9ee4" exitCode=143 Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.415491 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8986cac6-b070-4ea8-a961-24481fbb7bdd","Type":"ContainerDied","Data":"126f1fe0b63d9c60ba34b2fc8a728348a966ea5fb359156765cd0bce9ddb9ee4"} Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.417779 4775 generic.go:334] "Generic (PLEG): container finished" podID="def5a435-381a-4636-932e-8e4881989de3" containerID="36093f2c01943f0cca62904703c8989a7da89a2ab6946a9b8cec36ca74fb80b4" exitCode=143 Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.417813 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"def5a435-381a-4636-932e-8e4881989de3","Type":"ContainerDied","Data":"36093f2c01943f0cca62904703c8989a7da89a2ab6946a9b8cec36ca74fb80b4"} Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.455940 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.783336 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fababa8-2f3f-47dc-90cd-38105370efbe" path="/var/lib/kubelet/pods/6fababa8-2f3f-47dc-90cd-38105370efbe/volumes" Oct 02 04:15:23 crc kubenswrapper[4775]: I1002 04:15:23.938138 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 02 04:15:24 crc kubenswrapper[4775]: I1002 04:15:24.428151 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84","Type":"ContainerStarted","Data":"a4745e4e7257fe1decd3334cad61069430a3467c3ff1de806f656bd282de22bf"} Oct 02 04:15:24 crc kubenswrapper[4775]: I1002 04:15:24.428520 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84","Type":"ContainerStarted","Data":"3f9652e1b11367dabc5cdb032067d4252b445f9a382f232000aadd03a5177dd5"} Oct 02 04:15:24 crc kubenswrapper[4775]: I1002 04:15:24.429285 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:24 crc kubenswrapper[4775]: I1002 04:15:24.448327 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.4483033 podStartE2EDuration="2.4483033s" podCreationTimestamp="2025-10-02 04:15:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 04:15:24.439882041 +0000 UTC m=+9261.606626091" watchObservedRunningTime="2025-10-02 04:15:24.4483033 +0000 UTC m=+9261.615047340" Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.272730 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="def5a435-381a-4636-932e-8e4881989de3" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.85:8775/\": read tcp 10.217.0.2:45414->10.217.1.85:8775: read: connection reset by peer" Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.273138 4775 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="def5a435-381a-4636-932e-8e4881989de3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.85:8775/\": read tcp 10.217.0.2:45408->10.217.1.85:8775: read: connection reset by peer" Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.548198 4775 generic.go:334] "Generic (PLEG): container finished" podID="8986cac6-b070-4ea8-a961-24481fbb7bdd" containerID="276887ff0f0d687bacfb0513089001e62180ad08e2d3bcfc486bc4e5bd888e9f" exitCode=0 Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.548278 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8986cac6-b070-4ea8-a961-24481fbb7bdd","Type":"ContainerDied","Data":"276887ff0f0d687bacfb0513089001e62180ad08e2d3bcfc486bc4e5bd888e9f"} Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.605105 4775 generic.go:334] "Generic (PLEG): container finished" podID="def5a435-381a-4636-932e-8e4881989de3" containerID="4c136b0ad6b8a589bc80d338cd4a23c9b9f0075532593ecff610f11104b1f4a9" exitCode=0 Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.605148 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"def5a435-381a-4636-932e-8e4881989de3","Type":"ContainerDied","Data":"4c136b0ad6b8a589bc80d338cd4a23c9b9f0075532593ecff610f11104b1f4a9"} Oct 02 04:15:26 crc kubenswrapper[4775]: E1002 04:15:26.627205 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2d5246027af2edbc8f1f0bcfb9f5c32756879a301b041860ced369a8eae8dc35" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 04:15:26 crc kubenswrapper[4775]: E1002 04:15:26.669336 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2d5246027af2edbc8f1f0bcfb9f5c32756879a301b041860ced369a8eae8dc35" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 04:15:26 crc kubenswrapper[4775]: E1002 04:15:26.691585 4775 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="2d5246027af2edbc8f1f0bcfb9f5c32756879a301b041860ced369a8eae8dc35" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 02 04:15:26 crc kubenswrapper[4775]: E1002 04:15:26.691843 4775 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="02de0e68-b4be-42b1-80eb-b4cb50dbd456" containerName="nova-scheduler-scheduler" Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.862515 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.879043 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jthvb\" (UniqueName: \"kubernetes.io/projected/8986cac6-b070-4ea8-a961-24481fbb7bdd-kube-api-access-jthvb\") pod \"8986cac6-b070-4ea8-a961-24481fbb7bdd\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.879135 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8986cac6-b070-4ea8-a961-24481fbb7bdd-logs\") pod \"8986cac6-b070-4ea8-a961-24481fbb7bdd\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.879248 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8986cac6-b070-4ea8-a961-24481fbb7bdd-combined-ca-bundle\") pod \"8986cac6-b070-4ea8-a961-24481fbb7bdd\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.879305 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8986cac6-b070-4ea8-a961-24481fbb7bdd-config-data\") pod \"8986cac6-b070-4ea8-a961-24481fbb7bdd\" (UID: \"8986cac6-b070-4ea8-a961-24481fbb7bdd\") " Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.883427 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8986cac6-b070-4ea8-a961-24481fbb7bdd-logs" (OuterVolumeSpecName: "logs") pod "8986cac6-b070-4ea8-a961-24481fbb7bdd" (UID: "8986cac6-b070-4ea8-a961-24481fbb7bdd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.904237 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8986cac6-b070-4ea8-a961-24481fbb7bdd-kube-api-access-jthvb" (OuterVolumeSpecName: "kube-api-access-jthvb") pod "8986cac6-b070-4ea8-a961-24481fbb7bdd" (UID: "8986cac6-b070-4ea8-a961-24481fbb7bdd"). InnerVolumeSpecName "kube-api-access-jthvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.958416 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.958939 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8986cac6-b070-4ea8-a961-24481fbb7bdd-config-data" (OuterVolumeSpecName: "config-data") pod "8986cac6-b070-4ea8-a961-24481fbb7bdd" (UID: "8986cac6-b070-4ea8-a961-24481fbb7bdd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.962325 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8986cac6-b070-4ea8-a961-24481fbb7bdd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8986cac6-b070-4ea8-a961-24481fbb7bdd" (UID: "8986cac6-b070-4ea8-a961-24481fbb7bdd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.982332 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jthvb\" (UniqueName: \"kubernetes.io/projected/8986cac6-b070-4ea8-a961-24481fbb7bdd-kube-api-access-jthvb\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.982364 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8986cac6-b070-4ea8-a961-24481fbb7bdd-logs\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.982378 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8986cac6-b070-4ea8-a961-24481fbb7bdd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:26 crc kubenswrapper[4775]: I1002 04:15:26.982386 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8986cac6-b070-4ea8-a961-24481fbb7bdd-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.083426 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/def5a435-381a-4636-932e-8e4881989de3-logs\") pod \"def5a435-381a-4636-932e-8e4881989de3\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.083585 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljlqn\" (UniqueName: \"kubernetes.io/projected/def5a435-381a-4636-932e-8e4881989de3-kube-api-access-ljlqn\") pod \"def5a435-381a-4636-932e-8e4881989de3\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.083660 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/def5a435-381a-4636-932e-8e4881989de3-config-data\") pod \"def5a435-381a-4636-932e-8e4881989de3\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.083757 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/def5a435-381a-4636-932e-8e4881989de3-combined-ca-bundle\") pod \"def5a435-381a-4636-932e-8e4881989de3\" (UID: \"def5a435-381a-4636-932e-8e4881989de3\") " Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.084706 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/def5a435-381a-4636-932e-8e4881989de3-logs" (OuterVolumeSpecName: "logs") pod "def5a435-381a-4636-932e-8e4881989de3" (UID: "def5a435-381a-4636-932e-8e4881989de3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.089112 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/def5a435-381a-4636-932e-8e4881989de3-kube-api-access-ljlqn" (OuterVolumeSpecName: "kube-api-access-ljlqn") pod "def5a435-381a-4636-932e-8e4881989de3" (UID: "def5a435-381a-4636-932e-8e4881989de3"). InnerVolumeSpecName "kube-api-access-ljlqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.126095 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/def5a435-381a-4636-932e-8e4881989de3-config-data" (OuterVolumeSpecName: "config-data") pod "def5a435-381a-4636-932e-8e4881989de3" (UID: "def5a435-381a-4636-932e-8e4881989de3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.135102 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/def5a435-381a-4636-932e-8e4881989de3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "def5a435-381a-4636-932e-8e4881989de3" (UID: "def5a435-381a-4636-932e-8e4881989de3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.188424 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/def5a435-381a-4636-932e-8e4881989de3-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.188458 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/def5a435-381a-4636-932e-8e4881989de3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.188469 4775 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/def5a435-381a-4636-932e-8e4881989de3-logs\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.188477 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljlqn\" (UniqueName: \"kubernetes.io/projected/def5a435-381a-4636-932e-8e4881989de3-kube-api-access-ljlqn\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.627522 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"8986cac6-b070-4ea8-a961-24481fbb7bdd","Type":"ContainerDied","Data":"49734b270b32bcae1851d56616d6f946eaf31914db2772c5124d03d9e0cd5089"} Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.627599 4775 scope.go:117] "RemoveContainer" containerID="276887ff0f0d687bacfb0513089001e62180ad08e2d3bcfc486bc4e5bd888e9f" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.627797 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.633616 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"def5a435-381a-4636-932e-8e4881989de3","Type":"ContainerDied","Data":"298fd9920f4ecbc49ae3edaaccc776395e881b56c77c9178b9ebab80fcee16c2"} Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.633704 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.641872 4775 generic.go:334] "Generic (PLEG): container finished" podID="4e865292-9dda-49fc-acdf-3afc04304aac" containerID="6e4b2842c6945dc728e9ab20f6eae59f1424306f45f162bb0205a42c38e05645" exitCode=0 Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.641908 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4e865292-9dda-49fc-acdf-3afc04304aac","Type":"ContainerDied","Data":"6e4b2842c6945dc728e9ab20f6eae59f1424306f45f162bb0205a42c38e05645"} Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.641931 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"4e865292-9dda-49fc-acdf-3afc04304aac","Type":"ContainerDied","Data":"c89a4734eb7e14a5d749a8ae821c124e6672db8a27a6f24bd494d1b09799be33"} Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.641941 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c89a4734eb7e14a5d749a8ae821c124e6672db8a27a6f24bd494d1b09799be33" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.702218 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.729231 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.731205 4775 scope.go:117] "RemoveContainer" containerID="126f1fe0b63d9c60ba34b2fc8a728348a966ea5fb359156765cd0bce9ddb9ee4" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.796519 4775 scope.go:117] "RemoveContainer" containerID="4c136b0ad6b8a589bc80d338cd4a23c9b9f0075532593ecff610f11104b1f4a9" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.810050 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.810093 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.815227 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.824660 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 02 04:15:27 crc kubenswrapper[4775]: E1002 04:15:27.825164 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def5a435-381a-4636-932e-8e4881989de3" containerName="nova-metadata-log" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.825192 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="def5a435-381a-4636-932e-8e4881989de3" containerName="nova-metadata-log" Oct 02 04:15:27 crc kubenswrapper[4775]: E1002 04:15:27.825215 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8986cac6-b070-4ea8-a961-24481fbb7bdd" containerName="nova-api-log" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.825225 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8986cac6-b070-4ea8-a961-24481fbb7bdd" containerName="nova-api-log" Oct 02 04:15:27 crc kubenswrapper[4775]: E1002 04:15:27.825248 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e865292-9dda-49fc-acdf-3afc04304aac" containerName="nova-cell0-conductor-conductor" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.825256 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e865292-9dda-49fc-acdf-3afc04304aac" containerName="nova-cell0-conductor-conductor" Oct 02 04:15:27 crc kubenswrapper[4775]: E1002 04:15:27.825283 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="def5a435-381a-4636-932e-8e4881989de3" containerName="nova-metadata-metadata" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.825291 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="def5a435-381a-4636-932e-8e4881989de3" containerName="nova-metadata-metadata" Oct 02 04:15:27 crc kubenswrapper[4775]: E1002 04:15:27.825320 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8986cac6-b070-4ea8-a961-24481fbb7bdd" containerName="nova-api-api" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.825327 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="8986cac6-b070-4ea8-a961-24481fbb7bdd" containerName="nova-api-api" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.825569 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e865292-9dda-49fc-acdf-3afc04304aac" containerName="nova-cell0-conductor-conductor" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.825602 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="def5a435-381a-4636-932e-8e4881989de3" containerName="nova-metadata-log" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.825617 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8986cac6-b070-4ea8-a961-24481fbb7bdd" containerName="nova-api-log" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.825636 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="8986cac6-b070-4ea8-a961-24481fbb7bdd" containerName="nova-api-api" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.825647 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="def5a435-381a-4636-932e-8e4881989de3" containerName="nova-metadata-metadata" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.827334 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.831389 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.834717 4775 scope.go:117] "RemoveContainer" containerID="36093f2c01943f0cca62904703c8989a7da89a2ab6946a9b8cec36ca74fb80b4" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.835983 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.838197 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.839949 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.847738 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.875450 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.903184 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mn6wm\" (UniqueName: \"kubernetes.io/projected/4e865292-9dda-49fc-acdf-3afc04304aac-kube-api-access-mn6wm\") pod \"4e865292-9dda-49fc-acdf-3afc04304aac\" (UID: \"4e865292-9dda-49fc-acdf-3afc04304aac\") " Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.903293 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e865292-9dda-49fc-acdf-3afc04304aac-config-data\") pod \"4e865292-9dda-49fc-acdf-3afc04304aac\" (UID: \"4e865292-9dda-49fc-acdf-3afc04304aac\") " Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.903368 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e865292-9dda-49fc-acdf-3afc04304aac-combined-ca-bundle\") pod \"4e865292-9dda-49fc-acdf-3afc04304aac\" (UID: \"4e865292-9dda-49fc-acdf-3afc04304aac\") " Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.908411 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e865292-9dda-49fc-acdf-3afc04304aac-kube-api-access-mn6wm" (OuterVolumeSpecName: "kube-api-access-mn6wm") pod "4e865292-9dda-49fc-acdf-3afc04304aac" (UID: "4e865292-9dda-49fc-acdf-3afc04304aac"). InnerVolumeSpecName "kube-api-access-mn6wm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.940146 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e865292-9dda-49fc-acdf-3afc04304aac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4e865292-9dda-49fc-acdf-3afc04304aac" (UID: "4e865292-9dda-49fc-acdf-3afc04304aac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:15:27 crc kubenswrapper[4775]: I1002 04:15:27.947177 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4e865292-9dda-49fc-acdf-3afc04304aac-config-data" (OuterVolumeSpecName: "config-data") pod "4e865292-9dda-49fc-acdf-3afc04304aac" (UID: "4e865292-9dda-49fc-acdf-3afc04304aac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.006512 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7pf7\" (UniqueName: \"kubernetes.io/projected/20d47d00-0c95-4858-963c-ac98a3b56306-kube-api-access-c7pf7\") pod \"nova-api-0\" (UID: \"20d47d00-0c95-4858-963c-ac98a3b56306\") " pod="openstack/nova-api-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.006581 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cflf\" (UniqueName: \"kubernetes.io/projected/283844d0-a97d-4bc0-b297-5e078862a5b9-kube-api-access-5cflf\") pod \"nova-metadata-0\" (UID: \"283844d0-a97d-4bc0-b297-5e078862a5b9\") " pod="openstack/nova-metadata-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.006651 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/283844d0-a97d-4bc0-b297-5e078862a5b9-config-data\") pod \"nova-metadata-0\" (UID: \"283844d0-a97d-4bc0-b297-5e078862a5b9\") " pod="openstack/nova-metadata-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.006736 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d47d00-0c95-4858-963c-ac98a3b56306-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"20d47d00-0c95-4858-963c-ac98a3b56306\") " pod="openstack/nova-api-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.006770 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d47d00-0c95-4858-963c-ac98a3b56306-config-data\") pod \"nova-api-0\" (UID: \"20d47d00-0c95-4858-963c-ac98a3b56306\") " pod="openstack/nova-api-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.006831 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20d47d00-0c95-4858-963c-ac98a3b56306-logs\") pod \"nova-api-0\" (UID: \"20d47d00-0c95-4858-963c-ac98a3b56306\") " pod="openstack/nova-api-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.006911 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/283844d0-a97d-4bc0-b297-5e078862a5b9-logs\") pod \"nova-metadata-0\" (UID: \"283844d0-a97d-4bc0-b297-5e078862a5b9\") " pod="openstack/nova-metadata-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.007013 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/283844d0-a97d-4bc0-b297-5e078862a5b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"283844d0-a97d-4bc0-b297-5e078862a5b9\") " pod="openstack/nova-metadata-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.007161 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn6wm\" (UniqueName: \"kubernetes.io/projected/4e865292-9dda-49fc-acdf-3afc04304aac-kube-api-access-mn6wm\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.007190 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4e865292-9dda-49fc-acdf-3afc04304aac-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.007208 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4e865292-9dda-49fc-acdf-3afc04304aac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.109984 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/283844d0-a97d-4bc0-b297-5e078862a5b9-logs\") pod \"nova-metadata-0\" (UID: \"283844d0-a97d-4bc0-b297-5e078862a5b9\") " pod="openstack/nova-metadata-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.110053 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/283844d0-a97d-4bc0-b297-5e078862a5b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"283844d0-a97d-4bc0-b297-5e078862a5b9\") " pod="openstack/nova-metadata-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.110144 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7pf7\" (UniqueName: \"kubernetes.io/projected/20d47d00-0c95-4858-963c-ac98a3b56306-kube-api-access-c7pf7\") pod \"nova-api-0\" (UID: \"20d47d00-0c95-4858-963c-ac98a3b56306\") " pod="openstack/nova-api-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.110174 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cflf\" (UniqueName: \"kubernetes.io/projected/283844d0-a97d-4bc0-b297-5e078862a5b9-kube-api-access-5cflf\") pod \"nova-metadata-0\" (UID: \"283844d0-a97d-4bc0-b297-5e078862a5b9\") " pod="openstack/nova-metadata-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.110209 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/283844d0-a97d-4bc0-b297-5e078862a5b9-config-data\") pod \"nova-metadata-0\" (UID: \"283844d0-a97d-4bc0-b297-5e078862a5b9\") " pod="openstack/nova-metadata-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.110253 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d47d00-0c95-4858-963c-ac98a3b56306-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"20d47d00-0c95-4858-963c-ac98a3b56306\") " pod="openstack/nova-api-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.110275 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d47d00-0c95-4858-963c-ac98a3b56306-config-data\") pod \"nova-api-0\" (UID: \"20d47d00-0c95-4858-963c-ac98a3b56306\") " pod="openstack/nova-api-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.110851 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20d47d00-0c95-4858-963c-ac98a3b56306-logs\") pod \"nova-api-0\" (UID: \"20d47d00-0c95-4858-963c-ac98a3b56306\") " pod="openstack/nova-api-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.110896 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/20d47d00-0c95-4858-963c-ac98a3b56306-logs\") pod \"nova-api-0\" (UID: \"20d47d00-0c95-4858-963c-ac98a3b56306\") " pod="openstack/nova-api-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.110853 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/283844d0-a97d-4bc0-b297-5e078862a5b9-logs\") pod \"nova-metadata-0\" (UID: \"283844d0-a97d-4bc0-b297-5e078862a5b9\") " pod="openstack/nova-metadata-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.113971 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/20d47d00-0c95-4858-963c-ac98a3b56306-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"20d47d00-0c95-4858-963c-ac98a3b56306\") " pod="openstack/nova-api-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.114977 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/283844d0-a97d-4bc0-b297-5e078862a5b9-config-data\") pod \"nova-metadata-0\" (UID: \"283844d0-a97d-4bc0-b297-5e078862a5b9\") " pod="openstack/nova-metadata-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.115416 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/283844d0-a97d-4bc0-b297-5e078862a5b9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"283844d0-a97d-4bc0-b297-5e078862a5b9\") " pod="openstack/nova-metadata-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.126247 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/20d47d00-0c95-4858-963c-ac98a3b56306-config-data\") pod \"nova-api-0\" (UID: \"20d47d00-0c95-4858-963c-ac98a3b56306\") " pod="openstack/nova-api-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.134145 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cflf\" (UniqueName: \"kubernetes.io/projected/283844d0-a97d-4bc0-b297-5e078862a5b9-kube-api-access-5cflf\") pod \"nova-metadata-0\" (UID: \"283844d0-a97d-4bc0-b297-5e078862a5b9\") " pod="openstack/nova-metadata-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.134272 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7pf7\" (UniqueName: \"kubernetes.io/projected/20d47d00-0c95-4858-963c-ac98a3b56306-kube-api-access-c7pf7\") pod \"nova-api-0\" (UID: \"20d47d00-0c95-4858-963c-ac98a3b56306\") " pod="openstack/nova-api-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.159477 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.174133 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.656758 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:28 crc kubenswrapper[4775]: W1002 04:15:28.701464 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20d47d00_0c95_4858_963c_ac98a3b56306.slice/crio-3b4d04f2b72f04b105d20f8eb3411ca9621c7dd45e9586c2478e2bfd2c396303 WatchSource:0}: Error finding container 3b4d04f2b72f04b105d20f8eb3411ca9621c7dd45e9586c2478e2bfd2c396303: Status 404 returned error can't find the container with id 3b4d04f2b72f04b105d20f8eb3411ca9621c7dd45e9586c2478e2bfd2c396303 Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.704304 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.717381 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.728124 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.739992 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.741429 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.745024 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.770109 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.794039 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 02 04:15:28 crc kubenswrapper[4775]: W1002 04:15:28.799933 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod283844d0_a97d_4bc0_b297_5e078862a5b9.slice/crio-19ca85a155bbba598b5be9d5345750517ffa3c8817cbcc12411fa43a5d75e018 WatchSource:0}: Error finding container 19ca85a155bbba598b5be9d5345750517ffa3c8817cbcc12411fa43a5d75e018: Status 404 returned error can't find the container with id 19ca85a155bbba598b5be9d5345750517ffa3c8817cbcc12411fa43a5d75e018 Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.927279 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dbfa9c9-66fa-48ad-956f-52297602546e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1dbfa9c9-66fa-48ad-956f-52297602546e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.927401 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dbfa9c9-66fa-48ad-956f-52297602546e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1dbfa9c9-66fa-48ad-956f-52297602546e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:28 crc kubenswrapper[4775]: I1002 04:15:28.927991 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8cld\" (UniqueName: \"kubernetes.io/projected/1dbfa9c9-66fa-48ad-956f-52297602546e-kube-api-access-n8cld\") pod \"nova-cell0-conductor-0\" (UID: \"1dbfa9c9-66fa-48ad-956f-52297602546e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.029623 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8cld\" (UniqueName: \"kubernetes.io/projected/1dbfa9c9-66fa-48ad-956f-52297602546e-kube-api-access-n8cld\") pod \"nova-cell0-conductor-0\" (UID: \"1dbfa9c9-66fa-48ad-956f-52297602546e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.029712 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dbfa9c9-66fa-48ad-956f-52297602546e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1dbfa9c9-66fa-48ad-956f-52297602546e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.029746 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dbfa9c9-66fa-48ad-956f-52297602546e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1dbfa9c9-66fa-48ad-956f-52297602546e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.033365 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1dbfa9c9-66fa-48ad-956f-52297602546e-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1dbfa9c9-66fa-48ad-956f-52297602546e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.038190 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1dbfa9c9-66fa-48ad-956f-52297602546e-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1dbfa9c9-66fa-48ad-956f-52297602546e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.049788 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8cld\" (UniqueName: \"kubernetes.io/projected/1dbfa9c9-66fa-48ad-956f-52297602546e-kube-api-access-n8cld\") pod \"nova-cell0-conductor-0\" (UID: \"1dbfa9c9-66fa-48ad-956f-52297602546e\") " pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.236276 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.677932 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20d47d00-0c95-4858-963c-ac98a3b56306","Type":"ContainerStarted","Data":"066f6cd170beda852dd6a21ef2119d7280b922447015b89174692bccb5467bae"} Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.678334 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20d47d00-0c95-4858-963c-ac98a3b56306","Type":"ContainerStarted","Data":"50a434da08db5eddfa6bdf7e5fefbc6bc05aebe88139066363e4ddd9112c06bf"} Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.678350 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"20d47d00-0c95-4858-963c-ac98a3b56306","Type":"ContainerStarted","Data":"3b4d04f2b72f04b105d20f8eb3411ca9621c7dd45e9586c2478e2bfd2c396303"} Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.683175 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"283844d0-a97d-4bc0-b297-5e078862a5b9","Type":"ContainerStarted","Data":"adc3430009e01afa7c434ffb09e6b386e234204eec60e511d2ac9e94898a4d8e"} Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.683548 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"283844d0-a97d-4bc0-b297-5e078862a5b9","Type":"ContainerStarted","Data":"c339fbf65b16792278de20d0a4941df7551091d6925c6ec042a7250ed259b2d8"} Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.683572 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"283844d0-a97d-4bc0-b297-5e078862a5b9","Type":"ContainerStarted","Data":"19ca85a155bbba598b5be9d5345750517ffa3c8817cbcc12411fa43a5d75e018"} Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.698196 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.698174958 podStartE2EDuration="2.698174958s" podCreationTimestamp="2025-10-02 04:15:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 04:15:29.69786642 +0000 UTC m=+9266.864610460" watchObservedRunningTime="2025-10-02 04:15:29.698174958 +0000 UTC m=+9266.864919008" Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.723751 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.724919 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.7249060419999998 podStartE2EDuration="2.724906042s" podCreationTimestamp="2025-10-02 04:15:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 04:15:29.720544229 +0000 UTC m=+9266.887288269" watchObservedRunningTime="2025-10-02 04:15:29.724906042 +0000 UTC m=+9266.891650082" Oct 02 04:15:29 crc kubenswrapper[4775]: W1002 04:15:29.728642 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1dbfa9c9_66fa_48ad_956f_52297602546e.slice/crio-519a178e5f8604634ac75394f94c34d3176773c58bd56e543cbcf20282d7ece7 WatchSource:0}: Error finding container 519a178e5f8604634ac75394f94c34d3176773c58bd56e543cbcf20282d7ece7: Status 404 returned error can't find the container with id 519a178e5f8604634ac75394f94c34d3176773c58bd56e543cbcf20282d7ece7 Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.792807 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e865292-9dda-49fc-acdf-3afc04304aac" path="/var/lib/kubelet/pods/4e865292-9dda-49fc-acdf-3afc04304aac/volumes" Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.793757 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8986cac6-b070-4ea8-a961-24481fbb7bdd" path="/var/lib/kubelet/pods/8986cac6-b070-4ea8-a961-24481fbb7bdd/volumes" Oct 02 04:15:29 crc kubenswrapper[4775]: I1002 04:15:29.794772 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="def5a435-381a-4636-932e-8e4881989de3" path="/var/lib/kubelet/pods/def5a435-381a-4636-932e-8e4881989de3/volumes" Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.701154 4775 generic.go:334] "Generic (PLEG): container finished" podID="02de0e68-b4be-42b1-80eb-b4cb50dbd456" containerID="2d5246027af2edbc8f1f0bcfb9f5c32756879a301b041860ced369a8eae8dc35" exitCode=0 Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.701405 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02de0e68-b4be-42b1-80eb-b4cb50dbd456","Type":"ContainerDied","Data":"2d5246027af2edbc8f1f0bcfb9f5c32756879a301b041860ced369a8eae8dc35"} Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.702379 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"02de0e68-b4be-42b1-80eb-b4cb50dbd456","Type":"ContainerDied","Data":"96ad8cd1898f5a30da72eaf36eeb585010c87fc803f6f93c740b42da44ecd9e1"} Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.702397 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96ad8cd1898f5a30da72eaf36eeb585010c87fc803f6f93c740b42da44ecd9e1" Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.708083 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1dbfa9c9-66fa-48ad-956f-52297602546e","Type":"ContainerStarted","Data":"ac5bd9b0237a17af7d77f108b3400ef84bfc77c44b4bd831e9d687e8f4959696"} Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.708113 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1dbfa9c9-66fa-48ad-956f-52297602546e","Type":"ContainerStarted","Data":"519a178e5f8604634ac75394f94c34d3176773c58bd56e543cbcf20282d7ece7"} Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.708381 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.744200 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.744173437 podStartE2EDuration="2.744173437s" podCreationTimestamp="2025-10-02 04:15:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 04:15:30.726612131 +0000 UTC m=+9267.893356191" watchObservedRunningTime="2025-10-02 04:15:30.744173437 +0000 UTC m=+9267.910917487" Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.756869 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.878846 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02de0e68-b4be-42b1-80eb-b4cb50dbd456-config-data\") pod \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\" (UID: \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\") " Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.878976 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnwcp\" (UniqueName: \"kubernetes.io/projected/02de0e68-b4be-42b1-80eb-b4cb50dbd456-kube-api-access-bnwcp\") pod \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\" (UID: \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\") " Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.879075 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02de0e68-b4be-42b1-80eb-b4cb50dbd456-combined-ca-bundle\") pod \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\" (UID: \"02de0e68-b4be-42b1-80eb-b4cb50dbd456\") " Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.886647 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02de0e68-b4be-42b1-80eb-b4cb50dbd456-kube-api-access-bnwcp" (OuterVolumeSpecName: "kube-api-access-bnwcp") pod "02de0e68-b4be-42b1-80eb-b4cb50dbd456" (UID: "02de0e68-b4be-42b1-80eb-b4cb50dbd456"). InnerVolumeSpecName "kube-api-access-bnwcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.922385 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02de0e68-b4be-42b1-80eb-b4cb50dbd456-config-data" (OuterVolumeSpecName: "config-data") pod "02de0e68-b4be-42b1-80eb-b4cb50dbd456" (UID: "02de0e68-b4be-42b1-80eb-b4cb50dbd456"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.943339 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02de0e68-b4be-42b1-80eb-b4cb50dbd456-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02de0e68-b4be-42b1-80eb-b4cb50dbd456" (UID: "02de0e68-b4be-42b1-80eb-b4cb50dbd456"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.982712 4775 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02de0e68-b4be-42b1-80eb-b4cb50dbd456-config-data\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.982744 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnwcp\" (UniqueName: \"kubernetes.io/projected/02de0e68-b4be-42b1-80eb-b4cb50dbd456-kube-api-access-bnwcp\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:30 crc kubenswrapper[4775]: I1002 04:15:30.982756 4775 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02de0e68-b4be-42b1-80eb-b4cb50dbd456-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 04:15:31 crc kubenswrapper[4775]: I1002 04:15:31.719584 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 04:15:31 crc kubenswrapper[4775]: I1002 04:15:31.759188 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 04:15:31 crc kubenswrapper[4775]: I1002 04:15:31.779094 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 04:15:31 crc kubenswrapper[4775]: I1002 04:15:31.795161 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 04:15:31 crc kubenswrapper[4775]: E1002 04:15:31.795908 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02de0e68-b4be-42b1-80eb-b4cb50dbd456" containerName="nova-scheduler-scheduler" Oct 02 04:15:31 crc kubenswrapper[4775]: I1002 04:15:31.795938 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="02de0e68-b4be-42b1-80eb-b4cb50dbd456" containerName="nova-scheduler-scheduler" Oct 02 04:15:31 crc kubenswrapper[4775]: I1002 04:15:31.796387 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="02de0e68-b4be-42b1-80eb-b4cb50dbd456" containerName="nova-scheduler-scheduler" Oct 02 04:15:31 crc kubenswrapper[4775]: I1002 04:15:31.797656 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 04:15:31 crc kubenswrapper[4775]: I1002 04:15:31.800408 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 02 04:15:31 crc kubenswrapper[4775]: I1002 04:15:31.808337 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 04:15:31 crc kubenswrapper[4775]: I1002 04:15:31.904326 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880173a7-686a-4c2f-8d7a-1b929fcd7c03-config-data\") pod \"nova-scheduler-0\" (UID: \"880173a7-686a-4c2f-8d7a-1b929fcd7c03\") " pod="openstack/nova-scheduler-0" Oct 02 04:15:31 crc kubenswrapper[4775]: I1002 04:15:31.904463 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9bzk\" (UniqueName: \"kubernetes.io/projected/880173a7-686a-4c2f-8d7a-1b929fcd7c03-kube-api-access-w9bzk\") pod \"nova-scheduler-0\" (UID: \"880173a7-686a-4c2f-8d7a-1b929fcd7c03\") " pod="openstack/nova-scheduler-0" Oct 02 04:15:31 crc kubenswrapper[4775]: I1002 04:15:31.904489 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880173a7-686a-4c2f-8d7a-1b929fcd7c03-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"880173a7-686a-4c2f-8d7a-1b929fcd7c03\") " pod="openstack/nova-scheduler-0" Oct 02 04:15:32 crc kubenswrapper[4775]: I1002 04:15:32.006361 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880173a7-686a-4c2f-8d7a-1b929fcd7c03-config-data\") pod \"nova-scheduler-0\" (UID: \"880173a7-686a-4c2f-8d7a-1b929fcd7c03\") " pod="openstack/nova-scheduler-0" Oct 02 04:15:32 crc kubenswrapper[4775]: I1002 04:15:32.006805 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9bzk\" (UniqueName: \"kubernetes.io/projected/880173a7-686a-4c2f-8d7a-1b929fcd7c03-kube-api-access-w9bzk\") pod \"nova-scheduler-0\" (UID: \"880173a7-686a-4c2f-8d7a-1b929fcd7c03\") " pod="openstack/nova-scheduler-0" Oct 02 04:15:32 crc kubenswrapper[4775]: I1002 04:15:32.006831 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880173a7-686a-4c2f-8d7a-1b929fcd7c03-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"880173a7-686a-4c2f-8d7a-1b929fcd7c03\") " pod="openstack/nova-scheduler-0" Oct 02 04:15:32 crc kubenswrapper[4775]: I1002 04:15:32.012897 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/880173a7-686a-4c2f-8d7a-1b929fcd7c03-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"880173a7-686a-4c2f-8d7a-1b929fcd7c03\") " pod="openstack/nova-scheduler-0" Oct 02 04:15:32 crc kubenswrapper[4775]: I1002 04:15:32.027524 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9bzk\" (UniqueName: \"kubernetes.io/projected/880173a7-686a-4c2f-8d7a-1b929fcd7c03-kube-api-access-w9bzk\") pod \"nova-scheduler-0\" (UID: \"880173a7-686a-4c2f-8d7a-1b929fcd7c03\") " pod="openstack/nova-scheduler-0" Oct 02 04:15:32 crc kubenswrapper[4775]: I1002 04:15:32.028441 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/880173a7-686a-4c2f-8d7a-1b929fcd7c03-config-data\") pod \"nova-scheduler-0\" (UID: \"880173a7-686a-4c2f-8d7a-1b929fcd7c03\") " pod="openstack/nova-scheduler-0" Oct 02 04:15:32 crc kubenswrapper[4775]: I1002 04:15:32.124676 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 02 04:15:32 crc kubenswrapper[4775]: I1002 04:15:32.674836 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 02 04:15:32 crc kubenswrapper[4775]: W1002 04:15:32.683638 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod880173a7_686a_4c2f_8d7a_1b929fcd7c03.slice/crio-55982db0281130b8575807fd1ecb887ab175f3a7039e5a5019526ba237a85802 WatchSource:0}: Error finding container 55982db0281130b8575807fd1ecb887ab175f3a7039e5a5019526ba237a85802: Status 404 returned error can't find the container with id 55982db0281130b8575807fd1ecb887ab175f3a7039e5a5019526ba237a85802 Oct 02 04:15:32 crc kubenswrapper[4775]: I1002 04:15:32.732404 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"880173a7-686a-4c2f-8d7a-1b929fcd7c03","Type":"ContainerStarted","Data":"55982db0281130b8575807fd1ecb887ab175f3a7039e5a5019526ba237a85802"} Oct 02 04:15:33 crc kubenswrapper[4775]: I1002 04:15:33.160611 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 04:15:33 crc kubenswrapper[4775]: I1002 04:15:33.161042 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 02 04:15:33 crc kubenswrapper[4775]: I1002 04:15:33.505533 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 02 04:15:33 crc kubenswrapper[4775]: I1002 04:15:33.747043 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"880173a7-686a-4c2f-8d7a-1b929fcd7c03","Type":"ContainerStarted","Data":"10867de1b29ae08906b833ca9c6cefb36595ead928bfef1af5b8020f49673ba1"} Oct 02 04:15:33 crc kubenswrapper[4775]: I1002 04:15:33.771282 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.771260921 podStartE2EDuration="2.771260921s" podCreationTimestamp="2025-10-02 04:15:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-02 04:15:33.764608558 +0000 UTC m=+9270.931352628" watchObservedRunningTime="2025-10-02 04:15:33.771260921 +0000 UTC m=+9270.938004971" Oct 02 04:15:33 crc kubenswrapper[4775]: I1002 04:15:33.788149 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02de0e68-b4be-42b1-80eb-b4cb50dbd456" path="/var/lib/kubelet/pods/02de0e68-b4be-42b1-80eb-b4cb50dbd456/volumes" Oct 02 04:15:34 crc kubenswrapper[4775]: I1002 04:15:34.279880 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 02 04:15:34 crc kubenswrapper[4775]: I1002 04:15:34.767477 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:15:34 crc kubenswrapper[4775]: E1002 04:15:34.767813 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:15:37 crc kubenswrapper[4775]: I1002 04:15:37.124894 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 02 04:15:38 crc kubenswrapper[4775]: I1002 04:15:38.160503 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 04:15:38 crc kubenswrapper[4775]: I1002 04:15:38.160591 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 02 04:15:38 crc kubenswrapper[4775]: I1002 04:15:38.175531 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 04:15:38 crc kubenswrapper[4775]: I1002 04:15:38.175598 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 02 04:15:39 crc kubenswrapper[4775]: I1002 04:15:39.324284 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="20d47d00-0c95-4858-963c-ac98a3b56306" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 04:15:39 crc kubenswrapper[4775]: I1002 04:15:39.324405 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="283844d0-a97d-4bc0-b297-5e078862a5b9" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.194:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 04:15:39 crc kubenswrapper[4775]: I1002 04:15:39.324417 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="283844d0-a97d-4bc0-b297-5e078862a5b9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.194:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 04:15:39 crc kubenswrapper[4775]: I1002 04:15:39.324843 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="20d47d00-0c95-4858-963c-ac98a3b56306" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.195:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 02 04:15:42 crc kubenswrapper[4775]: I1002 04:15:42.125200 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 02 04:15:42 crc kubenswrapper[4775]: I1002 04:15:42.166394 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 02 04:15:42 crc kubenswrapper[4775]: I1002 04:15:42.949912 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 02 04:15:45 crc kubenswrapper[4775]: I1002 04:15:45.765313 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:15:45 crc kubenswrapper[4775]: E1002 04:15:45.765877 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:15:48 crc kubenswrapper[4775]: I1002 04:15:48.163397 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 04:15:48 crc kubenswrapper[4775]: I1002 04:15:48.164048 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 02 04:15:48 crc kubenswrapper[4775]: I1002 04:15:48.168013 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 04:15:48 crc kubenswrapper[4775]: I1002 04:15:48.169062 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 02 04:15:48 crc kubenswrapper[4775]: I1002 04:15:48.179685 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 04:15:48 crc kubenswrapper[4775]: I1002 04:15:48.180277 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 04:15:48 crc kubenswrapper[4775]: I1002 04:15:48.181448 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 02 04:15:48 crc kubenswrapper[4775]: I1002 04:15:48.184799 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 04:15:48 crc kubenswrapper[4775]: I1002 04:15:48.979822 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 02 04:15:48 crc kubenswrapper[4775]: I1002 04:15:48.986493 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.224703 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv"] Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.228068 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.229797 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.230875 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.231049 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.231141 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.239393 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.239862 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.240937 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-68zrg" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.241281 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv"] Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.339351 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.339402 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.339450 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.339474 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.339587 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.339791 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.340021 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.340053 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2mfq\" (UniqueName: \"kubernetes.io/projected/b5d0f695-e66e-464e-bf9b-301a66a74f1b-kube-api-access-b2mfq\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.340228 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.340288 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.340418 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.442873 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.442942 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2mfq\" (UniqueName: \"kubernetes.io/projected/b5d0f695-e66e-464e-bf9b-301a66a74f1b-kube-api-access-b2mfq\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.443024 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.443050 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.443105 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.443150 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.443189 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.443238 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.443268 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.443296 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.443349 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.444104 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.444452 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.451756 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.451768 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.452760 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.453491 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.454684 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.456289 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.456943 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.460946 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.462771 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2mfq\" (UniqueName: \"kubernetes.io/projected/b5d0f695-e66e-464e-bf9b-301a66a74f1b-kube-api-access-b2mfq\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:50 crc kubenswrapper[4775]: I1002 04:15:50.554939 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:15:51 crc kubenswrapper[4775]: I1002 04:15:51.181628 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv"] Oct 02 04:15:52 crc kubenswrapper[4775]: I1002 04:15:52.028328 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" event={"ID":"b5d0f695-e66e-464e-bf9b-301a66a74f1b","Type":"ContainerStarted","Data":"2c67ddb141f99b2433b909453a75d977e0c5a250424f3b702c349ac3a831ea2c"} Oct 02 04:15:53 crc kubenswrapper[4775]: I1002 04:15:53.042650 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" event={"ID":"b5d0f695-e66e-464e-bf9b-301a66a74f1b","Type":"ContainerStarted","Data":"9b358381c9f09ca411bbcf33af99c6782e31ce86d970ce90b1c95573393a2d2e"} Oct 02 04:15:53 crc kubenswrapper[4775]: I1002 04:15:53.080319 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" podStartSLOduration=2.250483545 podStartE2EDuration="3.080288828s" podCreationTimestamp="2025-10-02 04:15:50 +0000 UTC" firstStartedPulling="2025-10-02 04:15:51.180430841 +0000 UTC m=+9288.347174901" lastFinishedPulling="2025-10-02 04:15:52.010236134 +0000 UTC m=+9289.176980184" observedRunningTime="2025-10-02 04:15:53.064051996 +0000 UTC m=+9290.230796046" watchObservedRunningTime="2025-10-02 04:15:53.080288828 +0000 UTC m=+9290.247032908" Oct 02 04:15:59 crc kubenswrapper[4775]: I1002 04:15:59.765563 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:15:59 crc kubenswrapper[4775]: E1002 04:15:59.766451 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:16:10 crc kubenswrapper[4775]: I1002 04:16:10.766892 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:16:10 crc kubenswrapper[4775]: E1002 04:16:10.767914 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:16:18 crc kubenswrapper[4775]: I1002 04:16:18.733528 4775 scope.go:117] "RemoveContainer" containerID="2d5246027af2edbc8f1f0bcfb9f5c32756879a301b041860ced369a8eae8dc35" Oct 02 04:16:18 crc kubenswrapper[4775]: I1002 04:16:18.800484 4775 scope.go:117] "RemoveContainer" containerID="6e4b2842c6945dc728e9ab20f6eae59f1424306f45f162bb0205a42c38e05645" Oct 02 04:16:22 crc kubenswrapper[4775]: I1002 04:16:22.766386 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:16:22 crc kubenswrapper[4775]: E1002 04:16:22.767662 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:16:33 crc kubenswrapper[4775]: I1002 04:16:33.773333 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:16:33 crc kubenswrapper[4775]: E1002 04:16:33.774417 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:16:48 crc kubenswrapper[4775]: I1002 04:16:48.768164 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:16:48 crc kubenswrapper[4775]: E1002 04:16:48.769572 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:17:02 crc kubenswrapper[4775]: I1002 04:17:02.765865 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:17:02 crc kubenswrapper[4775]: E1002 04:17:02.767664 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:17:15 crc kubenswrapper[4775]: I1002 04:17:15.764809 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:17:15 crc kubenswrapper[4775]: E1002 04:17:15.766044 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:17:28 crc kubenswrapper[4775]: I1002 04:17:28.766042 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:17:28 crc kubenswrapper[4775]: E1002 04:17:28.767500 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:17:42 crc kubenswrapper[4775]: I1002 04:17:42.774863 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:17:42 crc kubenswrapper[4775]: E1002 04:17:42.785208 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:17:45 crc kubenswrapper[4775]: I1002 04:17:45.232550 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d9p8j"] Oct 02 04:17:45 crc kubenswrapper[4775]: I1002 04:17:45.237230 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:17:45 crc kubenswrapper[4775]: I1002 04:17:45.252934 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d9p8j"] Oct 02 04:17:45 crc kubenswrapper[4775]: I1002 04:17:45.350680 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5700e992-86c2-43bf-8cbb-5c8285511f4c-catalog-content\") pod \"certified-operators-d9p8j\" (UID: \"5700e992-86c2-43bf-8cbb-5c8285511f4c\") " pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:17:45 crc kubenswrapper[4775]: I1002 04:17:45.350786 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5700e992-86c2-43bf-8cbb-5c8285511f4c-utilities\") pod \"certified-operators-d9p8j\" (UID: \"5700e992-86c2-43bf-8cbb-5c8285511f4c\") " pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:17:45 crc kubenswrapper[4775]: I1002 04:17:45.350829 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2mvb\" (UniqueName: \"kubernetes.io/projected/5700e992-86c2-43bf-8cbb-5c8285511f4c-kube-api-access-c2mvb\") pod \"certified-operators-d9p8j\" (UID: \"5700e992-86c2-43bf-8cbb-5c8285511f4c\") " pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:17:45 crc kubenswrapper[4775]: I1002 04:17:45.453828 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5700e992-86c2-43bf-8cbb-5c8285511f4c-catalog-content\") pod \"certified-operators-d9p8j\" (UID: \"5700e992-86c2-43bf-8cbb-5c8285511f4c\") " pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:17:45 crc kubenswrapper[4775]: I1002 04:17:45.453934 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5700e992-86c2-43bf-8cbb-5c8285511f4c-utilities\") pod \"certified-operators-d9p8j\" (UID: \"5700e992-86c2-43bf-8cbb-5c8285511f4c\") " pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:17:45 crc kubenswrapper[4775]: I1002 04:17:45.454026 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2mvb\" (UniqueName: \"kubernetes.io/projected/5700e992-86c2-43bf-8cbb-5c8285511f4c-kube-api-access-c2mvb\") pod \"certified-operators-d9p8j\" (UID: \"5700e992-86c2-43bf-8cbb-5c8285511f4c\") " pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:17:45 crc kubenswrapper[4775]: I1002 04:17:45.455131 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5700e992-86c2-43bf-8cbb-5c8285511f4c-catalog-content\") pod \"certified-operators-d9p8j\" (UID: \"5700e992-86c2-43bf-8cbb-5c8285511f4c\") " pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:17:45 crc kubenswrapper[4775]: I1002 04:17:45.455234 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5700e992-86c2-43bf-8cbb-5c8285511f4c-utilities\") pod \"certified-operators-d9p8j\" (UID: \"5700e992-86c2-43bf-8cbb-5c8285511f4c\") " pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:17:45 crc kubenswrapper[4775]: I1002 04:17:45.498635 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2mvb\" (UniqueName: \"kubernetes.io/projected/5700e992-86c2-43bf-8cbb-5c8285511f4c-kube-api-access-c2mvb\") pod \"certified-operators-d9p8j\" (UID: \"5700e992-86c2-43bf-8cbb-5c8285511f4c\") " pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:17:45 crc kubenswrapper[4775]: I1002 04:17:45.577347 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.162617 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d9p8j"] Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.587706 4775 generic.go:334] "Generic (PLEG): container finished" podID="5700e992-86c2-43bf-8cbb-5c8285511f4c" containerID="31ed8c28adb9a8c9ef8714d19431046a3baaf5a2432141ea880b104fefad0867" exitCode=0 Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.587786 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d9p8j" event={"ID":"5700e992-86c2-43bf-8cbb-5c8285511f4c","Type":"ContainerDied","Data":"31ed8c28adb9a8c9ef8714d19431046a3baaf5a2432141ea880b104fefad0867"} Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.587841 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d9p8j" event={"ID":"5700e992-86c2-43bf-8cbb-5c8285511f4c","Type":"ContainerStarted","Data":"8085a84978e008d79a467455ab4a4c609768502289df6eb97a9d2869886cacbb"} Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.641911 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-djv7s"] Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.645507 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.688585 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-djv7s"] Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.792632 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec363e50-fb34-4304-a26e-72e2d41c84b0-utilities\") pod \"redhat-operators-djv7s\" (UID: \"ec363e50-fb34-4304-a26e-72e2d41c84b0\") " pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.792786 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec363e50-fb34-4304-a26e-72e2d41c84b0-catalog-content\") pod \"redhat-operators-djv7s\" (UID: \"ec363e50-fb34-4304-a26e-72e2d41c84b0\") " pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.792814 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbnj8\" (UniqueName: \"kubernetes.io/projected/ec363e50-fb34-4304-a26e-72e2d41c84b0-kube-api-access-jbnj8\") pod \"redhat-operators-djv7s\" (UID: \"ec363e50-fb34-4304-a26e-72e2d41c84b0\") " pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.894451 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec363e50-fb34-4304-a26e-72e2d41c84b0-catalog-content\") pod \"redhat-operators-djv7s\" (UID: \"ec363e50-fb34-4304-a26e-72e2d41c84b0\") " pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.894759 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbnj8\" (UniqueName: \"kubernetes.io/projected/ec363e50-fb34-4304-a26e-72e2d41c84b0-kube-api-access-jbnj8\") pod \"redhat-operators-djv7s\" (UID: \"ec363e50-fb34-4304-a26e-72e2d41c84b0\") " pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.895428 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec363e50-fb34-4304-a26e-72e2d41c84b0-utilities\") pod \"redhat-operators-djv7s\" (UID: \"ec363e50-fb34-4304-a26e-72e2d41c84b0\") " pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.895710 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec363e50-fb34-4304-a26e-72e2d41c84b0-catalog-content\") pod \"redhat-operators-djv7s\" (UID: \"ec363e50-fb34-4304-a26e-72e2d41c84b0\") " pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.895865 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec363e50-fb34-4304-a26e-72e2d41c84b0-utilities\") pod \"redhat-operators-djv7s\" (UID: \"ec363e50-fb34-4304-a26e-72e2d41c84b0\") " pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.918775 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbnj8\" (UniqueName: \"kubernetes.io/projected/ec363e50-fb34-4304-a26e-72e2d41c84b0-kube-api-access-jbnj8\") pod \"redhat-operators-djv7s\" (UID: \"ec363e50-fb34-4304-a26e-72e2d41c84b0\") " pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:17:46 crc kubenswrapper[4775]: I1002 04:17:46.980663 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:17:47 crc kubenswrapper[4775]: W1002 04:17:47.504103 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podec363e50_fb34_4304_a26e_72e2d41c84b0.slice/crio-b3c35decf150dafcf29e133cef08696e7575349c88096d4136cfd9bcf6a4d40e WatchSource:0}: Error finding container b3c35decf150dafcf29e133cef08696e7575349c88096d4136cfd9bcf6a4d40e: Status 404 returned error can't find the container with id b3c35decf150dafcf29e133cef08696e7575349c88096d4136cfd9bcf6a4d40e Oct 02 04:17:47 crc kubenswrapper[4775]: I1002 04:17:47.507348 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-djv7s"] Oct 02 04:17:47 crc kubenswrapper[4775]: I1002 04:17:47.598379 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d9p8j" event={"ID":"5700e992-86c2-43bf-8cbb-5c8285511f4c","Type":"ContainerStarted","Data":"0d2de1ac149c2160b2abfd358743f8a02703a20d5c2d9704f24b01c60d6da8eb"} Oct 02 04:17:47 crc kubenswrapper[4775]: I1002 04:17:47.600019 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djv7s" event={"ID":"ec363e50-fb34-4304-a26e-72e2d41c84b0","Type":"ContainerStarted","Data":"b3c35decf150dafcf29e133cef08696e7575349c88096d4136cfd9bcf6a4d40e"} Oct 02 04:17:48 crc kubenswrapper[4775]: I1002 04:17:48.610882 4775 generic.go:334] "Generic (PLEG): container finished" podID="ec363e50-fb34-4304-a26e-72e2d41c84b0" containerID="3c126df49851dea425ccb541790944b93c996d0f77e8edd21f3cc435117dd170" exitCode=0 Oct 02 04:17:48 crc kubenswrapper[4775]: I1002 04:17:48.610944 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djv7s" event={"ID":"ec363e50-fb34-4304-a26e-72e2d41c84b0","Type":"ContainerDied","Data":"3c126df49851dea425ccb541790944b93c996d0f77e8edd21f3cc435117dd170"} Oct 02 04:17:49 crc kubenswrapper[4775]: I1002 04:17:49.628381 4775 generic.go:334] "Generic (PLEG): container finished" podID="5700e992-86c2-43bf-8cbb-5c8285511f4c" containerID="0d2de1ac149c2160b2abfd358743f8a02703a20d5c2d9704f24b01c60d6da8eb" exitCode=0 Oct 02 04:17:49 crc kubenswrapper[4775]: I1002 04:17:49.628727 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d9p8j" event={"ID":"5700e992-86c2-43bf-8cbb-5c8285511f4c","Type":"ContainerDied","Data":"0d2de1ac149c2160b2abfd358743f8a02703a20d5c2d9704f24b01c60d6da8eb"} Oct 02 04:17:50 crc kubenswrapper[4775]: I1002 04:17:50.645898 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d9p8j" event={"ID":"5700e992-86c2-43bf-8cbb-5c8285511f4c","Type":"ContainerStarted","Data":"318452b80f5ab7e32306c9e8036a1ed13b5ce37adce97b1da9a5c36b1d7c87d1"} Oct 02 04:17:50 crc kubenswrapper[4775]: I1002 04:17:50.650870 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djv7s" event={"ID":"ec363e50-fb34-4304-a26e-72e2d41c84b0","Type":"ContainerStarted","Data":"90f390eee7663f7d922beeb1ad82261c623b08d8675519f7781b87762141d4b5"} Oct 02 04:17:50 crc kubenswrapper[4775]: I1002 04:17:50.675137 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d9p8j" podStartSLOduration=1.9582369979999998 podStartE2EDuration="5.675115849s" podCreationTimestamp="2025-10-02 04:17:45 +0000 UTC" firstStartedPulling="2025-10-02 04:17:46.590778624 +0000 UTC m=+9403.757522664" lastFinishedPulling="2025-10-02 04:17:50.307657465 +0000 UTC m=+9407.474401515" observedRunningTime="2025-10-02 04:17:50.672487531 +0000 UTC m=+9407.839231581" watchObservedRunningTime="2025-10-02 04:17:50.675115849 +0000 UTC m=+9407.841859899" Oct 02 04:17:53 crc kubenswrapper[4775]: I1002 04:17:53.698809 4775 generic.go:334] "Generic (PLEG): container finished" podID="ec363e50-fb34-4304-a26e-72e2d41c84b0" containerID="90f390eee7663f7d922beeb1ad82261c623b08d8675519f7781b87762141d4b5" exitCode=0 Oct 02 04:17:53 crc kubenswrapper[4775]: I1002 04:17:53.698928 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djv7s" event={"ID":"ec363e50-fb34-4304-a26e-72e2d41c84b0","Type":"ContainerDied","Data":"90f390eee7663f7d922beeb1ad82261c623b08d8675519f7781b87762141d4b5"} Oct 02 04:17:54 crc kubenswrapper[4775]: I1002 04:17:54.712523 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djv7s" event={"ID":"ec363e50-fb34-4304-a26e-72e2d41c84b0","Type":"ContainerStarted","Data":"a0c9754566e7ee66f8d0206434bc448cb5bd940cd612687e0447acccb8dd4bf9"} Oct 02 04:17:54 crc kubenswrapper[4775]: I1002 04:17:54.743164 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-djv7s" podStartSLOduration=3.130992692 podStartE2EDuration="8.743147131s" podCreationTimestamp="2025-10-02 04:17:46 +0000 UTC" firstStartedPulling="2025-10-02 04:17:48.614572229 +0000 UTC m=+9405.781316269" lastFinishedPulling="2025-10-02 04:17:54.226726628 +0000 UTC m=+9411.393470708" observedRunningTime="2025-10-02 04:17:54.734435835 +0000 UTC m=+9411.901179945" watchObservedRunningTime="2025-10-02 04:17:54.743147131 +0000 UTC m=+9411.909891171" Oct 02 04:17:55 crc kubenswrapper[4775]: I1002 04:17:55.578408 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:17:55 crc kubenswrapper[4775]: I1002 04:17:55.578475 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:17:56 crc kubenswrapper[4775]: I1002 04:17:56.670632 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-d9p8j" podUID="5700e992-86c2-43bf-8cbb-5c8285511f4c" containerName="registry-server" probeResult="failure" output=< Oct 02 04:17:56 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 04:17:56 crc kubenswrapper[4775]: > Oct 02 04:17:56 crc kubenswrapper[4775]: I1002 04:17:56.765887 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:17:56 crc kubenswrapper[4775]: E1002 04:17:56.766235 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:17:56 crc kubenswrapper[4775]: I1002 04:17:56.981406 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:17:56 crc kubenswrapper[4775]: I1002 04:17:56.981667 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:17:58 crc kubenswrapper[4775]: I1002 04:17:58.066311 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-djv7s" podUID="ec363e50-fb34-4304-a26e-72e2d41c84b0" containerName="registry-server" probeResult="failure" output=< Oct 02 04:17:58 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 04:17:58 crc kubenswrapper[4775]: > Oct 02 04:18:05 crc kubenswrapper[4775]: I1002 04:18:05.659045 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:18:05 crc kubenswrapper[4775]: I1002 04:18:05.747933 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:18:06 crc kubenswrapper[4775]: I1002 04:18:06.795633 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d9p8j"] Oct 02 04:18:06 crc kubenswrapper[4775]: I1002 04:18:06.884442 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d9p8j" podUID="5700e992-86c2-43bf-8cbb-5c8285511f4c" containerName="registry-server" containerID="cri-o://318452b80f5ab7e32306c9e8036a1ed13b5ce37adce97b1da9a5c36b1d7c87d1" gracePeriod=2 Oct 02 04:18:07 crc kubenswrapper[4775]: I1002 04:18:07.767007 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:18:07 crc kubenswrapper[4775]: I1002 04:18:07.902547 4775 generic.go:334] "Generic (PLEG): container finished" podID="5700e992-86c2-43bf-8cbb-5c8285511f4c" containerID="318452b80f5ab7e32306c9e8036a1ed13b5ce37adce97b1da9a5c36b1d7c87d1" exitCode=0 Oct 02 04:18:07 crc kubenswrapper[4775]: I1002 04:18:07.902630 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d9p8j" event={"ID":"5700e992-86c2-43bf-8cbb-5c8285511f4c","Type":"ContainerDied","Data":"318452b80f5ab7e32306c9e8036a1ed13b5ce37adce97b1da9a5c36b1d7c87d1"} Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.065441 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-djv7s" podUID="ec363e50-fb34-4304-a26e-72e2d41c84b0" containerName="registry-server" probeResult="failure" output=< Oct 02 04:18:08 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 04:18:08 crc kubenswrapper[4775]: > Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.446685 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.551761 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5700e992-86c2-43bf-8cbb-5c8285511f4c-utilities\") pod \"5700e992-86c2-43bf-8cbb-5c8285511f4c\" (UID: \"5700e992-86c2-43bf-8cbb-5c8285511f4c\") " Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.552009 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5700e992-86c2-43bf-8cbb-5c8285511f4c-catalog-content\") pod \"5700e992-86c2-43bf-8cbb-5c8285511f4c\" (UID: \"5700e992-86c2-43bf-8cbb-5c8285511f4c\") " Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.552040 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2mvb\" (UniqueName: \"kubernetes.io/projected/5700e992-86c2-43bf-8cbb-5c8285511f4c-kube-api-access-c2mvb\") pod \"5700e992-86c2-43bf-8cbb-5c8285511f4c\" (UID: \"5700e992-86c2-43bf-8cbb-5c8285511f4c\") " Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.558429 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5700e992-86c2-43bf-8cbb-5c8285511f4c-kube-api-access-c2mvb" (OuterVolumeSpecName: "kube-api-access-c2mvb") pod "5700e992-86c2-43bf-8cbb-5c8285511f4c" (UID: "5700e992-86c2-43bf-8cbb-5c8285511f4c"). InnerVolumeSpecName "kube-api-access-c2mvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.559178 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5700e992-86c2-43bf-8cbb-5c8285511f4c-utilities" (OuterVolumeSpecName: "utilities") pod "5700e992-86c2-43bf-8cbb-5c8285511f4c" (UID: "5700e992-86c2-43bf-8cbb-5c8285511f4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.609208 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5700e992-86c2-43bf-8cbb-5c8285511f4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5700e992-86c2-43bf-8cbb-5c8285511f4c" (UID: "5700e992-86c2-43bf-8cbb-5c8285511f4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.654421 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5700e992-86c2-43bf-8cbb-5c8285511f4c-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.654766 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5700e992-86c2-43bf-8cbb-5c8285511f4c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.654852 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2mvb\" (UniqueName: \"kubernetes.io/projected/5700e992-86c2-43bf-8cbb-5c8285511f4c-kube-api-access-c2mvb\") on node \"crc\" DevicePath \"\"" Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.919433 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"f513efcfa13070dc5964ae5079d55032a787ae801794c0913b3a3fd25e50f61a"} Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.923687 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d9p8j" event={"ID":"5700e992-86c2-43bf-8cbb-5c8285511f4c","Type":"ContainerDied","Data":"8085a84978e008d79a467455ab4a4c609768502289df6eb97a9d2869886cacbb"} Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.923741 4775 scope.go:117] "RemoveContainer" containerID="318452b80f5ab7e32306c9e8036a1ed13b5ce37adce97b1da9a5c36b1d7c87d1" Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.923804 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d9p8j" Oct 02 04:18:08 crc kubenswrapper[4775]: I1002 04:18:08.983379 4775 scope.go:117] "RemoveContainer" containerID="0d2de1ac149c2160b2abfd358743f8a02703a20d5c2d9704f24b01c60d6da8eb" Oct 02 04:18:09 crc kubenswrapper[4775]: I1002 04:18:09.016173 4775 scope.go:117] "RemoveContainer" containerID="31ed8c28adb9a8c9ef8714d19431046a3baaf5a2432141ea880b104fefad0867" Oct 02 04:18:09 crc kubenswrapper[4775]: I1002 04:18:09.017041 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d9p8j"] Oct 02 04:18:09 crc kubenswrapper[4775]: I1002 04:18:09.031447 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d9p8j"] Oct 02 04:18:09 crc kubenswrapper[4775]: I1002 04:18:09.780581 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5700e992-86c2-43bf-8cbb-5c8285511f4c" path="/var/lib/kubelet/pods/5700e992-86c2-43bf-8cbb-5c8285511f4c/volumes" Oct 02 04:18:17 crc kubenswrapper[4775]: I1002 04:18:17.056699 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:18:17 crc kubenswrapper[4775]: I1002 04:18:17.132194 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:18:19 crc kubenswrapper[4775]: I1002 04:18:19.403593 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-djv7s"] Oct 02 04:18:19 crc kubenswrapper[4775]: I1002 04:18:19.404458 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-djv7s" podUID="ec363e50-fb34-4304-a26e-72e2d41c84b0" containerName="registry-server" containerID="cri-o://a0c9754566e7ee66f8d0206434bc448cb5bd940cd612687e0447acccb8dd4bf9" gracePeriod=2 Oct 02 04:18:19 crc kubenswrapper[4775]: I1002 04:18:19.975050 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.073393 4775 generic.go:334] "Generic (PLEG): container finished" podID="ec363e50-fb34-4304-a26e-72e2d41c84b0" containerID="a0c9754566e7ee66f8d0206434bc448cb5bd940cd612687e0447acccb8dd4bf9" exitCode=0 Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.073434 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djv7s" event={"ID":"ec363e50-fb34-4304-a26e-72e2d41c84b0","Type":"ContainerDied","Data":"a0c9754566e7ee66f8d0206434bc448cb5bd940cd612687e0447acccb8dd4bf9"} Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.073459 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-djv7s" event={"ID":"ec363e50-fb34-4304-a26e-72e2d41c84b0","Type":"ContainerDied","Data":"b3c35decf150dafcf29e133cef08696e7575349c88096d4136cfd9bcf6a4d40e"} Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.073477 4775 scope.go:117] "RemoveContainer" containerID="a0c9754566e7ee66f8d0206434bc448cb5bd940cd612687e0447acccb8dd4bf9" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.073531 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-djv7s" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.120286 4775 scope.go:117] "RemoveContainer" containerID="90f390eee7663f7d922beeb1ad82261c623b08d8675519f7781b87762141d4b5" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.127733 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbnj8\" (UniqueName: \"kubernetes.io/projected/ec363e50-fb34-4304-a26e-72e2d41c84b0-kube-api-access-jbnj8\") pod \"ec363e50-fb34-4304-a26e-72e2d41c84b0\" (UID: \"ec363e50-fb34-4304-a26e-72e2d41c84b0\") " Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.128281 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec363e50-fb34-4304-a26e-72e2d41c84b0-utilities\") pod \"ec363e50-fb34-4304-a26e-72e2d41c84b0\" (UID: \"ec363e50-fb34-4304-a26e-72e2d41c84b0\") " Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.128435 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec363e50-fb34-4304-a26e-72e2d41c84b0-catalog-content\") pod \"ec363e50-fb34-4304-a26e-72e2d41c84b0\" (UID: \"ec363e50-fb34-4304-a26e-72e2d41c84b0\") " Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.129080 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec363e50-fb34-4304-a26e-72e2d41c84b0-utilities" (OuterVolumeSpecName: "utilities") pod "ec363e50-fb34-4304-a26e-72e2d41c84b0" (UID: "ec363e50-fb34-4304-a26e-72e2d41c84b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.129803 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec363e50-fb34-4304-a26e-72e2d41c84b0-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.134259 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec363e50-fb34-4304-a26e-72e2d41c84b0-kube-api-access-jbnj8" (OuterVolumeSpecName: "kube-api-access-jbnj8") pod "ec363e50-fb34-4304-a26e-72e2d41c84b0" (UID: "ec363e50-fb34-4304-a26e-72e2d41c84b0"). InnerVolumeSpecName "kube-api-access-jbnj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.150122 4775 scope.go:117] "RemoveContainer" containerID="3c126df49851dea425ccb541790944b93c996d0f77e8edd21f3cc435117dd170" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.214354 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec363e50-fb34-4304-a26e-72e2d41c84b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec363e50-fb34-4304-a26e-72e2d41c84b0" (UID: "ec363e50-fb34-4304-a26e-72e2d41c84b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.230230 4775 scope.go:117] "RemoveContainer" containerID="a0c9754566e7ee66f8d0206434bc448cb5bd940cd612687e0447acccb8dd4bf9" Oct 02 04:18:20 crc kubenswrapper[4775]: E1002 04:18:20.230872 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0c9754566e7ee66f8d0206434bc448cb5bd940cd612687e0447acccb8dd4bf9\": container with ID starting with a0c9754566e7ee66f8d0206434bc448cb5bd940cd612687e0447acccb8dd4bf9 not found: ID does not exist" containerID="a0c9754566e7ee66f8d0206434bc448cb5bd940cd612687e0447acccb8dd4bf9" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.230928 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0c9754566e7ee66f8d0206434bc448cb5bd940cd612687e0447acccb8dd4bf9"} err="failed to get container status \"a0c9754566e7ee66f8d0206434bc448cb5bd940cd612687e0447acccb8dd4bf9\": rpc error: code = NotFound desc = could not find container \"a0c9754566e7ee66f8d0206434bc448cb5bd940cd612687e0447acccb8dd4bf9\": container with ID starting with a0c9754566e7ee66f8d0206434bc448cb5bd940cd612687e0447acccb8dd4bf9 not found: ID does not exist" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.230991 4775 scope.go:117] "RemoveContainer" containerID="90f390eee7663f7d922beeb1ad82261c623b08d8675519f7781b87762141d4b5" Oct 02 04:18:20 crc kubenswrapper[4775]: E1002 04:18:20.231415 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90f390eee7663f7d922beeb1ad82261c623b08d8675519f7781b87762141d4b5\": container with ID starting with 90f390eee7663f7d922beeb1ad82261c623b08d8675519f7781b87762141d4b5 not found: ID does not exist" containerID="90f390eee7663f7d922beeb1ad82261c623b08d8675519f7781b87762141d4b5" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.231457 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90f390eee7663f7d922beeb1ad82261c623b08d8675519f7781b87762141d4b5"} err="failed to get container status \"90f390eee7663f7d922beeb1ad82261c623b08d8675519f7781b87762141d4b5\": rpc error: code = NotFound desc = could not find container \"90f390eee7663f7d922beeb1ad82261c623b08d8675519f7781b87762141d4b5\": container with ID starting with 90f390eee7663f7d922beeb1ad82261c623b08d8675519f7781b87762141d4b5 not found: ID does not exist" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.231483 4775 scope.go:117] "RemoveContainer" containerID="3c126df49851dea425ccb541790944b93c996d0f77e8edd21f3cc435117dd170" Oct 02 04:18:20 crc kubenswrapper[4775]: E1002 04:18:20.232009 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c126df49851dea425ccb541790944b93c996d0f77e8edd21f3cc435117dd170\": container with ID starting with 3c126df49851dea425ccb541790944b93c996d0f77e8edd21f3cc435117dd170 not found: ID does not exist" containerID="3c126df49851dea425ccb541790944b93c996d0f77e8edd21f3cc435117dd170" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.232071 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c126df49851dea425ccb541790944b93c996d0f77e8edd21f3cc435117dd170"} err="failed to get container status \"3c126df49851dea425ccb541790944b93c996d0f77e8edd21f3cc435117dd170\": rpc error: code = NotFound desc = could not find container \"3c126df49851dea425ccb541790944b93c996d0f77e8edd21f3cc435117dd170\": container with ID starting with 3c126df49851dea425ccb541790944b93c996d0f77e8edd21f3cc435117dd170 not found: ID does not exist" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.232256 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec363e50-fb34-4304-a26e-72e2d41c84b0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.232301 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbnj8\" (UniqueName: \"kubernetes.io/projected/ec363e50-fb34-4304-a26e-72e2d41c84b0-kube-api-access-jbnj8\") on node \"crc\" DevicePath \"\"" Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.426577 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-djv7s"] Oct 02 04:18:20 crc kubenswrapper[4775]: I1002 04:18:20.439425 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-djv7s"] Oct 02 04:18:21 crc kubenswrapper[4775]: I1002 04:18:21.778864 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec363e50-fb34-4304-a26e-72e2d41c84b0" path="/var/lib/kubelet/pods/ec363e50-fb34-4304-a26e-72e2d41c84b0/volumes" Oct 02 04:20:37 crc kubenswrapper[4775]: I1002 04:20:37.233976 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:20:37 crc kubenswrapper[4775]: I1002 04:20:37.234646 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:21:07 crc kubenswrapper[4775]: I1002 04:21:07.233945 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:21:07 crc kubenswrapper[4775]: I1002 04:21:07.234783 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:21:37 crc kubenswrapper[4775]: I1002 04:21:37.233539 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:21:37 crc kubenswrapper[4775]: I1002 04:21:37.234450 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:21:37 crc kubenswrapper[4775]: I1002 04:21:37.234519 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 04:21:37 crc kubenswrapper[4775]: I1002 04:21:37.235748 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f513efcfa13070dc5964ae5079d55032a787ae801794c0913b3a3fd25e50f61a"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 04:21:37 crc kubenswrapper[4775]: I1002 04:21:37.235840 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://f513efcfa13070dc5964ae5079d55032a787ae801794c0913b3a3fd25e50f61a" gracePeriod=600 Oct 02 04:21:38 crc kubenswrapper[4775]: I1002 04:21:38.733437 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="f513efcfa13070dc5964ae5079d55032a787ae801794c0913b3a3fd25e50f61a" exitCode=0 Oct 02 04:21:38 crc kubenswrapper[4775]: I1002 04:21:38.733636 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"f513efcfa13070dc5964ae5079d55032a787ae801794c0913b3a3fd25e50f61a"} Oct 02 04:21:38 crc kubenswrapper[4775]: I1002 04:21:38.734144 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a"} Oct 02 04:21:38 crc kubenswrapper[4775]: I1002 04:21:38.734172 4775 scope.go:117] "RemoveContainer" containerID="67a7b4750214afd7c9fb2241e8a4e523907d6afdf060f684620b856c0a9aa404" Oct 02 04:24:07 crc kubenswrapper[4775]: I1002 04:24:07.234476 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:24:07 crc kubenswrapper[4775]: I1002 04:24:07.235146 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.098385 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-szmmt"] Oct 02 04:24:28 crc kubenswrapper[4775]: E1002 04:24:28.099807 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5700e992-86c2-43bf-8cbb-5c8285511f4c" containerName="registry-server" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.099830 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5700e992-86c2-43bf-8cbb-5c8285511f4c" containerName="registry-server" Oct 02 04:24:28 crc kubenswrapper[4775]: E1002 04:24:28.099887 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec363e50-fb34-4304-a26e-72e2d41c84b0" containerName="extract-utilities" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.099902 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec363e50-fb34-4304-a26e-72e2d41c84b0" containerName="extract-utilities" Oct 02 04:24:28 crc kubenswrapper[4775]: E1002 04:24:28.099942 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5700e992-86c2-43bf-8cbb-5c8285511f4c" containerName="extract-utilities" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.099978 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5700e992-86c2-43bf-8cbb-5c8285511f4c" containerName="extract-utilities" Oct 02 04:24:28 crc kubenswrapper[4775]: E1002 04:24:28.100007 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5700e992-86c2-43bf-8cbb-5c8285511f4c" containerName="extract-content" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.100020 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="5700e992-86c2-43bf-8cbb-5c8285511f4c" containerName="extract-content" Oct 02 04:24:28 crc kubenswrapper[4775]: E1002 04:24:28.100045 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec363e50-fb34-4304-a26e-72e2d41c84b0" containerName="registry-server" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.100057 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec363e50-fb34-4304-a26e-72e2d41c84b0" containerName="registry-server" Oct 02 04:24:28 crc kubenswrapper[4775]: E1002 04:24:28.100089 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec363e50-fb34-4304-a26e-72e2d41c84b0" containerName="extract-content" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.100101 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec363e50-fb34-4304-a26e-72e2d41c84b0" containerName="extract-content" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.100531 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="5700e992-86c2-43bf-8cbb-5c8285511f4c" containerName="registry-server" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.100572 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec363e50-fb34-4304-a26e-72e2d41c84b0" containerName="registry-server" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.103789 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.116406 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-szmmt"] Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.188080 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b799345f-7e1d-4f40-8fea-1376c08ce062-utilities\") pod \"redhat-marketplace-szmmt\" (UID: \"b799345f-7e1d-4f40-8fea-1376c08ce062\") " pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.188480 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mp2d\" (UniqueName: \"kubernetes.io/projected/b799345f-7e1d-4f40-8fea-1376c08ce062-kube-api-access-5mp2d\") pod \"redhat-marketplace-szmmt\" (UID: \"b799345f-7e1d-4f40-8fea-1376c08ce062\") " pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.188589 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b799345f-7e1d-4f40-8fea-1376c08ce062-catalog-content\") pod \"redhat-marketplace-szmmt\" (UID: \"b799345f-7e1d-4f40-8fea-1376c08ce062\") " pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.290828 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b799345f-7e1d-4f40-8fea-1376c08ce062-utilities\") pod \"redhat-marketplace-szmmt\" (UID: \"b799345f-7e1d-4f40-8fea-1376c08ce062\") " pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.290914 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mp2d\" (UniqueName: \"kubernetes.io/projected/b799345f-7e1d-4f40-8fea-1376c08ce062-kube-api-access-5mp2d\") pod \"redhat-marketplace-szmmt\" (UID: \"b799345f-7e1d-4f40-8fea-1376c08ce062\") " pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.291086 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b799345f-7e1d-4f40-8fea-1376c08ce062-catalog-content\") pod \"redhat-marketplace-szmmt\" (UID: \"b799345f-7e1d-4f40-8fea-1376c08ce062\") " pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.292708 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b799345f-7e1d-4f40-8fea-1376c08ce062-catalog-content\") pod \"redhat-marketplace-szmmt\" (UID: \"b799345f-7e1d-4f40-8fea-1376c08ce062\") " pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.293614 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b799345f-7e1d-4f40-8fea-1376c08ce062-utilities\") pod \"redhat-marketplace-szmmt\" (UID: \"b799345f-7e1d-4f40-8fea-1376c08ce062\") " pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.319874 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mp2d\" (UniqueName: \"kubernetes.io/projected/b799345f-7e1d-4f40-8fea-1376c08ce062-kube-api-access-5mp2d\") pod \"redhat-marketplace-szmmt\" (UID: \"b799345f-7e1d-4f40-8fea-1376c08ce062\") " pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.443059 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:28 crc kubenswrapper[4775]: I1002 04:24:28.993583 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-szmmt"] Oct 02 04:24:29 crc kubenswrapper[4775]: I1002 04:24:29.058036 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-szmmt" event={"ID":"b799345f-7e1d-4f40-8fea-1376c08ce062","Type":"ContainerStarted","Data":"ba75f1765dd9f2a02ebcfd087532e4d3a3475ad771269820fa55771daa95e8f2"} Oct 02 04:24:30 crc kubenswrapper[4775]: I1002 04:24:30.075066 4775 generic.go:334] "Generic (PLEG): container finished" podID="b799345f-7e1d-4f40-8fea-1376c08ce062" containerID="3c59a6e1d9f94e288409558f3ef05986c32ce83382e2e36689d0aa059d21e72d" exitCode=0 Oct 02 04:24:30 crc kubenswrapper[4775]: I1002 04:24:30.075213 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-szmmt" event={"ID":"b799345f-7e1d-4f40-8fea-1376c08ce062","Type":"ContainerDied","Data":"3c59a6e1d9f94e288409558f3ef05986c32ce83382e2e36689d0aa059d21e72d"} Oct 02 04:24:30 crc kubenswrapper[4775]: I1002 04:24:30.079223 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 04:24:31 crc kubenswrapper[4775]: I1002 04:24:31.090636 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-szmmt" event={"ID":"b799345f-7e1d-4f40-8fea-1376c08ce062","Type":"ContainerStarted","Data":"aa0a19bc7812e09c9ccc7a93cb1ea5aac1eb923818e232ec34c0cba4acf7319b"} Oct 02 04:24:32 crc kubenswrapper[4775]: I1002 04:24:32.112099 4775 generic.go:334] "Generic (PLEG): container finished" podID="b799345f-7e1d-4f40-8fea-1376c08ce062" containerID="aa0a19bc7812e09c9ccc7a93cb1ea5aac1eb923818e232ec34c0cba4acf7319b" exitCode=0 Oct 02 04:24:32 crc kubenswrapper[4775]: I1002 04:24:32.112198 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-szmmt" event={"ID":"b799345f-7e1d-4f40-8fea-1376c08ce062","Type":"ContainerDied","Data":"aa0a19bc7812e09c9ccc7a93cb1ea5aac1eb923818e232ec34c0cba4acf7319b"} Oct 02 04:24:34 crc kubenswrapper[4775]: I1002 04:24:34.147918 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-szmmt" event={"ID":"b799345f-7e1d-4f40-8fea-1376c08ce062","Type":"ContainerStarted","Data":"a39bf049c6f5a2c6f95b6709d351f026453fc0aa1e8c3d9ce391cd390b05291f"} Oct 02 04:24:34 crc kubenswrapper[4775]: I1002 04:24:34.189626 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-szmmt" podStartSLOduration=3.651964969 podStartE2EDuration="6.189596328s" podCreationTimestamp="2025-10-02 04:24:28 +0000 UTC" firstStartedPulling="2025-10-02 04:24:30.078975008 +0000 UTC m=+9807.245719048" lastFinishedPulling="2025-10-02 04:24:32.616606357 +0000 UTC m=+9809.783350407" observedRunningTime="2025-10-02 04:24:34.178703145 +0000 UTC m=+9811.345447225" watchObservedRunningTime="2025-10-02 04:24:34.189596328 +0000 UTC m=+9811.356340388" Oct 02 04:24:37 crc kubenswrapper[4775]: I1002 04:24:37.233588 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:24:37 crc kubenswrapper[4775]: I1002 04:24:37.234479 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:24:38 crc kubenswrapper[4775]: I1002 04:24:38.443175 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:38 crc kubenswrapper[4775]: I1002 04:24:38.443508 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:38 crc kubenswrapper[4775]: I1002 04:24:38.491714 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:39 crc kubenswrapper[4775]: I1002 04:24:39.300330 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:39 crc kubenswrapper[4775]: I1002 04:24:39.393905 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-szmmt"] Oct 02 04:24:41 crc kubenswrapper[4775]: I1002 04:24:41.230233 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-szmmt" podUID="b799345f-7e1d-4f40-8fea-1376c08ce062" containerName="registry-server" containerID="cri-o://a39bf049c6f5a2c6f95b6709d351f026453fc0aa1e8c3d9ce391cd390b05291f" gracePeriod=2 Oct 02 04:24:41 crc kubenswrapper[4775]: I1002 04:24:41.835537 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.029413 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b799345f-7e1d-4f40-8fea-1376c08ce062-catalog-content\") pod \"b799345f-7e1d-4f40-8fea-1376c08ce062\" (UID: \"b799345f-7e1d-4f40-8fea-1376c08ce062\") " Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.029793 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b799345f-7e1d-4f40-8fea-1376c08ce062-utilities\") pod \"b799345f-7e1d-4f40-8fea-1376c08ce062\" (UID: \"b799345f-7e1d-4f40-8fea-1376c08ce062\") " Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.029845 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mp2d\" (UniqueName: \"kubernetes.io/projected/b799345f-7e1d-4f40-8fea-1376c08ce062-kube-api-access-5mp2d\") pod \"b799345f-7e1d-4f40-8fea-1376c08ce062\" (UID: \"b799345f-7e1d-4f40-8fea-1376c08ce062\") " Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.030603 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b799345f-7e1d-4f40-8fea-1376c08ce062-utilities" (OuterVolumeSpecName: "utilities") pod "b799345f-7e1d-4f40-8fea-1376c08ce062" (UID: "b799345f-7e1d-4f40-8fea-1376c08ce062"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.031162 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b799345f-7e1d-4f40-8fea-1376c08ce062-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.040159 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b799345f-7e1d-4f40-8fea-1376c08ce062-kube-api-access-5mp2d" (OuterVolumeSpecName: "kube-api-access-5mp2d") pod "b799345f-7e1d-4f40-8fea-1376c08ce062" (UID: "b799345f-7e1d-4f40-8fea-1376c08ce062"). InnerVolumeSpecName "kube-api-access-5mp2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.041257 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b799345f-7e1d-4f40-8fea-1376c08ce062-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b799345f-7e1d-4f40-8fea-1376c08ce062" (UID: "b799345f-7e1d-4f40-8fea-1376c08ce062"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.135732 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mp2d\" (UniqueName: \"kubernetes.io/projected/b799345f-7e1d-4f40-8fea-1376c08ce062-kube-api-access-5mp2d\") on node \"crc\" DevicePath \"\"" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.135804 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b799345f-7e1d-4f40-8fea-1376c08ce062-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.244339 4775 generic.go:334] "Generic (PLEG): container finished" podID="b799345f-7e1d-4f40-8fea-1376c08ce062" containerID="a39bf049c6f5a2c6f95b6709d351f026453fc0aa1e8c3d9ce391cd390b05291f" exitCode=0 Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.244378 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-szmmt" event={"ID":"b799345f-7e1d-4f40-8fea-1376c08ce062","Type":"ContainerDied","Data":"a39bf049c6f5a2c6f95b6709d351f026453fc0aa1e8c3d9ce391cd390b05291f"} Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.244399 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-szmmt" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.244421 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-szmmt" event={"ID":"b799345f-7e1d-4f40-8fea-1376c08ce062","Type":"ContainerDied","Data":"ba75f1765dd9f2a02ebcfd087532e4d3a3475ad771269820fa55771daa95e8f2"} Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.244441 4775 scope.go:117] "RemoveContainer" containerID="a39bf049c6f5a2c6f95b6709d351f026453fc0aa1e8c3d9ce391cd390b05291f" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.288999 4775 scope.go:117] "RemoveContainer" containerID="aa0a19bc7812e09c9ccc7a93cb1ea5aac1eb923818e232ec34c0cba4acf7319b" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.296913 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-szmmt"] Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.308643 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-szmmt"] Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.326549 4775 scope.go:117] "RemoveContainer" containerID="3c59a6e1d9f94e288409558f3ef05986c32ce83382e2e36689d0aa059d21e72d" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.419517 4775 scope.go:117] "RemoveContainer" containerID="a39bf049c6f5a2c6f95b6709d351f026453fc0aa1e8c3d9ce391cd390b05291f" Oct 02 04:24:42 crc kubenswrapper[4775]: E1002 04:24:42.420584 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a39bf049c6f5a2c6f95b6709d351f026453fc0aa1e8c3d9ce391cd390b05291f\": container with ID starting with a39bf049c6f5a2c6f95b6709d351f026453fc0aa1e8c3d9ce391cd390b05291f not found: ID does not exist" containerID="a39bf049c6f5a2c6f95b6709d351f026453fc0aa1e8c3d9ce391cd390b05291f" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.420660 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a39bf049c6f5a2c6f95b6709d351f026453fc0aa1e8c3d9ce391cd390b05291f"} err="failed to get container status \"a39bf049c6f5a2c6f95b6709d351f026453fc0aa1e8c3d9ce391cd390b05291f\": rpc error: code = NotFound desc = could not find container \"a39bf049c6f5a2c6f95b6709d351f026453fc0aa1e8c3d9ce391cd390b05291f\": container with ID starting with a39bf049c6f5a2c6f95b6709d351f026453fc0aa1e8c3d9ce391cd390b05291f not found: ID does not exist" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.420749 4775 scope.go:117] "RemoveContainer" containerID="aa0a19bc7812e09c9ccc7a93cb1ea5aac1eb923818e232ec34c0cba4acf7319b" Oct 02 04:24:42 crc kubenswrapper[4775]: E1002 04:24:42.421194 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa0a19bc7812e09c9ccc7a93cb1ea5aac1eb923818e232ec34c0cba4acf7319b\": container with ID starting with aa0a19bc7812e09c9ccc7a93cb1ea5aac1eb923818e232ec34c0cba4acf7319b not found: ID does not exist" containerID="aa0a19bc7812e09c9ccc7a93cb1ea5aac1eb923818e232ec34c0cba4acf7319b" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.421235 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa0a19bc7812e09c9ccc7a93cb1ea5aac1eb923818e232ec34c0cba4acf7319b"} err="failed to get container status \"aa0a19bc7812e09c9ccc7a93cb1ea5aac1eb923818e232ec34c0cba4acf7319b\": rpc error: code = NotFound desc = could not find container \"aa0a19bc7812e09c9ccc7a93cb1ea5aac1eb923818e232ec34c0cba4acf7319b\": container with ID starting with aa0a19bc7812e09c9ccc7a93cb1ea5aac1eb923818e232ec34c0cba4acf7319b not found: ID does not exist" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.421262 4775 scope.go:117] "RemoveContainer" containerID="3c59a6e1d9f94e288409558f3ef05986c32ce83382e2e36689d0aa059d21e72d" Oct 02 04:24:42 crc kubenswrapper[4775]: E1002 04:24:42.421605 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c59a6e1d9f94e288409558f3ef05986c32ce83382e2e36689d0aa059d21e72d\": container with ID starting with 3c59a6e1d9f94e288409558f3ef05986c32ce83382e2e36689d0aa059d21e72d not found: ID does not exist" containerID="3c59a6e1d9f94e288409558f3ef05986c32ce83382e2e36689d0aa059d21e72d" Oct 02 04:24:42 crc kubenswrapper[4775]: I1002 04:24:42.421707 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c59a6e1d9f94e288409558f3ef05986c32ce83382e2e36689d0aa059d21e72d"} err="failed to get container status \"3c59a6e1d9f94e288409558f3ef05986c32ce83382e2e36689d0aa059d21e72d\": rpc error: code = NotFound desc = could not find container \"3c59a6e1d9f94e288409558f3ef05986c32ce83382e2e36689d0aa059d21e72d\": container with ID starting with 3c59a6e1d9f94e288409558f3ef05986c32ce83382e2e36689d0aa059d21e72d not found: ID does not exist" Oct 02 04:24:43 crc kubenswrapper[4775]: I1002 04:24:43.783045 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b799345f-7e1d-4f40-8fea-1376c08ce062" path="/var/lib/kubelet/pods/b799345f-7e1d-4f40-8fea-1376c08ce062/volumes" Oct 02 04:25:07 crc kubenswrapper[4775]: I1002 04:25:07.233705 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:25:07 crc kubenswrapper[4775]: I1002 04:25:07.234660 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:25:07 crc kubenswrapper[4775]: I1002 04:25:07.234755 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 04:25:07 crc kubenswrapper[4775]: I1002 04:25:07.236361 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 04:25:07 crc kubenswrapper[4775]: I1002 04:25:07.236467 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" gracePeriod=600 Oct 02 04:25:07 crc kubenswrapper[4775]: E1002 04:25:07.368615 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:25:07 crc kubenswrapper[4775]: I1002 04:25:07.625329 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" exitCode=0 Oct 02 04:25:07 crc kubenswrapper[4775]: I1002 04:25:07.625389 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a"} Oct 02 04:25:07 crc kubenswrapper[4775]: I1002 04:25:07.625428 4775 scope.go:117] "RemoveContainer" containerID="f513efcfa13070dc5964ae5079d55032a787ae801794c0913b3a3fd25e50f61a" Oct 02 04:25:07 crc kubenswrapper[4775]: I1002 04:25:07.626320 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:25:07 crc kubenswrapper[4775]: E1002 04:25:07.626699 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:25:21 crc kubenswrapper[4775]: I1002 04:25:21.766447 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:25:21 crc kubenswrapper[4775]: E1002 04:25:21.767532 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:25:35 crc kubenswrapper[4775]: I1002 04:25:35.766299 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:25:35 crc kubenswrapper[4775]: E1002 04:25:35.767401 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:25:48 crc kubenswrapper[4775]: I1002 04:25:48.765869 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:25:48 crc kubenswrapper[4775]: E1002 04:25:48.766691 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:26:00 crc kubenswrapper[4775]: I1002 04:26:00.765764 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:26:00 crc kubenswrapper[4775]: E1002 04:26:00.766861 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:26:12 crc kubenswrapper[4775]: I1002 04:26:12.765090 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:26:12 crc kubenswrapper[4775]: E1002 04:26:12.766224 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:26:26 crc kubenswrapper[4775]: I1002 04:26:26.765709 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:26:26 crc kubenswrapper[4775]: E1002 04:26:26.766596 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:26:37 crc kubenswrapper[4775]: I1002 04:26:37.765670 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:26:37 crc kubenswrapper[4775]: E1002 04:26:37.767077 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:26:51 crc kubenswrapper[4775]: I1002 04:26:51.768275 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:26:51 crc kubenswrapper[4775]: E1002 04:26:51.769512 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:27:05 crc kubenswrapper[4775]: I1002 04:27:05.766641 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:27:05 crc kubenswrapper[4775]: E1002 04:27:05.769361 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:27:18 crc kubenswrapper[4775]: I1002 04:27:18.766741 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:27:18 crc kubenswrapper[4775]: E1002 04:27:18.767721 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:27:32 crc kubenswrapper[4775]: I1002 04:27:32.766592 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:27:32 crc kubenswrapper[4775]: E1002 04:27:32.767734 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:27:33 crc kubenswrapper[4775]: I1002 04:27:33.716890 4775 generic.go:334] "Generic (PLEG): container finished" podID="b5d0f695-e66e-464e-bf9b-301a66a74f1b" containerID="9b358381c9f09ca411bbcf33af99c6782e31ce86d970ce90b1c95573393a2d2e" exitCode=0 Oct 02 04:27:33 crc kubenswrapper[4775]: I1002 04:27:33.717113 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" event={"ID":"b5d0f695-e66e-464e-bf9b-301a66a74f1b","Type":"ContainerDied","Data":"9b358381c9f09ca411bbcf33af99c6782e31ce86d970ce90b1c95573393a2d2e"} Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.272712 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.351823 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-combined-ca-bundle\") pod \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.351913 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-inventory\") pod \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.351991 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-compute-config-0\") pod \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.352022 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-ssh-key\") pod \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.352078 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-compute-config-1\") pod \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.352133 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cells-global-config-1\") pod \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.352189 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-ceph\") pod \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.352216 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-migration-ssh-key-0\") pod \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.352238 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cells-global-config-0\") pod \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.352321 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2mfq\" (UniqueName: \"kubernetes.io/projected/b5d0f695-e66e-464e-bf9b-301a66a74f1b-kube-api-access-b2mfq\") pod \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.352361 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-migration-ssh-key-1\") pod \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\" (UID: \"b5d0f695-e66e-464e-bf9b-301a66a74f1b\") " Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.361457 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-ceph" (OuterVolumeSpecName: "ceph") pod "b5d0f695-e66e-464e-bf9b-301a66a74f1b" (UID: "b5d0f695-e66e-464e-bf9b-301a66a74f1b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.362797 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "b5d0f695-e66e-464e-bf9b-301a66a74f1b" (UID: "b5d0f695-e66e-464e-bf9b-301a66a74f1b"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.365254 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5d0f695-e66e-464e-bf9b-301a66a74f1b-kube-api-access-b2mfq" (OuterVolumeSpecName: "kube-api-access-b2mfq") pod "b5d0f695-e66e-464e-bf9b-301a66a74f1b" (UID: "b5d0f695-e66e-464e-bf9b-301a66a74f1b"). InnerVolumeSpecName "kube-api-access-b2mfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.383330 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "b5d0f695-e66e-464e-bf9b-301a66a74f1b" (UID: "b5d0f695-e66e-464e-bf9b-301a66a74f1b"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.401646 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "b5d0f695-e66e-464e-bf9b-301a66a74f1b" (UID: "b5d0f695-e66e-464e-bf9b-301a66a74f1b"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.403245 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b5d0f695-e66e-464e-bf9b-301a66a74f1b" (UID: "b5d0f695-e66e-464e-bf9b-301a66a74f1b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.409995 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "b5d0f695-e66e-464e-bf9b-301a66a74f1b" (UID: "b5d0f695-e66e-464e-bf9b-301a66a74f1b"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.414034 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "b5d0f695-e66e-464e-bf9b-301a66a74f1b" (UID: "b5d0f695-e66e-464e-bf9b-301a66a74f1b"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.427248 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "b5d0f695-e66e-464e-bf9b-301a66a74f1b" (UID: "b5d0f695-e66e-464e-bf9b-301a66a74f1b"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.440055 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "b5d0f695-e66e-464e-bf9b-301a66a74f1b" (UID: "b5d0f695-e66e-464e-bf9b-301a66a74f1b"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.441139 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-inventory" (OuterVolumeSpecName: "inventory") pod "b5d0f695-e66e-464e-bf9b-301a66a74f1b" (UID: "b5d0f695-e66e-464e-bf9b-301a66a74f1b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.454254 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.454293 4775 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.454305 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.454313 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.454323 4775 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-ceph\") on node \"crc\" DevicePath \"\"" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.454332 4775 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.454340 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.454349 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2mfq\" (UniqueName: \"kubernetes.io/projected/b5d0f695-e66e-464e-bf9b-301a66a74f1b-kube-api-access-b2mfq\") on node \"crc\" DevicePath \"\"" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.454357 4775 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.454365 4775 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.454373 4775 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5d0f695-e66e-464e-bf9b-301a66a74f1b-inventory\") on node \"crc\" DevicePath \"\"" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.743866 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" event={"ID":"b5d0f695-e66e-464e-bf9b-301a66a74f1b","Type":"ContainerDied","Data":"2c67ddb141f99b2433b909453a75d977e0c5a250424f3b702c349ac3a831ea2c"} Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.743917 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c67ddb141f99b2433b909453a75d977e0c5a250424f3b702c349ac3a831ea2c" Oct 02 04:27:35 crc kubenswrapper[4775]: I1002 04:27:35.744039 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv" Oct 02 04:27:43 crc kubenswrapper[4775]: I1002 04:27:43.782241 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:27:43 crc kubenswrapper[4775]: E1002 04:27:43.783566 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:27:57 crc kubenswrapper[4775]: I1002 04:27:57.766457 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:27:57 crc kubenswrapper[4775]: E1002 04:27:57.767145 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:28:09 crc kubenswrapper[4775]: I1002 04:28:09.765490 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:28:09 crc kubenswrapper[4775]: E1002 04:28:09.766409 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:28:24 crc kubenswrapper[4775]: I1002 04:28:24.766031 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:28:24 crc kubenswrapper[4775]: E1002 04:28:24.766828 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:28:35 crc kubenswrapper[4775]: I1002 04:28:35.766485 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:28:35 crc kubenswrapper[4775]: E1002 04:28:35.768079 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:28:49 crc kubenswrapper[4775]: I1002 04:28:49.769572 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:28:49 crc kubenswrapper[4775]: E1002 04:28:49.772396 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:29:04 crc kubenswrapper[4775]: I1002 04:29:04.765813 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:29:04 crc kubenswrapper[4775]: E1002 04:29:04.766935 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.205640 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5fgvl"] Oct 02 04:29:10 crc kubenswrapper[4775]: E1002 04:29:10.207055 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b799345f-7e1d-4f40-8fea-1376c08ce062" containerName="extract-utilities" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.207082 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b799345f-7e1d-4f40-8fea-1376c08ce062" containerName="extract-utilities" Oct 02 04:29:10 crc kubenswrapper[4775]: E1002 04:29:10.207121 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b799345f-7e1d-4f40-8fea-1376c08ce062" containerName="registry-server" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.207132 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b799345f-7e1d-4f40-8fea-1376c08ce062" containerName="registry-server" Oct 02 04:29:10 crc kubenswrapper[4775]: E1002 04:29:10.207149 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5d0f695-e66e-464e-bf9b-301a66a74f1b" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.207163 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5d0f695-e66e-464e-bf9b-301a66a74f1b" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 02 04:29:10 crc kubenswrapper[4775]: E1002 04:29:10.207188 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b799345f-7e1d-4f40-8fea-1376c08ce062" containerName="extract-content" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.207200 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="b799345f-7e1d-4f40-8fea-1376c08ce062" containerName="extract-content" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.207578 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b799345f-7e1d-4f40-8fea-1376c08ce062" containerName="registry-server" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.207610 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5d0f695-e66e-464e-bf9b-301a66a74f1b" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.210038 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.219324 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5fgvl"] Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.320577 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/184d89af-1876-42f3-86c6-1369ebda54bf-catalog-content\") pod \"certified-operators-5fgvl\" (UID: \"184d89af-1876-42f3-86c6-1369ebda54bf\") " pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.321013 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/184d89af-1876-42f3-86c6-1369ebda54bf-utilities\") pod \"certified-operators-5fgvl\" (UID: \"184d89af-1876-42f3-86c6-1369ebda54bf\") " pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.321217 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqwd9\" (UniqueName: \"kubernetes.io/projected/184d89af-1876-42f3-86c6-1369ebda54bf-kube-api-access-tqwd9\") pod \"certified-operators-5fgvl\" (UID: \"184d89af-1876-42f3-86c6-1369ebda54bf\") " pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.423461 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/184d89af-1876-42f3-86c6-1369ebda54bf-utilities\") pod \"certified-operators-5fgvl\" (UID: \"184d89af-1876-42f3-86c6-1369ebda54bf\") " pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.423551 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqwd9\" (UniqueName: \"kubernetes.io/projected/184d89af-1876-42f3-86c6-1369ebda54bf-kube-api-access-tqwd9\") pod \"certified-operators-5fgvl\" (UID: \"184d89af-1876-42f3-86c6-1369ebda54bf\") " pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.423623 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/184d89af-1876-42f3-86c6-1369ebda54bf-catalog-content\") pod \"certified-operators-5fgvl\" (UID: \"184d89af-1876-42f3-86c6-1369ebda54bf\") " pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.424187 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/184d89af-1876-42f3-86c6-1369ebda54bf-catalog-content\") pod \"certified-operators-5fgvl\" (UID: \"184d89af-1876-42f3-86c6-1369ebda54bf\") " pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.424315 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/184d89af-1876-42f3-86c6-1369ebda54bf-utilities\") pod \"certified-operators-5fgvl\" (UID: \"184d89af-1876-42f3-86c6-1369ebda54bf\") " pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.441680 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqwd9\" (UniqueName: \"kubernetes.io/projected/184d89af-1876-42f3-86c6-1369ebda54bf-kube-api-access-tqwd9\") pod \"certified-operators-5fgvl\" (UID: \"184d89af-1876-42f3-86c6-1369ebda54bf\") " pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:10 crc kubenswrapper[4775]: I1002 04:29:10.550888 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:11 crc kubenswrapper[4775]: I1002 04:29:11.155498 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5fgvl"] Oct 02 04:29:11 crc kubenswrapper[4775]: I1002 04:29:11.997857 4775 generic.go:334] "Generic (PLEG): container finished" podID="184d89af-1876-42f3-86c6-1369ebda54bf" containerID="20338b7dd0980c4daa56c109aba3139ee5cd902ebdf9593de5eec2f0eb1a67c8" exitCode=0 Oct 02 04:29:11 crc kubenswrapper[4775]: I1002 04:29:11.997947 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5fgvl" event={"ID":"184d89af-1876-42f3-86c6-1369ebda54bf","Type":"ContainerDied","Data":"20338b7dd0980c4daa56c109aba3139ee5cd902ebdf9593de5eec2f0eb1a67c8"} Oct 02 04:29:11 crc kubenswrapper[4775]: I1002 04:29:11.998267 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5fgvl" event={"ID":"184d89af-1876-42f3-86c6-1369ebda54bf","Type":"ContainerStarted","Data":"c63404d26d2744faf714c454456b14930314ee94197b1bd0254973c067edb059"} Oct 02 04:29:14 crc kubenswrapper[4775]: I1002 04:29:14.033065 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5fgvl" event={"ID":"184d89af-1876-42f3-86c6-1369ebda54bf","Type":"ContainerStarted","Data":"1a0abeab5d6f6484be90bd9da2052f3f817c97e931f16c0210a2ce4c81e7b9dc"} Oct 02 04:29:15 crc kubenswrapper[4775]: I1002 04:29:15.052533 4775 generic.go:334] "Generic (PLEG): container finished" podID="184d89af-1876-42f3-86c6-1369ebda54bf" containerID="1a0abeab5d6f6484be90bd9da2052f3f817c97e931f16c0210a2ce4c81e7b9dc" exitCode=0 Oct 02 04:29:15 crc kubenswrapper[4775]: I1002 04:29:15.052695 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5fgvl" event={"ID":"184d89af-1876-42f3-86c6-1369ebda54bf","Type":"ContainerDied","Data":"1a0abeab5d6f6484be90bd9da2052f3f817c97e931f16c0210a2ce4c81e7b9dc"} Oct 02 04:29:15 crc kubenswrapper[4775]: I1002 04:29:15.765484 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:29:15 crc kubenswrapper[4775]: E1002 04:29:15.765940 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:29:16 crc kubenswrapper[4775]: I1002 04:29:16.064588 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5fgvl" event={"ID":"184d89af-1876-42f3-86c6-1369ebda54bf","Type":"ContainerStarted","Data":"db4a3b284bede2824872f6751899b299d3cf9f27e4e03dfd7b1960a539a77392"} Oct 02 04:29:16 crc kubenswrapper[4775]: I1002 04:29:16.094466 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5fgvl" podStartSLOduration=2.574500887 podStartE2EDuration="6.094406455s" podCreationTimestamp="2025-10-02 04:29:10 +0000 UTC" firstStartedPulling="2025-10-02 04:29:12.000998653 +0000 UTC m=+10089.167742733" lastFinishedPulling="2025-10-02 04:29:15.520904231 +0000 UTC m=+10092.687648301" observedRunningTime="2025-10-02 04:29:16.085034072 +0000 UTC m=+10093.251778132" watchObservedRunningTime="2025-10-02 04:29:16.094406455 +0000 UTC m=+10093.261150495" Oct 02 04:29:20 crc kubenswrapper[4775]: I1002 04:29:20.551371 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:20 crc kubenswrapper[4775]: I1002 04:29:20.552257 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:20 crc kubenswrapper[4775]: I1002 04:29:20.641204 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:21 crc kubenswrapper[4775]: I1002 04:29:21.203462 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:21 crc kubenswrapper[4775]: I1002 04:29:21.285350 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5fgvl"] Oct 02 04:29:23 crc kubenswrapper[4775]: I1002 04:29:23.153507 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5fgvl" podUID="184d89af-1876-42f3-86c6-1369ebda54bf" containerName="registry-server" containerID="cri-o://db4a3b284bede2824872f6751899b299d3cf9f27e4e03dfd7b1960a539a77392" gracePeriod=2 Oct 02 04:29:23 crc kubenswrapper[4775]: I1002 04:29:23.761613 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:23 crc kubenswrapper[4775]: I1002 04:29:23.879132 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqwd9\" (UniqueName: \"kubernetes.io/projected/184d89af-1876-42f3-86c6-1369ebda54bf-kube-api-access-tqwd9\") pod \"184d89af-1876-42f3-86c6-1369ebda54bf\" (UID: \"184d89af-1876-42f3-86c6-1369ebda54bf\") " Oct 02 04:29:23 crc kubenswrapper[4775]: I1002 04:29:23.879229 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/184d89af-1876-42f3-86c6-1369ebda54bf-catalog-content\") pod \"184d89af-1876-42f3-86c6-1369ebda54bf\" (UID: \"184d89af-1876-42f3-86c6-1369ebda54bf\") " Oct 02 04:29:23 crc kubenswrapper[4775]: I1002 04:29:23.879519 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/184d89af-1876-42f3-86c6-1369ebda54bf-utilities\") pod \"184d89af-1876-42f3-86c6-1369ebda54bf\" (UID: \"184d89af-1876-42f3-86c6-1369ebda54bf\") " Oct 02 04:29:23 crc kubenswrapper[4775]: I1002 04:29:23.881074 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/184d89af-1876-42f3-86c6-1369ebda54bf-utilities" (OuterVolumeSpecName: "utilities") pod "184d89af-1876-42f3-86c6-1369ebda54bf" (UID: "184d89af-1876-42f3-86c6-1369ebda54bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:29:23 crc kubenswrapper[4775]: I1002 04:29:23.888591 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/184d89af-1876-42f3-86c6-1369ebda54bf-kube-api-access-tqwd9" (OuterVolumeSpecName: "kube-api-access-tqwd9") pod "184d89af-1876-42f3-86c6-1369ebda54bf" (UID: "184d89af-1876-42f3-86c6-1369ebda54bf"). InnerVolumeSpecName "kube-api-access-tqwd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:29:23 crc kubenswrapper[4775]: I1002 04:29:23.932821 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/184d89af-1876-42f3-86c6-1369ebda54bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "184d89af-1876-42f3-86c6-1369ebda54bf" (UID: "184d89af-1876-42f3-86c6-1369ebda54bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:29:23 crc kubenswrapper[4775]: I1002 04:29:23.982483 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/184d89af-1876-42f3-86c6-1369ebda54bf-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 04:29:23 crc kubenswrapper[4775]: I1002 04:29:23.982526 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqwd9\" (UniqueName: \"kubernetes.io/projected/184d89af-1876-42f3-86c6-1369ebda54bf-kube-api-access-tqwd9\") on node \"crc\" DevicePath \"\"" Oct 02 04:29:23 crc kubenswrapper[4775]: I1002 04:29:23.982537 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/184d89af-1876-42f3-86c6-1369ebda54bf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.168997 4775 generic.go:334] "Generic (PLEG): container finished" podID="184d89af-1876-42f3-86c6-1369ebda54bf" containerID="db4a3b284bede2824872f6751899b299d3cf9f27e4e03dfd7b1960a539a77392" exitCode=0 Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.169056 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5fgvl" event={"ID":"184d89af-1876-42f3-86c6-1369ebda54bf","Type":"ContainerDied","Data":"db4a3b284bede2824872f6751899b299d3cf9f27e4e03dfd7b1960a539a77392"} Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.169097 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5fgvl" Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.169109 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5fgvl" event={"ID":"184d89af-1876-42f3-86c6-1369ebda54bf","Type":"ContainerDied","Data":"c63404d26d2744faf714c454456b14930314ee94197b1bd0254973c067edb059"} Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.169143 4775 scope.go:117] "RemoveContainer" containerID="db4a3b284bede2824872f6751899b299d3cf9f27e4e03dfd7b1960a539a77392" Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.202875 4775 scope.go:117] "RemoveContainer" containerID="1a0abeab5d6f6484be90bd9da2052f3f817c97e931f16c0210a2ce4c81e7b9dc" Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.229184 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5fgvl"] Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.241800 4775 scope.go:117] "RemoveContainer" containerID="20338b7dd0980c4daa56c109aba3139ee5cd902ebdf9593de5eec2f0eb1a67c8" Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.244906 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5fgvl"] Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.317070 4775 scope.go:117] "RemoveContainer" containerID="db4a3b284bede2824872f6751899b299d3cf9f27e4e03dfd7b1960a539a77392" Oct 02 04:29:24 crc kubenswrapper[4775]: E1002 04:29:24.317700 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db4a3b284bede2824872f6751899b299d3cf9f27e4e03dfd7b1960a539a77392\": container with ID starting with db4a3b284bede2824872f6751899b299d3cf9f27e4e03dfd7b1960a539a77392 not found: ID does not exist" containerID="db4a3b284bede2824872f6751899b299d3cf9f27e4e03dfd7b1960a539a77392" Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.317745 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db4a3b284bede2824872f6751899b299d3cf9f27e4e03dfd7b1960a539a77392"} err="failed to get container status \"db4a3b284bede2824872f6751899b299d3cf9f27e4e03dfd7b1960a539a77392\": rpc error: code = NotFound desc = could not find container \"db4a3b284bede2824872f6751899b299d3cf9f27e4e03dfd7b1960a539a77392\": container with ID starting with db4a3b284bede2824872f6751899b299d3cf9f27e4e03dfd7b1960a539a77392 not found: ID does not exist" Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.317796 4775 scope.go:117] "RemoveContainer" containerID="1a0abeab5d6f6484be90bd9da2052f3f817c97e931f16c0210a2ce4c81e7b9dc" Oct 02 04:29:24 crc kubenswrapper[4775]: E1002 04:29:24.318274 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a0abeab5d6f6484be90bd9da2052f3f817c97e931f16c0210a2ce4c81e7b9dc\": container with ID starting with 1a0abeab5d6f6484be90bd9da2052f3f817c97e931f16c0210a2ce4c81e7b9dc not found: ID does not exist" containerID="1a0abeab5d6f6484be90bd9da2052f3f817c97e931f16c0210a2ce4c81e7b9dc" Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.318339 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a0abeab5d6f6484be90bd9da2052f3f817c97e931f16c0210a2ce4c81e7b9dc"} err="failed to get container status \"1a0abeab5d6f6484be90bd9da2052f3f817c97e931f16c0210a2ce4c81e7b9dc\": rpc error: code = NotFound desc = could not find container \"1a0abeab5d6f6484be90bd9da2052f3f817c97e931f16c0210a2ce4c81e7b9dc\": container with ID starting with 1a0abeab5d6f6484be90bd9da2052f3f817c97e931f16c0210a2ce4c81e7b9dc not found: ID does not exist" Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.318377 4775 scope.go:117] "RemoveContainer" containerID="20338b7dd0980c4daa56c109aba3139ee5cd902ebdf9593de5eec2f0eb1a67c8" Oct 02 04:29:24 crc kubenswrapper[4775]: E1002 04:29:24.319249 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20338b7dd0980c4daa56c109aba3139ee5cd902ebdf9593de5eec2f0eb1a67c8\": container with ID starting with 20338b7dd0980c4daa56c109aba3139ee5cd902ebdf9593de5eec2f0eb1a67c8 not found: ID does not exist" containerID="20338b7dd0980c4daa56c109aba3139ee5cd902ebdf9593de5eec2f0eb1a67c8" Oct 02 04:29:24 crc kubenswrapper[4775]: I1002 04:29:24.319281 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20338b7dd0980c4daa56c109aba3139ee5cd902ebdf9593de5eec2f0eb1a67c8"} err="failed to get container status \"20338b7dd0980c4daa56c109aba3139ee5cd902ebdf9593de5eec2f0eb1a67c8\": rpc error: code = NotFound desc = could not find container \"20338b7dd0980c4daa56c109aba3139ee5cd902ebdf9593de5eec2f0eb1a67c8\": container with ID starting with 20338b7dd0980c4daa56c109aba3139ee5cd902ebdf9593de5eec2f0eb1a67c8 not found: ID does not exist" Oct 02 04:29:25 crc kubenswrapper[4775]: I1002 04:29:25.790565 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="184d89af-1876-42f3-86c6-1369ebda54bf" path="/var/lib/kubelet/pods/184d89af-1876-42f3-86c6-1369ebda54bf/volumes" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.613377 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8ct8l"] Oct 02 04:29:27 crc kubenswrapper[4775]: E1002 04:29:27.614670 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="184d89af-1876-42f3-86c6-1369ebda54bf" containerName="extract-content" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.614686 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="184d89af-1876-42f3-86c6-1369ebda54bf" containerName="extract-content" Oct 02 04:29:27 crc kubenswrapper[4775]: E1002 04:29:27.614706 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="184d89af-1876-42f3-86c6-1369ebda54bf" containerName="registry-server" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.614714 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="184d89af-1876-42f3-86c6-1369ebda54bf" containerName="registry-server" Oct 02 04:29:27 crc kubenswrapper[4775]: E1002 04:29:27.614725 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="184d89af-1876-42f3-86c6-1369ebda54bf" containerName="extract-utilities" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.614734 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="184d89af-1876-42f3-86c6-1369ebda54bf" containerName="extract-utilities" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.615182 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="184d89af-1876-42f3-86c6-1369ebda54bf" containerName="registry-server" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.617915 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.636337 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8ct8l"] Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.679513 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/598dc99a-0eb3-4bad-a630-3a44b69775f6-catalog-content\") pod \"redhat-operators-8ct8l\" (UID: \"598dc99a-0eb3-4bad-a630-3a44b69775f6\") " pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.681828 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/598dc99a-0eb3-4bad-a630-3a44b69775f6-utilities\") pod \"redhat-operators-8ct8l\" (UID: \"598dc99a-0eb3-4bad-a630-3a44b69775f6\") " pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.681985 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdps4\" (UniqueName: \"kubernetes.io/projected/598dc99a-0eb3-4bad-a630-3a44b69775f6-kube-api-access-kdps4\") pod \"redhat-operators-8ct8l\" (UID: \"598dc99a-0eb3-4bad-a630-3a44b69775f6\") " pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.765309 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:29:27 crc kubenswrapper[4775]: E1002 04:29:27.766021 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.784336 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/598dc99a-0eb3-4bad-a630-3a44b69775f6-catalog-content\") pod \"redhat-operators-8ct8l\" (UID: \"598dc99a-0eb3-4bad-a630-3a44b69775f6\") " pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.784405 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/598dc99a-0eb3-4bad-a630-3a44b69775f6-utilities\") pod \"redhat-operators-8ct8l\" (UID: \"598dc99a-0eb3-4bad-a630-3a44b69775f6\") " pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.784544 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdps4\" (UniqueName: \"kubernetes.io/projected/598dc99a-0eb3-4bad-a630-3a44b69775f6-kube-api-access-kdps4\") pod \"redhat-operators-8ct8l\" (UID: \"598dc99a-0eb3-4bad-a630-3a44b69775f6\") " pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.785140 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/598dc99a-0eb3-4bad-a630-3a44b69775f6-catalog-content\") pod \"redhat-operators-8ct8l\" (UID: \"598dc99a-0eb3-4bad-a630-3a44b69775f6\") " pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.785382 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/598dc99a-0eb3-4bad-a630-3a44b69775f6-utilities\") pod \"redhat-operators-8ct8l\" (UID: \"598dc99a-0eb3-4bad-a630-3a44b69775f6\") " pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.817634 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdps4\" (UniqueName: \"kubernetes.io/projected/598dc99a-0eb3-4bad-a630-3a44b69775f6-kube-api-access-kdps4\") pod \"redhat-operators-8ct8l\" (UID: \"598dc99a-0eb3-4bad-a630-3a44b69775f6\") " pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:27 crc kubenswrapper[4775]: I1002 04:29:27.959616 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:28 crc kubenswrapper[4775]: I1002 04:29:28.458456 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8ct8l"] Oct 02 04:29:28 crc kubenswrapper[4775]: W1002 04:29:28.464434 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod598dc99a_0eb3_4bad_a630_3a44b69775f6.slice/crio-4aa8063e8df91980724ba409ea58cc9440840385447e2fee03899f847cd7b990 WatchSource:0}: Error finding container 4aa8063e8df91980724ba409ea58cc9440840385447e2fee03899f847cd7b990: Status 404 returned error can't find the container with id 4aa8063e8df91980724ba409ea58cc9440840385447e2fee03899f847cd7b990 Oct 02 04:29:29 crc kubenswrapper[4775]: I1002 04:29:29.237765 4775 generic.go:334] "Generic (PLEG): container finished" podID="598dc99a-0eb3-4bad-a630-3a44b69775f6" containerID="3e27b9819c998d1ec835b1df1f152840a70a92cbde9b68a07faf0627868df84d" exitCode=0 Oct 02 04:29:29 crc kubenswrapper[4775]: I1002 04:29:29.237835 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ct8l" event={"ID":"598dc99a-0eb3-4bad-a630-3a44b69775f6","Type":"ContainerDied","Data":"3e27b9819c998d1ec835b1df1f152840a70a92cbde9b68a07faf0627868df84d"} Oct 02 04:29:29 crc kubenswrapper[4775]: I1002 04:29:29.238275 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ct8l" event={"ID":"598dc99a-0eb3-4bad-a630-3a44b69775f6","Type":"ContainerStarted","Data":"4aa8063e8df91980724ba409ea58cc9440840385447e2fee03899f847cd7b990"} Oct 02 04:29:31 crc kubenswrapper[4775]: I1002 04:29:31.260683 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ct8l" event={"ID":"598dc99a-0eb3-4bad-a630-3a44b69775f6","Type":"ContainerStarted","Data":"d182b535e96bc26a023cd841415af05fe1863bd3f2846673acffb11e8904142a"} Oct 02 04:29:34 crc kubenswrapper[4775]: I1002 04:29:34.302847 4775 generic.go:334] "Generic (PLEG): container finished" podID="598dc99a-0eb3-4bad-a630-3a44b69775f6" containerID="d182b535e96bc26a023cd841415af05fe1863bd3f2846673acffb11e8904142a" exitCode=0 Oct 02 04:29:34 crc kubenswrapper[4775]: I1002 04:29:34.302922 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ct8l" event={"ID":"598dc99a-0eb3-4bad-a630-3a44b69775f6","Type":"ContainerDied","Data":"d182b535e96bc26a023cd841415af05fe1863bd3f2846673acffb11e8904142a"} Oct 02 04:29:34 crc kubenswrapper[4775]: I1002 04:29:34.306841 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 04:29:35 crc kubenswrapper[4775]: I1002 04:29:35.316754 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ct8l" event={"ID":"598dc99a-0eb3-4bad-a630-3a44b69775f6","Type":"ContainerStarted","Data":"85eca3cb43935ccd119b8a730985b8c72f8e6dc5056613640414220d439c89ef"} Oct 02 04:29:35 crc kubenswrapper[4775]: I1002 04:29:35.349028 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8ct8l" podStartSLOduration=2.7052507930000003 podStartE2EDuration="8.349003305s" podCreationTimestamp="2025-10-02 04:29:27 +0000 UTC" firstStartedPulling="2025-10-02 04:29:29.241600123 +0000 UTC m=+10106.408344163" lastFinishedPulling="2025-10-02 04:29:34.885352615 +0000 UTC m=+10112.052096675" observedRunningTime="2025-10-02 04:29:35.344103397 +0000 UTC m=+10112.510847457" watchObservedRunningTime="2025-10-02 04:29:35.349003305 +0000 UTC m=+10112.515747365" Oct 02 04:29:37 crc kubenswrapper[4775]: I1002 04:29:37.960406 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:37 crc kubenswrapper[4775]: I1002 04:29:37.961028 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:39 crc kubenswrapper[4775]: I1002 04:29:39.019512 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8ct8l" podUID="598dc99a-0eb3-4bad-a630-3a44b69775f6" containerName="registry-server" probeResult="failure" output=< Oct 02 04:29:39 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 04:29:39 crc kubenswrapper[4775]: > Oct 02 04:29:41 crc kubenswrapper[4775]: I1002 04:29:41.766033 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:29:41 crc kubenswrapper[4775]: E1002 04:29:41.767212 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:29:49 crc kubenswrapper[4775]: I1002 04:29:49.018631 4775 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-8ct8l" podUID="598dc99a-0eb3-4bad-a630-3a44b69775f6" containerName="registry-server" probeResult="failure" output=< Oct 02 04:29:49 crc kubenswrapper[4775]: timeout: failed to connect service ":50051" within 1s Oct 02 04:29:49 crc kubenswrapper[4775]: > Oct 02 04:29:51 crc kubenswrapper[4775]: I1002 04:29:51.123071 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tbjl6/must-gather-7px42"] Oct 02 04:29:51 crc kubenswrapper[4775]: I1002 04:29:51.125514 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/must-gather-7px42" Oct 02 04:29:51 crc kubenswrapper[4775]: I1002 04:29:51.127042 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-tbjl6"/"kube-root-ca.crt" Oct 02 04:29:51 crc kubenswrapper[4775]: I1002 04:29:51.127609 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-tbjl6"/"openshift-service-ca.crt" Oct 02 04:29:51 crc kubenswrapper[4775]: I1002 04:29:51.127778 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-tbjl6"/"default-dockercfg-kj5lq" Oct 02 04:29:51 crc kubenswrapper[4775]: I1002 04:29:51.131809 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tbjl6/must-gather-7px42"] Oct 02 04:29:51 crc kubenswrapper[4775]: I1002 04:29:51.237320 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c9sk\" (UniqueName: \"kubernetes.io/projected/1fd65775-68aa-4633-a43e-803f6b3ce9a2-kube-api-access-9c9sk\") pod \"must-gather-7px42\" (UID: \"1fd65775-68aa-4633-a43e-803f6b3ce9a2\") " pod="openshift-must-gather-tbjl6/must-gather-7px42" Oct 02 04:29:51 crc kubenswrapper[4775]: I1002 04:29:51.237424 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1fd65775-68aa-4633-a43e-803f6b3ce9a2-must-gather-output\") pod \"must-gather-7px42\" (UID: \"1fd65775-68aa-4633-a43e-803f6b3ce9a2\") " pod="openshift-must-gather-tbjl6/must-gather-7px42" Oct 02 04:29:51 crc kubenswrapper[4775]: I1002 04:29:51.339916 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c9sk\" (UniqueName: \"kubernetes.io/projected/1fd65775-68aa-4633-a43e-803f6b3ce9a2-kube-api-access-9c9sk\") pod \"must-gather-7px42\" (UID: \"1fd65775-68aa-4633-a43e-803f6b3ce9a2\") " pod="openshift-must-gather-tbjl6/must-gather-7px42" Oct 02 04:29:51 crc kubenswrapper[4775]: I1002 04:29:51.340068 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1fd65775-68aa-4633-a43e-803f6b3ce9a2-must-gather-output\") pod \"must-gather-7px42\" (UID: \"1fd65775-68aa-4633-a43e-803f6b3ce9a2\") " pod="openshift-must-gather-tbjl6/must-gather-7px42" Oct 02 04:29:51 crc kubenswrapper[4775]: I1002 04:29:51.340548 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1fd65775-68aa-4633-a43e-803f6b3ce9a2-must-gather-output\") pod \"must-gather-7px42\" (UID: \"1fd65775-68aa-4633-a43e-803f6b3ce9a2\") " pod="openshift-must-gather-tbjl6/must-gather-7px42" Oct 02 04:29:51 crc kubenswrapper[4775]: I1002 04:29:51.360114 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c9sk\" (UniqueName: \"kubernetes.io/projected/1fd65775-68aa-4633-a43e-803f6b3ce9a2-kube-api-access-9c9sk\") pod \"must-gather-7px42\" (UID: \"1fd65775-68aa-4633-a43e-803f6b3ce9a2\") " pod="openshift-must-gather-tbjl6/must-gather-7px42" Oct 02 04:29:51 crc kubenswrapper[4775]: I1002 04:29:51.443515 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/must-gather-7px42" Oct 02 04:29:51 crc kubenswrapper[4775]: I1002 04:29:51.926851 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-tbjl6/must-gather-7px42"] Oct 02 04:29:52 crc kubenswrapper[4775]: I1002 04:29:52.542878 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tbjl6/must-gather-7px42" event={"ID":"1fd65775-68aa-4633-a43e-803f6b3ce9a2","Type":"ContainerStarted","Data":"9d3154cd67fc773fd88e4919a103235fc97f0ca0d3b2edcbd4b8ff41084cc150"} Oct 02 04:29:52 crc kubenswrapper[4775]: I1002 04:29:52.766340 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:29:52 crc kubenswrapper[4775]: E1002 04:29:52.766818 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:29:58 crc kubenswrapper[4775]: I1002 04:29:58.054496 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:58 crc kubenswrapper[4775]: I1002 04:29:58.124444 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:29:58 crc kubenswrapper[4775]: I1002 04:29:58.804143 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8ct8l"] Oct 02 04:29:59 crc kubenswrapper[4775]: I1002 04:29:59.664870 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8ct8l" podUID="598dc99a-0eb3-4bad-a630-3a44b69775f6" containerName="registry-server" containerID="cri-o://85eca3cb43935ccd119b8a730985b8c72f8e6dc5056613640414220d439c89ef" gracePeriod=2 Oct 02 04:29:59 crc kubenswrapper[4775]: I1002 04:29:59.666362 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tbjl6/must-gather-7px42" event={"ID":"1fd65775-68aa-4633-a43e-803f6b3ce9a2","Type":"ContainerStarted","Data":"9cef27e990f26117f18d7d535484f057592574392b3d9425a31d9314fb5fb18f"} Oct 02 04:29:59 crc kubenswrapper[4775]: I1002 04:29:59.666411 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tbjl6/must-gather-7px42" event={"ID":"1fd65775-68aa-4633-a43e-803f6b3ce9a2","Type":"ContainerStarted","Data":"0878ae6e198eed664c32026ddeb8a8c0b26826d38c05a377ae2bb85df0ff5182"} Oct 02 04:29:59 crc kubenswrapper[4775]: I1002 04:29:59.704660 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tbjl6/must-gather-7px42" podStartSLOduration=1.932510144 podStartE2EDuration="8.704638702s" podCreationTimestamp="2025-10-02 04:29:51 +0000 UTC" firstStartedPulling="2025-10-02 04:29:51.932221908 +0000 UTC m=+10129.098965948" lastFinishedPulling="2025-10-02 04:29:58.704350466 +0000 UTC m=+10135.871094506" observedRunningTime="2025-10-02 04:29:59.69378723 +0000 UTC m=+10136.860531340" watchObservedRunningTime="2025-10-02 04:29:59.704638702 +0000 UTC m=+10136.871382762" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.198665 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv"] Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.200747 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.209468 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv"] Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.209517 4775 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.209877 4775 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.281891 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcghx\" (UniqueName: \"kubernetes.io/projected/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-kube-api-access-xcghx\") pod \"collect-profiles-29322990-xppvv\" (UID: \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.281950 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-secret-volume\") pod \"collect-profiles-29322990-xppvv\" (UID: \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.281987 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-config-volume\") pod \"collect-profiles-29322990-xppvv\" (UID: \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.385202 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-config-volume\") pod \"collect-profiles-29322990-xppvv\" (UID: \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.385402 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcghx\" (UniqueName: \"kubernetes.io/projected/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-kube-api-access-xcghx\") pod \"collect-profiles-29322990-xppvv\" (UID: \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.385441 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-secret-volume\") pod \"collect-profiles-29322990-xppvv\" (UID: \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.386827 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-config-volume\") pod \"collect-profiles-29322990-xppvv\" (UID: \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.391903 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-secret-volume\") pod \"collect-profiles-29322990-xppvv\" (UID: \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.404548 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcghx\" (UniqueName: \"kubernetes.io/projected/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-kube-api-access-xcghx\") pod \"collect-profiles-29322990-xppvv\" (UID: \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.469435 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.486990 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/598dc99a-0eb3-4bad-a630-3a44b69775f6-catalog-content\") pod \"598dc99a-0eb3-4bad-a630-3a44b69775f6\" (UID: \"598dc99a-0eb3-4bad-a630-3a44b69775f6\") " Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.487300 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/598dc99a-0eb3-4bad-a630-3a44b69775f6-utilities\") pod \"598dc99a-0eb3-4bad-a630-3a44b69775f6\" (UID: \"598dc99a-0eb3-4bad-a630-3a44b69775f6\") " Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.487568 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdps4\" (UniqueName: \"kubernetes.io/projected/598dc99a-0eb3-4bad-a630-3a44b69775f6-kube-api-access-kdps4\") pod \"598dc99a-0eb3-4bad-a630-3a44b69775f6\" (UID: \"598dc99a-0eb3-4bad-a630-3a44b69775f6\") " Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.488487 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/598dc99a-0eb3-4bad-a630-3a44b69775f6-utilities" (OuterVolumeSpecName: "utilities") pod "598dc99a-0eb3-4bad-a630-3a44b69775f6" (UID: "598dc99a-0eb3-4bad-a630-3a44b69775f6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.491871 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/598dc99a-0eb3-4bad-a630-3a44b69775f6-kube-api-access-kdps4" (OuterVolumeSpecName: "kube-api-access-kdps4") pod "598dc99a-0eb3-4bad-a630-3a44b69775f6" (UID: "598dc99a-0eb3-4bad-a630-3a44b69775f6"). InnerVolumeSpecName "kube-api-access-kdps4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.556345 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.564270 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/598dc99a-0eb3-4bad-a630-3a44b69775f6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "598dc99a-0eb3-4bad-a630-3a44b69775f6" (UID: "598dc99a-0eb3-4bad-a630-3a44b69775f6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.589645 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdps4\" (UniqueName: \"kubernetes.io/projected/598dc99a-0eb3-4bad-a630-3a44b69775f6-kube-api-access-kdps4\") on node \"crc\" DevicePath \"\"" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.589677 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/598dc99a-0eb3-4bad-a630-3a44b69775f6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.589686 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/598dc99a-0eb3-4bad-a630-3a44b69775f6-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.678539 4775 generic.go:334] "Generic (PLEG): container finished" podID="598dc99a-0eb3-4bad-a630-3a44b69775f6" containerID="85eca3cb43935ccd119b8a730985b8c72f8e6dc5056613640414220d439c89ef" exitCode=0 Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.678575 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ct8l" event={"ID":"598dc99a-0eb3-4bad-a630-3a44b69775f6","Type":"ContainerDied","Data":"85eca3cb43935ccd119b8a730985b8c72f8e6dc5056613640414220d439c89ef"} Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.678611 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8ct8l" event={"ID":"598dc99a-0eb3-4bad-a630-3a44b69775f6","Type":"ContainerDied","Data":"4aa8063e8df91980724ba409ea58cc9440840385447e2fee03899f847cd7b990"} Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.678627 4775 scope.go:117] "RemoveContainer" containerID="85eca3cb43935ccd119b8a730985b8c72f8e6dc5056613640414220d439c89ef" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.678639 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8ct8l" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.712370 4775 scope.go:117] "RemoveContainer" containerID="d182b535e96bc26a023cd841415af05fe1863bd3f2846673acffb11e8904142a" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.722878 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8ct8l"] Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.733924 4775 scope.go:117] "RemoveContainer" containerID="3e27b9819c998d1ec835b1df1f152840a70a92cbde9b68a07faf0627868df84d" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.741030 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8ct8l"] Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.764185 4775 scope.go:117] "RemoveContainer" containerID="85eca3cb43935ccd119b8a730985b8c72f8e6dc5056613640414220d439c89ef" Oct 02 04:30:00 crc kubenswrapper[4775]: E1002 04:30:00.768005 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85eca3cb43935ccd119b8a730985b8c72f8e6dc5056613640414220d439c89ef\": container with ID starting with 85eca3cb43935ccd119b8a730985b8c72f8e6dc5056613640414220d439c89ef not found: ID does not exist" containerID="85eca3cb43935ccd119b8a730985b8c72f8e6dc5056613640414220d439c89ef" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.768037 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85eca3cb43935ccd119b8a730985b8c72f8e6dc5056613640414220d439c89ef"} err="failed to get container status \"85eca3cb43935ccd119b8a730985b8c72f8e6dc5056613640414220d439c89ef\": rpc error: code = NotFound desc = could not find container \"85eca3cb43935ccd119b8a730985b8c72f8e6dc5056613640414220d439c89ef\": container with ID starting with 85eca3cb43935ccd119b8a730985b8c72f8e6dc5056613640414220d439c89ef not found: ID does not exist" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.768058 4775 scope.go:117] "RemoveContainer" containerID="d182b535e96bc26a023cd841415af05fe1863bd3f2846673acffb11e8904142a" Oct 02 04:30:00 crc kubenswrapper[4775]: E1002 04:30:00.768531 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d182b535e96bc26a023cd841415af05fe1863bd3f2846673acffb11e8904142a\": container with ID starting with d182b535e96bc26a023cd841415af05fe1863bd3f2846673acffb11e8904142a not found: ID does not exist" containerID="d182b535e96bc26a023cd841415af05fe1863bd3f2846673acffb11e8904142a" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.768574 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d182b535e96bc26a023cd841415af05fe1863bd3f2846673acffb11e8904142a"} err="failed to get container status \"d182b535e96bc26a023cd841415af05fe1863bd3f2846673acffb11e8904142a\": rpc error: code = NotFound desc = could not find container \"d182b535e96bc26a023cd841415af05fe1863bd3f2846673acffb11e8904142a\": container with ID starting with d182b535e96bc26a023cd841415af05fe1863bd3f2846673acffb11e8904142a not found: ID does not exist" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.768604 4775 scope.go:117] "RemoveContainer" containerID="3e27b9819c998d1ec835b1df1f152840a70a92cbde9b68a07faf0627868df84d" Oct 02 04:30:00 crc kubenswrapper[4775]: E1002 04:30:00.768918 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e27b9819c998d1ec835b1df1f152840a70a92cbde9b68a07faf0627868df84d\": container with ID starting with 3e27b9819c998d1ec835b1df1f152840a70a92cbde9b68a07faf0627868df84d not found: ID does not exist" containerID="3e27b9819c998d1ec835b1df1f152840a70a92cbde9b68a07faf0627868df84d" Oct 02 04:30:00 crc kubenswrapper[4775]: I1002 04:30:00.768937 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e27b9819c998d1ec835b1df1f152840a70a92cbde9b68a07faf0627868df84d"} err="failed to get container status \"3e27b9819c998d1ec835b1df1f152840a70a92cbde9b68a07faf0627868df84d\": rpc error: code = NotFound desc = could not find container \"3e27b9819c998d1ec835b1df1f152840a70a92cbde9b68a07faf0627868df84d\": container with ID starting with 3e27b9819c998d1ec835b1df1f152840a70a92cbde9b68a07faf0627868df84d not found: ID does not exist" Oct 02 04:30:01 crc kubenswrapper[4775]: I1002 04:30:01.117917 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv"] Oct 02 04:30:01 crc kubenswrapper[4775]: I1002 04:30:01.694307 4775 generic.go:334] "Generic (PLEG): container finished" podID="7be5a764-5041-47ea-a0e1-5c85a0f8a05f" containerID="56469ed290b1317047355c432e397e1b1350169b6e135f389ca6875d0a665290" exitCode=0 Oct 02 04:30:01 crc kubenswrapper[4775]: I1002 04:30:01.694373 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" event={"ID":"7be5a764-5041-47ea-a0e1-5c85a0f8a05f","Type":"ContainerDied","Data":"56469ed290b1317047355c432e397e1b1350169b6e135f389ca6875d0a665290"} Oct 02 04:30:01 crc kubenswrapper[4775]: I1002 04:30:01.694622 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" event={"ID":"7be5a764-5041-47ea-a0e1-5c85a0f8a05f","Type":"ContainerStarted","Data":"c027f0a466d450c791c41962b6833e159d3f698b0601d12af75bd8813b85a85c"} Oct 02 04:30:01 crc kubenswrapper[4775]: I1002 04:30:01.779986 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="598dc99a-0eb3-4bad-a630-3a44b69775f6" path="/var/lib/kubelet/pods/598dc99a-0eb3-4bad-a630-3a44b69775f6/volumes" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.081610 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.260513 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-config-volume\") pod \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\" (UID: \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\") " Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.260889 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcghx\" (UniqueName: \"kubernetes.io/projected/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-kube-api-access-xcghx\") pod \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\" (UID: \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\") " Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.261143 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-secret-volume\") pod \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\" (UID: \"7be5a764-5041-47ea-a0e1-5c85a0f8a05f\") " Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.261539 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-config-volume" (OuterVolumeSpecName: "config-volume") pod "7be5a764-5041-47ea-a0e1-5c85a0f8a05f" (UID: "7be5a764-5041-47ea-a0e1-5c85a0f8a05f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.266615 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7be5a764-5041-47ea-a0e1-5c85a0f8a05f" (UID: "7be5a764-5041-47ea-a0e1-5c85a0f8a05f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.267324 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-kube-api-access-xcghx" (OuterVolumeSpecName: "kube-api-access-xcghx") pod "7be5a764-5041-47ea-a0e1-5c85a0f8a05f" (UID: "7be5a764-5041-47ea-a0e1-5c85a0f8a05f"). InnerVolumeSpecName "kube-api-access-xcghx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.363401 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcghx\" (UniqueName: \"kubernetes.io/projected/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-kube-api-access-xcghx\") on node \"crc\" DevicePath \"\"" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.363446 4775 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.363459 4775 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7be5a764-5041-47ea-a0e1-5c85a0f8a05f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.714020 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" event={"ID":"7be5a764-5041-47ea-a0e1-5c85a0f8a05f","Type":"ContainerDied","Data":"c027f0a466d450c791c41962b6833e159d3f698b0601d12af75bd8813b85a85c"} Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.714310 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c027f0a466d450c791c41962b6833e159d3f698b0601d12af75bd8813b85a85c" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.714068 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29322990-xppvv" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.811157 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tbjl6/crc-debug-mvw6j"] Oct 02 04:30:03 crc kubenswrapper[4775]: E1002 04:30:03.811724 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="598dc99a-0eb3-4bad-a630-3a44b69775f6" containerName="registry-server" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.811748 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="598dc99a-0eb3-4bad-a630-3a44b69775f6" containerName="registry-server" Oct 02 04:30:03 crc kubenswrapper[4775]: E1002 04:30:03.811767 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="598dc99a-0eb3-4bad-a630-3a44b69775f6" containerName="extract-utilities" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.811778 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="598dc99a-0eb3-4bad-a630-3a44b69775f6" containerName="extract-utilities" Oct 02 04:30:03 crc kubenswrapper[4775]: E1002 04:30:03.811804 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7be5a764-5041-47ea-a0e1-5c85a0f8a05f" containerName="collect-profiles" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.811814 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="7be5a764-5041-47ea-a0e1-5c85a0f8a05f" containerName="collect-profiles" Oct 02 04:30:03 crc kubenswrapper[4775]: E1002 04:30:03.811839 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="598dc99a-0eb3-4bad-a630-3a44b69775f6" containerName="extract-content" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.811847 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="598dc99a-0eb3-4bad-a630-3a44b69775f6" containerName="extract-content" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.812174 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="598dc99a-0eb3-4bad-a630-3a44b69775f6" containerName="registry-server" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.812205 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="7be5a764-5041-47ea-a0e1-5c85a0f8a05f" containerName="collect-profiles" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.813361 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/crc-debug-mvw6j" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.877621 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh9r4\" (UniqueName: \"kubernetes.io/projected/2d8d41fb-1836-400e-88df-df77b07e17be-kube-api-access-wh9r4\") pod \"crc-debug-mvw6j\" (UID: \"2d8d41fb-1836-400e-88df-df77b07e17be\") " pod="openshift-must-gather-tbjl6/crc-debug-mvw6j" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.877688 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d8d41fb-1836-400e-88df-df77b07e17be-host\") pod \"crc-debug-mvw6j\" (UID: \"2d8d41fb-1836-400e-88df-df77b07e17be\") " pod="openshift-must-gather-tbjl6/crc-debug-mvw6j" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.979909 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh9r4\" (UniqueName: \"kubernetes.io/projected/2d8d41fb-1836-400e-88df-df77b07e17be-kube-api-access-wh9r4\") pod \"crc-debug-mvw6j\" (UID: \"2d8d41fb-1836-400e-88df-df77b07e17be\") " pod="openshift-must-gather-tbjl6/crc-debug-mvw6j" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.979988 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d8d41fb-1836-400e-88df-df77b07e17be-host\") pod \"crc-debug-mvw6j\" (UID: \"2d8d41fb-1836-400e-88df-df77b07e17be\") " pod="openshift-must-gather-tbjl6/crc-debug-mvw6j" Oct 02 04:30:03 crc kubenswrapper[4775]: I1002 04:30:03.980223 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d8d41fb-1836-400e-88df-df77b07e17be-host\") pod \"crc-debug-mvw6j\" (UID: \"2d8d41fb-1836-400e-88df-df77b07e17be\") " pod="openshift-must-gather-tbjl6/crc-debug-mvw6j" Oct 02 04:30:04 crc kubenswrapper[4775]: I1002 04:30:03.995676 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh9r4\" (UniqueName: \"kubernetes.io/projected/2d8d41fb-1836-400e-88df-df77b07e17be-kube-api-access-wh9r4\") pod \"crc-debug-mvw6j\" (UID: \"2d8d41fb-1836-400e-88df-df77b07e17be\") " pod="openshift-must-gather-tbjl6/crc-debug-mvw6j" Oct 02 04:30:04 crc kubenswrapper[4775]: I1002 04:30:04.136492 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/crc-debug-mvw6j" Oct 02 04:30:04 crc kubenswrapper[4775]: I1002 04:30:04.155584 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x"] Oct 02 04:30:04 crc kubenswrapper[4775]: I1002 04:30:04.168273 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29322945-bds7x"] Oct 02 04:30:04 crc kubenswrapper[4775]: I1002 04:30:04.723217 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tbjl6/crc-debug-mvw6j" event={"ID":"2d8d41fb-1836-400e-88df-df77b07e17be","Type":"ContainerStarted","Data":"59864500e018b96dbc7e574fc96c6f40026b67d88dce34d86581244cef0bb05c"} Oct 02 04:30:04 crc kubenswrapper[4775]: I1002 04:30:04.765895 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:30:04 crc kubenswrapper[4775]: E1002 04:30:04.766240 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:30:05 crc kubenswrapper[4775]: I1002 04:30:05.781415 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa9a4ec4-fd61-4b68-839f-b617df1204eb" path="/var/lib/kubelet/pods/fa9a4ec4-fd61-4b68-839f-b617df1204eb/volumes" Oct 02 04:30:15 crc kubenswrapper[4775]: I1002 04:30:15.837769 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tbjl6/crc-debug-mvw6j" event={"ID":"2d8d41fb-1836-400e-88df-df77b07e17be","Type":"ContainerStarted","Data":"2dc5066f3e291b5d8dd85bbc925f1cf783fc80c68b73889cbab9f6f71117bf3e"} Oct 02 04:30:15 crc kubenswrapper[4775]: I1002 04:30:15.870165 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-tbjl6/crc-debug-mvw6j" podStartSLOduration=2.346151049 podStartE2EDuration="12.870139441s" podCreationTimestamp="2025-10-02 04:30:03 +0000 UTC" firstStartedPulling="2025-10-02 04:30:04.211604833 +0000 UTC m=+10141.378348863" lastFinishedPulling="2025-10-02 04:30:14.735593215 +0000 UTC m=+10151.902337255" observedRunningTime="2025-10-02 04:30:15.854529555 +0000 UTC m=+10153.021273635" watchObservedRunningTime="2025-10-02 04:30:15.870139441 +0000 UTC m=+10153.036883511" Oct 02 04:30:16 crc kubenswrapper[4775]: I1002 04:30:16.765940 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:30:17 crc kubenswrapper[4775]: I1002 04:30:17.860824 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"f16b9f11307f31ca81e98942662cc69d6baccc40acb87cced0a5c54c35b4a97c"} Oct 02 04:30:19 crc kubenswrapper[4775]: I1002 04:30:19.365721 4775 scope.go:117] "RemoveContainer" containerID="0b8651815abbac281863662f159ddfd32dd0b8b07d1c687d2a16eca3dec0622d" Oct 02 04:30:37 crc kubenswrapper[4775]: I1002 04:30:37.541102 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vj9r5"] Oct 02 04:30:37 crc kubenswrapper[4775]: I1002 04:30:37.544621 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:37 crc kubenswrapper[4775]: I1002 04:30:37.560408 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vj9r5"] Oct 02 04:30:37 crc kubenswrapper[4775]: I1002 04:30:37.683662 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dxh8\" (UniqueName: \"kubernetes.io/projected/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-kube-api-access-4dxh8\") pod \"community-operators-vj9r5\" (UID: \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\") " pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:37 crc kubenswrapper[4775]: I1002 04:30:37.683715 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-utilities\") pod \"community-operators-vj9r5\" (UID: \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\") " pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:37 crc kubenswrapper[4775]: I1002 04:30:37.683868 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-catalog-content\") pod \"community-operators-vj9r5\" (UID: \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\") " pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:37 crc kubenswrapper[4775]: I1002 04:30:37.792319 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-catalog-content\") pod \"community-operators-vj9r5\" (UID: \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\") " pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:37 crc kubenswrapper[4775]: I1002 04:30:37.792413 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dxh8\" (UniqueName: \"kubernetes.io/projected/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-kube-api-access-4dxh8\") pod \"community-operators-vj9r5\" (UID: \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\") " pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:37 crc kubenswrapper[4775]: I1002 04:30:37.792448 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-utilities\") pod \"community-operators-vj9r5\" (UID: \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\") " pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:37 crc kubenswrapper[4775]: I1002 04:30:37.794798 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-utilities\") pod \"community-operators-vj9r5\" (UID: \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\") " pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:37 crc kubenswrapper[4775]: I1002 04:30:37.796035 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-catalog-content\") pod \"community-operators-vj9r5\" (UID: \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\") " pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:37 crc kubenswrapper[4775]: I1002 04:30:37.831505 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dxh8\" (UniqueName: \"kubernetes.io/projected/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-kube-api-access-4dxh8\") pod \"community-operators-vj9r5\" (UID: \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\") " pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:37 crc kubenswrapper[4775]: I1002 04:30:37.889789 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:38 crc kubenswrapper[4775]: I1002 04:30:38.443252 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vj9r5"] Oct 02 04:30:39 crc kubenswrapper[4775]: I1002 04:30:39.104693 4775 generic.go:334] "Generic (PLEG): container finished" podID="29a8a9e6-8652-44ef-b563-f5c9538fc8d5" containerID="3d0ff6bd6f0941fd9bc759c7dd0d22a79574d1bd2a1d53d1ccbade858dc16067" exitCode=0 Oct 02 04:30:39 crc kubenswrapper[4775]: I1002 04:30:39.104767 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj9r5" event={"ID":"29a8a9e6-8652-44ef-b563-f5c9538fc8d5","Type":"ContainerDied","Data":"3d0ff6bd6f0941fd9bc759c7dd0d22a79574d1bd2a1d53d1ccbade858dc16067"} Oct 02 04:30:39 crc kubenswrapper[4775]: I1002 04:30:39.105117 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj9r5" event={"ID":"29a8a9e6-8652-44ef-b563-f5c9538fc8d5","Type":"ContainerStarted","Data":"6f327be1b70b4d210eafcb23ad1a115afc013b5868340a3b9b91221d4e36e533"} Oct 02 04:30:41 crc kubenswrapper[4775]: I1002 04:30:41.127636 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj9r5" event={"ID":"29a8a9e6-8652-44ef-b563-f5c9538fc8d5","Type":"ContainerStarted","Data":"d766f6f6bfbd38f462305f096949a3fb0eacdc1afae079cde8d3a74241780f80"} Oct 02 04:30:42 crc kubenswrapper[4775]: I1002 04:30:42.141231 4775 generic.go:334] "Generic (PLEG): container finished" podID="29a8a9e6-8652-44ef-b563-f5c9538fc8d5" containerID="d766f6f6bfbd38f462305f096949a3fb0eacdc1afae079cde8d3a74241780f80" exitCode=0 Oct 02 04:30:42 crc kubenswrapper[4775]: I1002 04:30:42.141475 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj9r5" event={"ID":"29a8a9e6-8652-44ef-b563-f5c9538fc8d5","Type":"ContainerDied","Data":"d766f6f6bfbd38f462305f096949a3fb0eacdc1afae079cde8d3a74241780f80"} Oct 02 04:30:43 crc kubenswrapper[4775]: I1002 04:30:43.153556 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj9r5" event={"ID":"29a8a9e6-8652-44ef-b563-f5c9538fc8d5","Type":"ContainerStarted","Data":"d0a3a68ed3d7563f1f32d64c072c468556c0d81abea7e2765691f0169526b69f"} Oct 02 04:30:43 crc kubenswrapper[4775]: I1002 04:30:43.181652 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vj9r5" podStartSLOduration=2.672711344 podStartE2EDuration="6.181625646s" podCreationTimestamp="2025-10-02 04:30:37 +0000 UTC" firstStartedPulling="2025-10-02 04:30:39.106563301 +0000 UTC m=+10176.273307341" lastFinishedPulling="2025-10-02 04:30:42.615477603 +0000 UTC m=+10179.782221643" observedRunningTime="2025-10-02 04:30:43.17908519 +0000 UTC m=+10180.345829260" watchObservedRunningTime="2025-10-02 04:30:43.181625646 +0000 UTC m=+10180.348369706" Oct 02 04:30:47 crc kubenswrapper[4775]: I1002 04:30:47.890114 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:47 crc kubenswrapper[4775]: I1002 04:30:47.890485 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:47 crc kubenswrapper[4775]: I1002 04:30:47.950933 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:48 crc kubenswrapper[4775]: I1002 04:30:48.253516 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:48 crc kubenswrapper[4775]: I1002 04:30:48.309861 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vj9r5"] Oct 02 04:30:50 crc kubenswrapper[4775]: I1002 04:30:50.223498 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vj9r5" podUID="29a8a9e6-8652-44ef-b563-f5c9538fc8d5" containerName="registry-server" containerID="cri-o://d0a3a68ed3d7563f1f32d64c072c468556c0d81abea7e2765691f0169526b69f" gracePeriod=2 Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.152814 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.244846 4775 generic.go:334] "Generic (PLEG): container finished" podID="29a8a9e6-8652-44ef-b563-f5c9538fc8d5" containerID="d0a3a68ed3d7563f1f32d64c072c468556c0d81abea7e2765691f0169526b69f" exitCode=0 Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.244901 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj9r5" event={"ID":"29a8a9e6-8652-44ef-b563-f5c9538fc8d5","Type":"ContainerDied","Data":"d0a3a68ed3d7563f1f32d64c072c468556c0d81abea7e2765691f0169526b69f"} Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.244943 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vj9r5" event={"ID":"29a8a9e6-8652-44ef-b563-f5c9538fc8d5","Type":"ContainerDied","Data":"6f327be1b70b4d210eafcb23ad1a115afc013b5868340a3b9b91221d4e36e533"} Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.244982 4775 scope.go:117] "RemoveContainer" containerID="d0a3a68ed3d7563f1f32d64c072c468556c0d81abea7e2765691f0169526b69f" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.245151 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vj9r5" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.270025 4775 scope.go:117] "RemoveContainer" containerID="d766f6f6bfbd38f462305f096949a3fb0eacdc1afae079cde8d3a74241780f80" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.285911 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-utilities\") pod \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\" (UID: \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\") " Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.286268 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dxh8\" (UniqueName: \"kubernetes.io/projected/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-kube-api-access-4dxh8\") pod \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\" (UID: \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\") " Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.286802 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-catalog-content\") pod \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\" (UID: \"29a8a9e6-8652-44ef-b563-f5c9538fc8d5\") " Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.287459 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-utilities" (OuterVolumeSpecName: "utilities") pod "29a8a9e6-8652-44ef-b563-f5c9538fc8d5" (UID: "29a8a9e6-8652-44ef-b563-f5c9538fc8d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.296686 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-kube-api-access-4dxh8" (OuterVolumeSpecName: "kube-api-access-4dxh8") pod "29a8a9e6-8652-44ef-b563-f5c9538fc8d5" (UID: "29a8a9e6-8652-44ef-b563-f5c9538fc8d5"). InnerVolumeSpecName "kube-api-access-4dxh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.301400 4775 scope.go:117] "RemoveContainer" containerID="3d0ff6bd6f0941fd9bc759c7dd0d22a79574d1bd2a1d53d1ccbade858dc16067" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.340081 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "29a8a9e6-8652-44ef-b563-f5c9538fc8d5" (UID: "29a8a9e6-8652-44ef-b563-f5c9538fc8d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.381992 4775 scope.go:117] "RemoveContainer" containerID="d0a3a68ed3d7563f1f32d64c072c468556c0d81abea7e2765691f0169526b69f" Oct 02 04:30:51 crc kubenswrapper[4775]: E1002 04:30:51.383769 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0a3a68ed3d7563f1f32d64c072c468556c0d81abea7e2765691f0169526b69f\": container with ID starting with d0a3a68ed3d7563f1f32d64c072c468556c0d81abea7e2765691f0169526b69f not found: ID does not exist" containerID="d0a3a68ed3d7563f1f32d64c072c468556c0d81abea7e2765691f0169526b69f" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.384110 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0a3a68ed3d7563f1f32d64c072c468556c0d81abea7e2765691f0169526b69f"} err="failed to get container status \"d0a3a68ed3d7563f1f32d64c072c468556c0d81abea7e2765691f0169526b69f\": rpc error: code = NotFound desc = could not find container \"d0a3a68ed3d7563f1f32d64c072c468556c0d81abea7e2765691f0169526b69f\": container with ID starting with d0a3a68ed3d7563f1f32d64c072c468556c0d81abea7e2765691f0169526b69f not found: ID does not exist" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.384251 4775 scope.go:117] "RemoveContainer" containerID="d766f6f6bfbd38f462305f096949a3fb0eacdc1afae079cde8d3a74241780f80" Oct 02 04:30:51 crc kubenswrapper[4775]: E1002 04:30:51.384781 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d766f6f6bfbd38f462305f096949a3fb0eacdc1afae079cde8d3a74241780f80\": container with ID starting with d766f6f6bfbd38f462305f096949a3fb0eacdc1afae079cde8d3a74241780f80 not found: ID does not exist" containerID="d766f6f6bfbd38f462305f096949a3fb0eacdc1afae079cde8d3a74241780f80" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.384825 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d766f6f6bfbd38f462305f096949a3fb0eacdc1afae079cde8d3a74241780f80"} err="failed to get container status \"d766f6f6bfbd38f462305f096949a3fb0eacdc1afae079cde8d3a74241780f80\": rpc error: code = NotFound desc = could not find container \"d766f6f6bfbd38f462305f096949a3fb0eacdc1afae079cde8d3a74241780f80\": container with ID starting with d766f6f6bfbd38f462305f096949a3fb0eacdc1afae079cde8d3a74241780f80 not found: ID does not exist" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.384851 4775 scope.go:117] "RemoveContainer" containerID="3d0ff6bd6f0941fd9bc759c7dd0d22a79574d1bd2a1d53d1ccbade858dc16067" Oct 02 04:30:51 crc kubenswrapper[4775]: E1002 04:30:51.385176 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d0ff6bd6f0941fd9bc759c7dd0d22a79574d1bd2a1d53d1ccbade858dc16067\": container with ID starting with 3d0ff6bd6f0941fd9bc759c7dd0d22a79574d1bd2a1d53d1ccbade858dc16067 not found: ID does not exist" containerID="3d0ff6bd6f0941fd9bc759c7dd0d22a79574d1bd2a1d53d1ccbade858dc16067" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.385196 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d0ff6bd6f0941fd9bc759c7dd0d22a79574d1bd2a1d53d1ccbade858dc16067"} err="failed to get container status \"3d0ff6bd6f0941fd9bc759c7dd0d22a79574d1bd2a1d53d1ccbade858dc16067\": rpc error: code = NotFound desc = could not find container \"3d0ff6bd6f0941fd9bc759c7dd0d22a79574d1bd2a1d53d1ccbade858dc16067\": container with ID starting with 3d0ff6bd6f0941fd9bc759c7dd0d22a79574d1bd2a1d53d1ccbade858dc16067 not found: ID does not exist" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.389491 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dxh8\" (UniqueName: \"kubernetes.io/projected/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-kube-api-access-4dxh8\") on node \"crc\" DevicePath \"\"" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.389517 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.389529 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/29a8a9e6-8652-44ef-b563-f5c9538fc8d5-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.587035 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vj9r5"] Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.609728 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vj9r5"] Oct 02 04:30:51 crc kubenswrapper[4775]: I1002 04:30:51.783578 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29a8a9e6-8652-44ef-b563-f5c9538fc8d5" path="/var/lib/kubelet/pods/29a8a9e6-8652-44ef-b563-f5c9538fc8d5/volumes" Oct 02 04:31:30 crc kubenswrapper[4775]: I1002 04:31:30.371646 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_2e0082f9-da1f-4b8d-8177-aebe807d9a1f/init-config-reloader/0.log" Oct 02 04:31:30 crc kubenswrapper[4775]: I1002 04:31:30.540353 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_2e0082f9-da1f-4b8d-8177-aebe807d9a1f/init-config-reloader/0.log" Oct 02 04:31:30 crc kubenswrapper[4775]: I1002 04:31:30.572254 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_2e0082f9-da1f-4b8d-8177-aebe807d9a1f/alertmanager/0.log" Oct 02 04:31:30 crc kubenswrapper[4775]: I1002 04:31:30.738023 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_2e0082f9-da1f-4b8d-8177-aebe807d9a1f/config-reloader/0.log" Oct 02 04:31:30 crc kubenswrapper[4775]: I1002 04:31:30.925110 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7bb83c5f-2641-421a-ac8d-35e94398b856/aodh-api/0.log" Oct 02 04:31:31 crc kubenswrapper[4775]: I1002 04:31:31.014580 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7bb83c5f-2641-421a-ac8d-35e94398b856/aodh-evaluator/0.log" Oct 02 04:31:31 crc kubenswrapper[4775]: I1002 04:31:31.125855 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7bb83c5f-2641-421a-ac8d-35e94398b856/aodh-listener/0.log" Oct 02 04:31:31 crc kubenswrapper[4775]: I1002 04:31:31.209986 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_7bb83c5f-2641-421a-ac8d-35e94398b856/aodh-notifier/0.log" Oct 02 04:31:31 crc kubenswrapper[4775]: I1002 04:31:31.413331 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5f5d675bc4-lw5dk_7463fd90-1ce1-484f-b3ca-499e93e46f8e/barbican-api/0.log" Oct 02 04:31:31 crc kubenswrapper[4775]: I1002 04:31:31.601495 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5f5d675bc4-lw5dk_7463fd90-1ce1-484f-b3ca-499e93e46f8e/barbican-api-log/0.log" Oct 02 04:31:31 crc kubenswrapper[4775]: I1002 04:31:31.805273 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b5587698d-j95nw_445125c9-802a-4759-ab11-11845f420062/barbican-keystone-listener/0.log" Oct 02 04:31:31 crc kubenswrapper[4775]: I1002 04:31:31.884115 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6b5587698d-j95nw_445125c9-802a-4759-ab11-11845f420062/barbican-keystone-listener-log/0.log" Oct 02 04:31:32 crc kubenswrapper[4775]: I1002 04:31:32.080719 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6f755bf8ff-wb6kc_f069d67d-c773-4393-ac3d-ef520b69355b/barbican-worker/0.log" Oct 02 04:31:32 crc kubenswrapper[4775]: I1002 04:31:32.260889 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6f755bf8ff-wb6kc_f069d67d-c773-4393-ac3d-ef520b69355b/barbican-worker-log/0.log" Oct 02 04:31:32 crc kubenswrapper[4775]: I1002 04:31:32.480021 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-rldqh_5032d81f-b10d-4bf0-89f1-a6ee328ceaec/bootstrap-openstack-openstack-cell1/0.log" Oct 02 04:31:32 crc kubenswrapper[4775]: I1002 04:31:32.647875 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c7172345-ba1a-4856-8b88-009379970b34/ceilometer-central-agent/0.log" Oct 02 04:31:32 crc kubenswrapper[4775]: I1002 04:31:32.694323 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c7172345-ba1a-4856-8b88-009379970b34/ceilometer-notification-agent/0.log" Oct 02 04:31:32 crc kubenswrapper[4775]: I1002 04:31:32.839577 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c7172345-ba1a-4856-8b88-009379970b34/proxy-httpd/0.log" Oct 02 04:31:32 crc kubenswrapper[4775]: I1002 04:31:32.879939 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c7172345-ba1a-4856-8b88-009379970b34/sg-core/0.log" Oct 02 04:31:33 crc kubenswrapper[4775]: I1002 04:31:33.060555 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-r82b9_bb77ca06-d802-41b4-98f3-0cf0c6b5bb58/ceph-client-openstack-openstack-cell1/0.log" Oct 02 04:31:33 crc kubenswrapper[4775]: I1002 04:31:33.257581 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_88a07e5c-e647-4a80-9397-fc2cc1f54fcd/cinder-api/0.log" Oct 02 04:31:33 crc kubenswrapper[4775]: I1002 04:31:33.280449 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_88a07e5c-e647-4a80-9397-fc2cc1f54fcd/cinder-api-log/0.log" Oct 02 04:31:33 crc kubenswrapper[4775]: I1002 04:31:33.525490 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_dc936ddc-e6af-4e0a-85b0-732eddb8ca82/probe/0.log" Oct 02 04:31:33 crc kubenswrapper[4775]: I1002 04:31:33.608113 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_dc936ddc-e6af-4e0a-85b0-732eddb8ca82/cinder-backup/0.log" Oct 02 04:31:33 crc kubenswrapper[4775]: I1002 04:31:33.757692 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_abd6be45-e458-4dcf-94b1-39545386cfe2/cinder-scheduler/0.log" Oct 02 04:31:33 crc kubenswrapper[4775]: I1002 04:31:33.831391 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_abd6be45-e458-4dcf-94b1-39545386cfe2/probe/0.log" Oct 02 04:31:33 crc kubenswrapper[4775]: I1002 04:31:33.974403 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_e763e57d-3848-40c3-ac6f-29987d3070cf/cinder-volume/0.log" Oct 02 04:31:34 crc kubenswrapper[4775]: I1002 04:31:34.019174 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_e763e57d-3848-40c3-ac6f-29987d3070cf/probe/0.log" Oct 02 04:31:34 crc kubenswrapper[4775]: I1002 04:31:34.180392 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-x58ln_e3dc106c-ef5c-4394-91ee-d0ba9cf0c1da/configure-network-openstack-openstack-cell1/0.log" Oct 02 04:31:34 crc kubenswrapper[4775]: I1002 04:31:34.301916 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-gsmzl_1e299556-d363-499a-9aec-6f1241060e4a/configure-os-openstack-openstack-cell1/0.log" Oct 02 04:31:34 crc kubenswrapper[4775]: I1002 04:31:34.680367 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-54d96fb68f-shvgf_4808d138-0235-4e3e-97fe-8362aa73d26f/init/0.log" Oct 02 04:31:34 crc kubenswrapper[4775]: I1002 04:31:34.878854 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-54d96fb68f-shvgf_4808d138-0235-4e3e-97fe-8362aa73d26f/init/0.log" Oct 02 04:31:34 crc kubenswrapper[4775]: I1002 04:31:34.888602 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-54d96fb68f-shvgf_4808d138-0235-4e3e-97fe-8362aa73d26f/dnsmasq-dns/0.log" Oct 02 04:31:35 crc kubenswrapper[4775]: I1002 04:31:35.065808 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-5mg6v_e15f9838-d5f3-4b6c-b968-de51d37ffaf9/download-cache-openstack-openstack-cell1/0.log" Oct 02 04:31:35 crc kubenswrapper[4775]: I1002 04:31:35.192664 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_08ba6c24-402e-4dc1-b71a-47cb167ac6f2/glance-httpd/0.log" Oct 02 04:31:35 crc kubenswrapper[4775]: I1002 04:31:35.287439 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_08ba6c24-402e-4dc1-b71a-47cb167ac6f2/glance-log/0.log" Oct 02 04:31:35 crc kubenswrapper[4775]: I1002 04:31:35.402532 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c541a988-a6cf-4625-bc69-5d114ee4aaa8/glance-httpd/0.log" Oct 02 04:31:35 crc kubenswrapper[4775]: I1002 04:31:35.470920 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_c541a988-a6cf-4625-bc69-5d114ee4aaa8/glance-log/0.log" Oct 02 04:31:36 crc kubenswrapper[4775]: I1002 04:31:36.319931 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-9fcb5b9c4-kpksr_654406cb-d5cc-46d0-9319-cd678ba45b72/heat-cfnapi/0.log" Oct 02 04:31:36 crc kubenswrapper[4775]: I1002 04:31:36.434204 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-d7c784464-67wzm_37f06ca5-4d12-4f82-8403-3fc12c9bb1f0/heat-api/0.log" Oct 02 04:31:36 crc kubenswrapper[4775]: I1002 04:31:36.545870 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-68fb455f9-b7mpg_d843cd4a-9895-4385-90ca-6d9f91954adc/heat-engine/0.log" Oct 02 04:31:36 crc kubenswrapper[4775]: I1002 04:31:36.722204 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-598f77fd5-6jh6q_2b046c0b-b2f9-4808-84fa-5474dd3f6d19/horizon/0.log" Oct 02 04:31:36 crc kubenswrapper[4775]: I1002 04:31:36.834168 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-598f77fd5-6jh6q_2b046c0b-b2f9-4808-84fa-5474dd3f6d19/horizon-log/0.log" Oct 02 04:31:36 crc kubenswrapper[4775]: I1002 04:31:36.881941 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-shjt8_5e735a83-11ab-4b9d-b97c-a60174da0899/install-certs-openstack-openstack-cell1/0.log" Oct 02 04:31:37 crc kubenswrapper[4775]: I1002 04:31:37.078103 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-9jghw_d998f892-b126-406a-bfad-568524759afd/install-os-openstack-openstack-cell1/0.log" Oct 02 04:31:37 crc kubenswrapper[4775]: I1002 04:31:37.517445 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6959f9548-mcs66_4161a715-fa2c-452b-9259-ce6a0e165533/keystone-api/0.log" Oct 02 04:31:37 crc kubenswrapper[4775]: I1002 04:31:37.942873 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29322961-tcwf8_05e4796e-9130-4382-abe0-939e36fd4f1b/keystone-cron/0.log" Oct 02 04:31:38 crc kubenswrapper[4775]: I1002 04:31:38.219464 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_3fa0fce1-781d-41a9-8aaf-8c0589df9cf8/kube-state-metrics/0.log" Oct 02 04:31:38 crc kubenswrapper[4775]: I1002 04:31:38.292145 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-csddm_9c88413d-ac75-4d79-9d76-54e4b1c8c1fc/libvirt-openstack-openstack-cell1/0.log" Oct 02 04:31:38 crc kubenswrapper[4775]: I1002 04:31:38.464062 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_e902e4c2-791d-489a-b2b7-f2e2ed851356/manila-api/0.log" Oct 02 04:31:38 crc kubenswrapper[4775]: I1002 04:31:38.509280 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_e902e4c2-791d-489a-b2b7-f2e2ed851356/manila-api-log/0.log" Oct 02 04:31:38 crc kubenswrapper[4775]: I1002 04:31:38.720137 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_2faaa667-8d44-47aa-ab2f-27b1d49e6d70/manila-scheduler/0.log" Oct 02 04:31:38 crc kubenswrapper[4775]: I1002 04:31:38.727719 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_2faaa667-8d44-47aa-ab2f-27b1d49e6d70/probe/0.log" Oct 02 04:31:38 crc kubenswrapper[4775]: I1002 04:31:38.906398 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_6355bb29-85fc-474d-8370-af3339ff22c8/probe/0.log" Oct 02 04:31:38 crc kubenswrapper[4775]: I1002 04:31:38.946246 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_6355bb29-85fc-474d-8370-af3339ff22c8/manila-share/0.log" Oct 02 04:31:39 crc kubenswrapper[4775]: I1002 04:31:39.116659 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-copy-data_3f39e43d-dd89-41c5-9725-ffc4011d49e7/adoption/0.log" Oct 02 04:31:39 crc kubenswrapper[4775]: I1002 04:31:39.523304 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-776c686f69-g7qtq_aca10bd9-aa40-47a8-a461-c0bfc1fa4638/neutron-api/0.log" Oct 02 04:31:39 crc kubenswrapper[4775]: I1002 04:31:39.614619 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-776c686f69-g7qtq_aca10bd9-aa40-47a8-a461-c0bfc1fa4638/neutron-httpd/0.log" Oct 02 04:31:39 crc kubenswrapper[4775]: I1002 04:31:39.907169 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-p97fw_2258fa83-dbaf-4438-a9bf-42bb3458fa98/neutron-dhcp-openstack-openstack-cell1/0.log" Oct 02 04:31:40 crc kubenswrapper[4775]: I1002 04:31:40.193231 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-2zw6m_93ccc1b8-b113-45a6-a69f-4c9b9e6052c5/neutron-metadata-openstack-openstack-cell1/0.log" Oct 02 04:31:40 crc kubenswrapper[4775]: I1002 04:31:40.451732 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-qjjxk_4b83b696-fe3c-4696-a9ed-f30bbd0e1ba5/neutron-sriov-openstack-openstack-cell1/0.log" Oct 02 04:31:40 crc kubenswrapper[4775]: I1002 04:31:40.766343 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_20d47d00-0c95-4858-963c-ac98a3b56306/nova-api-api/0.log" Oct 02 04:31:40 crc kubenswrapper[4775]: I1002 04:31:40.807679 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_20d47d00-0c95-4858-963c-ac98a3b56306/nova-api-log/0.log" Oct 02 04:31:41 crc kubenswrapper[4775]: I1002 04:31:41.056289 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_1dbfa9c9-66fa-48ad-956f-52297602546e/nova-cell0-conductor-conductor/0.log" Oct 02 04:31:41 crc kubenswrapper[4775]: I1002 04:31:41.352495 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_4a5e074b-a9a5-4ddb-a167-d7fee4fd8d84/nova-cell1-conductor-conductor/0.log" Oct 02 04:31:41 crc kubenswrapper[4775]: I1002 04:31:41.651085 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_207f84a9-5284-46d4-856c-ab11901ae6e7/nova-cell1-novncproxy-novncproxy/0.log" Oct 02 04:31:41 crc kubenswrapper[4775]: I1002 04:31:41.976406 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-celldd6hv_b5d0f695-e66e-464e-bf9b-301a66a74f1b/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Oct 02 04:31:42 crc kubenswrapper[4775]: I1002 04:31:42.344773 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-dsmb2_00fb1524-9133-48e4-9fb7-63c91b3cf6fd/nova-cell1-openstack-openstack-cell1/0.log" Oct 02 04:31:42 crc kubenswrapper[4775]: I1002 04:31:42.491209 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_283844d0-a97d-4bc0-b297-5e078862a5b9/nova-metadata-log/0.log" Oct 02 04:31:42 crc kubenswrapper[4775]: I1002 04:31:42.652147 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_283844d0-a97d-4bc0-b297-5e078862a5b9/nova-metadata-metadata/0.log" Oct 02 04:31:42 crc kubenswrapper[4775]: I1002 04:31:42.800766 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_e7850678-5334-43e5-9262-d1442e4959b1/memcached/0.log" Oct 02 04:31:42 crc kubenswrapper[4775]: I1002 04:31:42.918449 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_880173a7-686a-4c2f-8d7a-1b929fcd7c03/nova-scheduler-scheduler/0.log" Oct 02 04:31:42 crc kubenswrapper[4775]: I1002 04:31:42.967318 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-9669b56-5jpsb_e534bc54-984a-4661-9e75-f898b8c78d33/init/0.log" Oct 02 04:31:43 crc kubenswrapper[4775]: I1002 04:31:43.212029 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-9669b56-5jpsb_e534bc54-984a-4661-9e75-f898b8c78d33/init/0.log" Oct 02 04:31:43 crc kubenswrapper[4775]: I1002 04:31:43.287645 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-9669b56-5jpsb_e534bc54-984a-4661-9e75-f898b8c78d33/octavia-api-provider-agent/0.log" Oct 02 04:31:43 crc kubenswrapper[4775]: I1002 04:31:43.458331 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-9669b56-5jpsb_e534bc54-984a-4661-9e75-f898b8c78d33/octavia-api/0.log" Oct 02 04:31:43 crc kubenswrapper[4775]: I1002 04:31:43.481991 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-vrtnf_2c72a379-38b8-4152-80f9-49c83f7b8719/init/0.log" Oct 02 04:31:43 crc kubenswrapper[4775]: I1002 04:31:43.690131 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-vrtnf_2c72a379-38b8-4152-80f9-49c83f7b8719/init/0.log" Oct 02 04:31:43 crc kubenswrapper[4775]: I1002 04:31:43.759341 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-vrtnf_2c72a379-38b8-4152-80f9-49c83f7b8719/octavia-healthmanager/0.log" Oct 02 04:31:43 crc kubenswrapper[4775]: I1002 04:31:43.877117 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-4tlkz_26dc7016-0e75-4919-957a-feac2af12ab0/init/0.log" Oct 02 04:31:44 crc kubenswrapper[4775]: I1002 04:31:44.043189 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-4tlkz_26dc7016-0e75-4919-957a-feac2af12ab0/init/0.log" Oct 02 04:31:44 crc kubenswrapper[4775]: I1002 04:31:44.101758 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-4tlkz_26dc7016-0e75-4919-957a-feac2af12ab0/octavia-housekeeping/0.log" Oct 02 04:31:44 crc kubenswrapper[4775]: I1002 04:31:44.232948 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-dcpms_462e5b8f-d3ae-4073-ba5f-18a1d992b634/init/0.log" Oct 02 04:31:44 crc kubenswrapper[4775]: I1002 04:31:44.398073 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-dcpms_462e5b8f-d3ae-4073-ba5f-18a1d992b634/init/0.log" Oct 02 04:31:44 crc kubenswrapper[4775]: I1002 04:31:44.415290 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-dcpms_462e5b8f-d3ae-4073-ba5f-18a1d992b634/octavia-amphora-httpd/0.log" Oct 02 04:31:44 crc kubenswrapper[4775]: I1002 04:31:44.561727 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-9l9kr_d1cf6eca-9072-43a2-8827-d2e0b90134b5/init/0.log" Oct 02 04:31:44 crc kubenswrapper[4775]: I1002 04:31:44.761558 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-9l9kr_d1cf6eca-9072-43a2-8827-d2e0b90134b5/init/0.log" Oct 02 04:31:44 crc kubenswrapper[4775]: I1002 04:31:44.786380 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-9l9kr_d1cf6eca-9072-43a2-8827-d2e0b90134b5/octavia-rsyslog/0.log" Oct 02 04:31:44 crc kubenswrapper[4775]: I1002 04:31:44.952598 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-crg8r_516737d3-2942-4df1-861e-a4e0dbaccbf7/init/0.log" Oct 02 04:31:45 crc kubenswrapper[4775]: I1002 04:31:45.109116 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-crg8r_516737d3-2942-4df1-861e-a4e0dbaccbf7/init/0.log" Oct 02 04:31:45 crc kubenswrapper[4775]: I1002 04:31:45.221423 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-crg8r_516737d3-2942-4df1-861e-a4e0dbaccbf7/octavia-worker/0.log" Oct 02 04:31:46 crc kubenswrapper[4775]: I1002 04:31:46.311859 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_63eca845-367d-47d2-a693-60b6b52252d3/mysql-bootstrap/0.log" Oct 02 04:31:46 crc kubenswrapper[4775]: I1002 04:31:46.447822 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_63eca845-367d-47d2-a693-60b6b52252d3/mysql-bootstrap/0.log" Oct 02 04:31:46 crc kubenswrapper[4775]: I1002 04:31:46.488749 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_63eca845-367d-47d2-a693-60b6b52252d3/galera/0.log" Oct 02 04:31:46 crc kubenswrapper[4775]: I1002 04:31:46.631508 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608/mysql-bootstrap/0.log" Oct 02 04:31:46 crc kubenswrapper[4775]: I1002 04:31:46.793061 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608/mysql-bootstrap/0.log" Oct 02 04:31:46 crc kubenswrapper[4775]: I1002 04:31:46.837143 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f87e89d2-f8ad-4ce8-bbc3-13a3bc1d6608/galera/0.log" Oct 02 04:31:46 crc kubenswrapper[4775]: I1002 04:31:46.986310 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_d77ec247-8a1e-4882-9017-755408b8dc34/openstackclient/0.log" Oct 02 04:31:47 crc kubenswrapper[4775]: I1002 04:31:47.079188 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-b27p9_bef5bd3c-dc22-4fe7-b499-e8fb17d17742/ovn-controller/0.log" Oct 02 04:31:47 crc kubenswrapper[4775]: I1002 04:31:47.270212 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-gm78r_269c1835-ec87-4102-970c-5073e78cb156/openstack-network-exporter/0.log" Oct 02 04:31:47 crc kubenswrapper[4775]: I1002 04:31:47.560085 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-w6gh2_ea76f5cd-b164-49a8-9da2-30702b113bf3/ovsdb-server-init/0.log" Oct 02 04:31:48 crc kubenswrapper[4775]: I1002 04:31:48.254651 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-w6gh2_ea76f5cd-b164-49a8-9da2-30702b113bf3/ovsdb-server-init/0.log" Oct 02 04:31:48 crc kubenswrapper[4775]: I1002 04:31:48.302231 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-w6gh2_ea76f5cd-b164-49a8-9da2-30702b113bf3/ovs-vswitchd/0.log" Oct 02 04:31:48 crc kubenswrapper[4775]: I1002 04:31:48.327887 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-w6gh2_ea76f5cd-b164-49a8-9da2-30702b113bf3/ovsdb-server/0.log" Oct 02 04:31:48 crc kubenswrapper[4775]: I1002 04:31:48.487529 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-copy-data_4dc66a88-f227-4e53-b079-20480022af30/adoption/0.log" Oct 02 04:31:48 crc kubenswrapper[4775]: I1002 04:31:48.660135 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671/openstack-network-exporter/0.log" Oct 02 04:31:48 crc kubenswrapper[4775]: I1002 04:31:48.713480 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_04bab7f7-0bbf-4cd8-9f93-c0f8f07b4671/ovn-northd/0.log" Oct 02 04:31:48 crc kubenswrapper[4775]: I1002 04:31:48.908985 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-dv978_8654436c-452a-43f6-8932-55b0a85d60d2/ovn-openstack-openstack-cell1/0.log" Oct 02 04:31:48 crc kubenswrapper[4775]: I1002 04:31:48.998281 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_790fc282-a15f-4530-a6e0-20b5e44401ad/openstack-network-exporter/0.log" Oct 02 04:31:49 crc kubenswrapper[4775]: I1002 04:31:49.146216 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_790fc282-a15f-4530-a6e0-20b5e44401ad/ovsdbserver-nb/0.log" Oct 02 04:31:49 crc kubenswrapper[4775]: I1002 04:31:49.231541 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_ddccc962-e478-4e58-b4d3-cfcf109fd0f7/openstack-network-exporter/0.log" Oct 02 04:31:49 crc kubenswrapper[4775]: I1002 04:31:49.346327 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_ddccc962-e478-4e58-b4d3-cfcf109fd0f7/ovsdbserver-nb/0.log" Oct 02 04:31:49 crc kubenswrapper[4775]: I1002 04:31:49.431406 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_730a72f5-678f-4ae4-8008-397b6ba49170/openstack-network-exporter/0.log" Oct 02 04:31:49 crc kubenswrapper[4775]: I1002 04:31:49.606053 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_730a72f5-678f-4ae4-8008-397b6ba49170/ovsdbserver-nb/0.log" Oct 02 04:31:49 crc kubenswrapper[4775]: I1002 04:31:49.721634 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1964d668-d2ef-4f06-9ffc-c49f14eb81c3/openstack-network-exporter/0.log" Oct 02 04:31:49 crc kubenswrapper[4775]: I1002 04:31:49.854412 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_1964d668-d2ef-4f06-9ffc-c49f14eb81c3/ovsdbserver-sb/0.log" Oct 02 04:31:49 crc kubenswrapper[4775]: I1002 04:31:49.905471 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_94de3ec6-4d84-4ef5-a127-284ff4e64ea2/openstack-network-exporter/0.log" Oct 02 04:31:50 crc kubenswrapper[4775]: I1002 04:31:50.070407 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_94de3ec6-4d84-4ef5-a127-284ff4e64ea2/ovsdbserver-sb/0.log" Oct 02 04:31:50 crc kubenswrapper[4775]: I1002 04:31:50.126377 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_b18a1684-4fbd-4a52-95de-8d63ce13b38d/openstack-network-exporter/0.log" Oct 02 04:31:50 crc kubenswrapper[4775]: I1002 04:31:50.244216 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_b18a1684-4fbd-4a52-95de-8d63ce13b38d/ovsdbserver-sb/0.log" Oct 02 04:31:50 crc kubenswrapper[4775]: I1002 04:31:50.413063 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7575468d44-vzv5w_ecdfed0f-2bee-4334-9e88-ea6ad912a1e4/placement-api/0.log" Oct 02 04:31:50 crc kubenswrapper[4775]: I1002 04:31:50.520408 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7575468d44-vzv5w_ecdfed0f-2bee-4334-9e88-ea6ad912a1e4/placement-log/0.log" Oct 02 04:31:50 crc kubenswrapper[4775]: I1002 04:31:50.636254 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-clcqdp_e6b88bf2-a076-4957-8cd5-7777afcceead/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Oct 02 04:31:50 crc kubenswrapper[4775]: I1002 04:31:50.792856 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_73818738-ba3d-43f6-b0d1-9c96fd768406/init-config-reloader/0.log" Oct 02 04:31:50 crc kubenswrapper[4775]: I1002 04:31:50.969789 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_73818738-ba3d-43f6-b0d1-9c96fd768406/config-reloader/0.log" Oct 02 04:31:50 crc kubenswrapper[4775]: I1002 04:31:50.989297 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_73818738-ba3d-43f6-b0d1-9c96fd768406/prometheus/0.log" Oct 02 04:31:51 crc kubenswrapper[4775]: I1002 04:31:51.034065 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_73818738-ba3d-43f6-b0d1-9c96fd768406/init-config-reloader/0.log" Oct 02 04:31:51 crc kubenswrapper[4775]: I1002 04:31:51.169720 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_73818738-ba3d-43f6-b0d1-9c96fd768406/thanos-sidecar/0.log" Oct 02 04:31:51 crc kubenswrapper[4775]: I1002 04:31:51.255234 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf4059c4-4a05-4598-bf74-3e1584829b7b/setup-container/0.log" Oct 02 04:31:51 crc kubenswrapper[4775]: I1002 04:31:51.403006 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf4059c4-4a05-4598-bf74-3e1584829b7b/setup-container/0.log" Oct 02 04:31:51 crc kubenswrapper[4775]: I1002 04:31:51.439318 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_bf4059c4-4a05-4598-bf74-3e1584829b7b/rabbitmq/0.log" Oct 02 04:31:51 crc kubenswrapper[4775]: I1002 04:31:51.849095 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_88fede09-aa59-48c1-9957-4ce95fcb95d9/setup-container/0.log" Oct 02 04:31:52 crc kubenswrapper[4775]: I1002 04:31:52.033885 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_88fede09-aa59-48c1-9957-4ce95fcb95d9/setup-container/0.log" Oct 02 04:31:52 crc kubenswrapper[4775]: I1002 04:31:52.047629 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_88fede09-aa59-48c1-9957-4ce95fcb95d9/rabbitmq/0.log" Oct 02 04:31:52 crc kubenswrapper[4775]: I1002 04:31:52.228659 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-x4bp2_986a4be9-d34f-4204-b1cc-bb7e7ad629ec/reboot-os-openstack-openstack-cell1/0.log" Oct 02 04:31:52 crc kubenswrapper[4775]: I1002 04:31:52.307905 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-679nv_79906cc8-51c5-4aaa-a552-15d2770dc621/run-os-openstack-openstack-cell1/0.log" Oct 02 04:31:52 crc kubenswrapper[4775]: I1002 04:31:52.475364 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-bzmkw_cb54d64b-0c87-401c-9e60-ea2c16ca0bfd/ssh-known-hosts-openstack/0.log" Oct 02 04:31:52 crc kubenswrapper[4775]: I1002 04:31:52.680558 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-6vllg_3d34ca5c-4e8e-441f-b1b6-a8ffeb95ab5d/telemetry-openstack-openstack-cell1/0.log" Oct 02 04:31:52 crc kubenswrapper[4775]: I1002 04:31:52.906480 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-7z87m_3959dab0-5e51-4cb7-a540-96ae7c7a28b9/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Oct 02 04:31:52 crc kubenswrapper[4775]: I1002 04:31:52.939919 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-qt4nc_663e6cdb-d635-4083-bf0a-7b00e61dc1ff/validate-network-openstack-openstack-cell1/0.log" Oct 02 04:32:20 crc kubenswrapper[4775]: I1002 04:32:20.207310 4775 generic.go:334] "Generic (PLEG): container finished" podID="2d8d41fb-1836-400e-88df-df77b07e17be" containerID="2dc5066f3e291b5d8dd85bbc925f1cf783fc80c68b73889cbab9f6f71117bf3e" exitCode=0 Oct 02 04:32:20 crc kubenswrapper[4775]: I1002 04:32:20.207426 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tbjl6/crc-debug-mvw6j" event={"ID":"2d8d41fb-1836-400e-88df-df77b07e17be","Type":"ContainerDied","Data":"2dc5066f3e291b5d8dd85bbc925f1cf783fc80c68b73889cbab9f6f71117bf3e"} Oct 02 04:32:21 crc kubenswrapper[4775]: I1002 04:32:21.604604 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/crc-debug-mvw6j" Oct 02 04:32:21 crc kubenswrapper[4775]: I1002 04:32:21.654203 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tbjl6/crc-debug-mvw6j"] Oct 02 04:32:21 crc kubenswrapper[4775]: I1002 04:32:21.666265 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tbjl6/crc-debug-mvw6j"] Oct 02 04:32:21 crc kubenswrapper[4775]: I1002 04:32:21.682431 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d8d41fb-1836-400e-88df-df77b07e17be-host\") pod \"2d8d41fb-1836-400e-88df-df77b07e17be\" (UID: \"2d8d41fb-1836-400e-88df-df77b07e17be\") " Oct 02 04:32:21 crc kubenswrapper[4775]: I1002 04:32:21.682846 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d8d41fb-1836-400e-88df-df77b07e17be-host" (OuterVolumeSpecName: "host") pod "2d8d41fb-1836-400e-88df-df77b07e17be" (UID: "2d8d41fb-1836-400e-88df-df77b07e17be"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 04:32:21 crc kubenswrapper[4775]: I1002 04:32:21.683075 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh9r4\" (UniqueName: \"kubernetes.io/projected/2d8d41fb-1836-400e-88df-df77b07e17be-kube-api-access-wh9r4\") pod \"2d8d41fb-1836-400e-88df-df77b07e17be\" (UID: \"2d8d41fb-1836-400e-88df-df77b07e17be\") " Oct 02 04:32:21 crc kubenswrapper[4775]: I1002 04:32:21.684346 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2d8d41fb-1836-400e-88df-df77b07e17be-host\") on node \"crc\" DevicePath \"\"" Oct 02 04:32:21 crc kubenswrapper[4775]: I1002 04:32:21.695296 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d8d41fb-1836-400e-88df-df77b07e17be-kube-api-access-wh9r4" (OuterVolumeSpecName: "kube-api-access-wh9r4") pod "2d8d41fb-1836-400e-88df-df77b07e17be" (UID: "2d8d41fb-1836-400e-88df-df77b07e17be"). InnerVolumeSpecName "kube-api-access-wh9r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:32:21 crc kubenswrapper[4775]: I1002 04:32:21.780740 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d8d41fb-1836-400e-88df-df77b07e17be" path="/var/lib/kubelet/pods/2d8d41fb-1836-400e-88df-df77b07e17be/volumes" Oct 02 04:32:21 crc kubenswrapper[4775]: I1002 04:32:21.786524 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh9r4\" (UniqueName: \"kubernetes.io/projected/2d8d41fb-1836-400e-88df-df77b07e17be-kube-api-access-wh9r4\") on node \"crc\" DevicePath \"\"" Oct 02 04:32:22 crc kubenswrapper[4775]: I1002 04:32:22.239167 4775 scope.go:117] "RemoveContainer" containerID="2dc5066f3e291b5d8dd85bbc925f1cf783fc80c68b73889cbab9f6f71117bf3e" Oct 02 04:32:22 crc kubenswrapper[4775]: I1002 04:32:22.239234 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/crc-debug-mvw6j" Oct 02 04:32:22 crc kubenswrapper[4775]: I1002 04:32:22.883067 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tbjl6/crc-debug-b8chk"] Oct 02 04:32:22 crc kubenswrapper[4775]: E1002 04:32:22.884003 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a8a9e6-8652-44ef-b563-f5c9538fc8d5" containerName="registry-server" Oct 02 04:32:22 crc kubenswrapper[4775]: I1002 04:32:22.884024 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a8a9e6-8652-44ef-b563-f5c9538fc8d5" containerName="registry-server" Oct 02 04:32:22 crc kubenswrapper[4775]: E1002 04:32:22.884074 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a8a9e6-8652-44ef-b563-f5c9538fc8d5" containerName="extract-content" Oct 02 04:32:22 crc kubenswrapper[4775]: I1002 04:32:22.884085 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a8a9e6-8652-44ef-b563-f5c9538fc8d5" containerName="extract-content" Oct 02 04:32:22 crc kubenswrapper[4775]: E1002 04:32:22.884112 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d8d41fb-1836-400e-88df-df77b07e17be" containerName="container-00" Oct 02 04:32:22 crc kubenswrapper[4775]: I1002 04:32:22.884123 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d8d41fb-1836-400e-88df-df77b07e17be" containerName="container-00" Oct 02 04:32:22 crc kubenswrapper[4775]: E1002 04:32:22.884156 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29a8a9e6-8652-44ef-b563-f5c9538fc8d5" containerName="extract-utilities" Oct 02 04:32:22 crc kubenswrapper[4775]: I1002 04:32:22.884167 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="29a8a9e6-8652-44ef-b563-f5c9538fc8d5" containerName="extract-utilities" Oct 02 04:32:22 crc kubenswrapper[4775]: I1002 04:32:22.884478 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d8d41fb-1836-400e-88df-df77b07e17be" containerName="container-00" Oct 02 04:32:22 crc kubenswrapper[4775]: I1002 04:32:22.884519 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="29a8a9e6-8652-44ef-b563-f5c9538fc8d5" containerName="registry-server" Oct 02 04:32:22 crc kubenswrapper[4775]: I1002 04:32:22.885665 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/crc-debug-b8chk" Oct 02 04:32:22 crc kubenswrapper[4775]: I1002 04:32:22.911192 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz8lz\" (UniqueName: \"kubernetes.io/projected/30a078fe-5166-4d1a-9540-3c74f8a627c1-kube-api-access-kz8lz\") pod \"crc-debug-b8chk\" (UID: \"30a078fe-5166-4d1a-9540-3c74f8a627c1\") " pod="openshift-must-gather-tbjl6/crc-debug-b8chk" Oct 02 04:32:22 crc kubenswrapper[4775]: I1002 04:32:22.911668 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/30a078fe-5166-4d1a-9540-3c74f8a627c1-host\") pod \"crc-debug-b8chk\" (UID: \"30a078fe-5166-4d1a-9540-3c74f8a627c1\") " pod="openshift-must-gather-tbjl6/crc-debug-b8chk" Oct 02 04:32:23 crc kubenswrapper[4775]: I1002 04:32:23.015073 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz8lz\" (UniqueName: \"kubernetes.io/projected/30a078fe-5166-4d1a-9540-3c74f8a627c1-kube-api-access-kz8lz\") pod \"crc-debug-b8chk\" (UID: \"30a078fe-5166-4d1a-9540-3c74f8a627c1\") " pod="openshift-must-gather-tbjl6/crc-debug-b8chk" Oct 02 04:32:23 crc kubenswrapper[4775]: I1002 04:32:23.015163 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/30a078fe-5166-4d1a-9540-3c74f8a627c1-host\") pod \"crc-debug-b8chk\" (UID: \"30a078fe-5166-4d1a-9540-3c74f8a627c1\") " pod="openshift-must-gather-tbjl6/crc-debug-b8chk" Oct 02 04:32:23 crc kubenswrapper[4775]: I1002 04:32:23.015648 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/30a078fe-5166-4d1a-9540-3c74f8a627c1-host\") pod \"crc-debug-b8chk\" (UID: \"30a078fe-5166-4d1a-9540-3c74f8a627c1\") " pod="openshift-must-gather-tbjl6/crc-debug-b8chk" Oct 02 04:32:23 crc kubenswrapper[4775]: I1002 04:32:23.039442 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz8lz\" (UniqueName: \"kubernetes.io/projected/30a078fe-5166-4d1a-9540-3c74f8a627c1-kube-api-access-kz8lz\") pod \"crc-debug-b8chk\" (UID: \"30a078fe-5166-4d1a-9540-3c74f8a627c1\") " pod="openshift-must-gather-tbjl6/crc-debug-b8chk" Oct 02 04:32:23 crc kubenswrapper[4775]: I1002 04:32:23.210872 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/crc-debug-b8chk" Oct 02 04:32:24 crc kubenswrapper[4775]: I1002 04:32:24.291869 4775 generic.go:334] "Generic (PLEG): container finished" podID="30a078fe-5166-4d1a-9540-3c74f8a627c1" containerID="41a876660cb21d50b93e7ffd5d4cffc46f1f6e23e7605909f55164ab2ac0d2b4" exitCode=0 Oct 02 04:32:24 crc kubenswrapper[4775]: I1002 04:32:24.292019 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tbjl6/crc-debug-b8chk" event={"ID":"30a078fe-5166-4d1a-9540-3c74f8a627c1","Type":"ContainerDied","Data":"41a876660cb21d50b93e7ffd5d4cffc46f1f6e23e7605909f55164ab2ac0d2b4"} Oct 02 04:32:24 crc kubenswrapper[4775]: I1002 04:32:24.292553 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tbjl6/crc-debug-b8chk" event={"ID":"30a078fe-5166-4d1a-9540-3c74f8a627c1","Type":"ContainerStarted","Data":"a787b45bbb6217f2ecf31f28c49ca71c7622b0269e8f7f98aaf627bc6ac4f514"} Oct 02 04:32:25 crc kubenswrapper[4775]: I1002 04:32:25.413016 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/crc-debug-b8chk" Oct 02 04:32:25 crc kubenswrapper[4775]: I1002 04:32:25.490095 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/30a078fe-5166-4d1a-9540-3c74f8a627c1-host\") pod \"30a078fe-5166-4d1a-9540-3c74f8a627c1\" (UID: \"30a078fe-5166-4d1a-9540-3c74f8a627c1\") " Oct 02 04:32:25 crc kubenswrapper[4775]: I1002 04:32:25.490216 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/30a078fe-5166-4d1a-9540-3c74f8a627c1-host" (OuterVolumeSpecName: "host") pod "30a078fe-5166-4d1a-9540-3c74f8a627c1" (UID: "30a078fe-5166-4d1a-9540-3c74f8a627c1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 04:32:25 crc kubenswrapper[4775]: I1002 04:32:25.490255 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kz8lz\" (UniqueName: \"kubernetes.io/projected/30a078fe-5166-4d1a-9540-3c74f8a627c1-kube-api-access-kz8lz\") pod \"30a078fe-5166-4d1a-9540-3c74f8a627c1\" (UID: \"30a078fe-5166-4d1a-9540-3c74f8a627c1\") " Oct 02 04:32:25 crc kubenswrapper[4775]: I1002 04:32:25.491055 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/30a078fe-5166-4d1a-9540-3c74f8a627c1-host\") on node \"crc\" DevicePath \"\"" Oct 02 04:32:25 crc kubenswrapper[4775]: I1002 04:32:25.497210 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30a078fe-5166-4d1a-9540-3c74f8a627c1-kube-api-access-kz8lz" (OuterVolumeSpecName: "kube-api-access-kz8lz") pod "30a078fe-5166-4d1a-9540-3c74f8a627c1" (UID: "30a078fe-5166-4d1a-9540-3c74f8a627c1"). InnerVolumeSpecName "kube-api-access-kz8lz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:32:25 crc kubenswrapper[4775]: I1002 04:32:25.591942 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kz8lz\" (UniqueName: \"kubernetes.io/projected/30a078fe-5166-4d1a-9540-3c74f8a627c1-kube-api-access-kz8lz\") on node \"crc\" DevicePath \"\"" Oct 02 04:32:26 crc kubenswrapper[4775]: I1002 04:32:26.317415 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tbjl6/crc-debug-b8chk" event={"ID":"30a078fe-5166-4d1a-9540-3c74f8a627c1","Type":"ContainerDied","Data":"a787b45bbb6217f2ecf31f28c49ca71c7622b0269e8f7f98aaf627bc6ac4f514"} Oct 02 04:32:26 crc kubenswrapper[4775]: I1002 04:32:26.317679 4775 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a787b45bbb6217f2ecf31f28c49ca71c7622b0269e8f7f98aaf627bc6ac4f514" Oct 02 04:32:26 crc kubenswrapper[4775]: I1002 04:32:26.317731 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/crc-debug-b8chk" Oct 02 04:32:35 crc kubenswrapper[4775]: I1002 04:32:35.199820 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tbjl6/crc-debug-b8chk"] Oct 02 04:32:35 crc kubenswrapper[4775]: I1002 04:32:35.208236 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tbjl6/crc-debug-b8chk"] Oct 02 04:32:35 crc kubenswrapper[4775]: I1002 04:32:35.788502 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30a078fe-5166-4d1a-9540-3c74f8a627c1" path="/var/lib/kubelet/pods/30a078fe-5166-4d1a-9540-3c74f8a627c1/volumes" Oct 02 04:32:36 crc kubenswrapper[4775]: I1002 04:32:36.420835 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-tbjl6/crc-debug-78w7z"] Oct 02 04:32:36 crc kubenswrapper[4775]: E1002 04:32:36.421531 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a078fe-5166-4d1a-9540-3c74f8a627c1" containerName="container-00" Oct 02 04:32:36 crc kubenswrapper[4775]: I1002 04:32:36.421544 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a078fe-5166-4d1a-9540-3c74f8a627c1" containerName="container-00" Oct 02 04:32:36 crc kubenswrapper[4775]: I1002 04:32:36.421790 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a078fe-5166-4d1a-9540-3c74f8a627c1" containerName="container-00" Oct 02 04:32:36 crc kubenswrapper[4775]: I1002 04:32:36.422467 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/crc-debug-78w7z" Oct 02 04:32:36 crc kubenswrapper[4775]: I1002 04:32:36.492208 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp9s7\" (UniqueName: \"kubernetes.io/projected/3619ed82-06f8-42e6-b7c6-7b3ab832580b-kube-api-access-bp9s7\") pod \"crc-debug-78w7z\" (UID: \"3619ed82-06f8-42e6-b7c6-7b3ab832580b\") " pod="openshift-must-gather-tbjl6/crc-debug-78w7z" Oct 02 04:32:36 crc kubenswrapper[4775]: I1002 04:32:36.492577 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3619ed82-06f8-42e6-b7c6-7b3ab832580b-host\") pod \"crc-debug-78w7z\" (UID: \"3619ed82-06f8-42e6-b7c6-7b3ab832580b\") " pod="openshift-must-gather-tbjl6/crc-debug-78w7z" Oct 02 04:32:36 crc kubenswrapper[4775]: I1002 04:32:36.593682 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp9s7\" (UniqueName: \"kubernetes.io/projected/3619ed82-06f8-42e6-b7c6-7b3ab832580b-kube-api-access-bp9s7\") pod \"crc-debug-78w7z\" (UID: \"3619ed82-06f8-42e6-b7c6-7b3ab832580b\") " pod="openshift-must-gather-tbjl6/crc-debug-78w7z" Oct 02 04:32:36 crc kubenswrapper[4775]: I1002 04:32:36.593808 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3619ed82-06f8-42e6-b7c6-7b3ab832580b-host\") pod \"crc-debug-78w7z\" (UID: \"3619ed82-06f8-42e6-b7c6-7b3ab832580b\") " pod="openshift-must-gather-tbjl6/crc-debug-78w7z" Oct 02 04:32:36 crc kubenswrapper[4775]: I1002 04:32:36.594074 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3619ed82-06f8-42e6-b7c6-7b3ab832580b-host\") pod \"crc-debug-78w7z\" (UID: \"3619ed82-06f8-42e6-b7c6-7b3ab832580b\") " pod="openshift-must-gather-tbjl6/crc-debug-78w7z" Oct 02 04:32:36 crc kubenswrapper[4775]: I1002 04:32:36.615783 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp9s7\" (UniqueName: \"kubernetes.io/projected/3619ed82-06f8-42e6-b7c6-7b3ab832580b-kube-api-access-bp9s7\") pod \"crc-debug-78w7z\" (UID: \"3619ed82-06f8-42e6-b7c6-7b3ab832580b\") " pod="openshift-must-gather-tbjl6/crc-debug-78w7z" Oct 02 04:32:36 crc kubenswrapper[4775]: I1002 04:32:36.760254 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/crc-debug-78w7z" Oct 02 04:32:37 crc kubenswrapper[4775]: I1002 04:32:37.234271 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:32:37 crc kubenswrapper[4775]: I1002 04:32:37.234689 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:32:37 crc kubenswrapper[4775]: I1002 04:32:37.484823 4775 generic.go:334] "Generic (PLEG): container finished" podID="3619ed82-06f8-42e6-b7c6-7b3ab832580b" containerID="cc1c03f2918b45138f5a72a8200ab07a247f40e7be5b6c11d045c5bc479f6c30" exitCode=0 Oct 02 04:32:37 crc kubenswrapper[4775]: I1002 04:32:37.484888 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tbjl6/crc-debug-78w7z" event={"ID":"3619ed82-06f8-42e6-b7c6-7b3ab832580b","Type":"ContainerDied","Data":"cc1c03f2918b45138f5a72a8200ab07a247f40e7be5b6c11d045c5bc479f6c30"} Oct 02 04:32:37 crc kubenswrapper[4775]: I1002 04:32:37.484929 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tbjl6/crc-debug-78w7z" event={"ID":"3619ed82-06f8-42e6-b7c6-7b3ab832580b","Type":"ContainerStarted","Data":"d669f58948d2ef21ef0f0ff5edb4d1c3bf05eb0a5475d2d06bb37d71656e8112"} Oct 02 04:32:37 crc kubenswrapper[4775]: I1002 04:32:37.541049 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tbjl6/crc-debug-78w7z"] Oct 02 04:32:37 crc kubenswrapper[4775]: I1002 04:32:37.556593 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tbjl6/crc-debug-78w7z"] Oct 02 04:32:38 crc kubenswrapper[4775]: I1002 04:32:38.639656 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/crc-debug-78w7z" Oct 02 04:32:38 crc kubenswrapper[4775]: I1002 04:32:38.778511 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3619ed82-06f8-42e6-b7c6-7b3ab832580b-host\") pod \"3619ed82-06f8-42e6-b7c6-7b3ab832580b\" (UID: \"3619ed82-06f8-42e6-b7c6-7b3ab832580b\") " Oct 02 04:32:38 crc kubenswrapper[4775]: I1002 04:32:38.778706 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bp9s7\" (UniqueName: \"kubernetes.io/projected/3619ed82-06f8-42e6-b7c6-7b3ab832580b-kube-api-access-bp9s7\") pod \"3619ed82-06f8-42e6-b7c6-7b3ab832580b\" (UID: \"3619ed82-06f8-42e6-b7c6-7b3ab832580b\") " Oct 02 04:32:38 crc kubenswrapper[4775]: I1002 04:32:38.778763 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3619ed82-06f8-42e6-b7c6-7b3ab832580b-host" (OuterVolumeSpecName: "host") pod "3619ed82-06f8-42e6-b7c6-7b3ab832580b" (UID: "3619ed82-06f8-42e6-b7c6-7b3ab832580b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 02 04:32:38 crc kubenswrapper[4775]: I1002 04:32:38.779515 4775 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3619ed82-06f8-42e6-b7c6-7b3ab832580b-host\") on node \"crc\" DevicePath \"\"" Oct 02 04:32:38 crc kubenswrapper[4775]: I1002 04:32:38.796620 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3619ed82-06f8-42e6-b7c6-7b3ab832580b-kube-api-access-bp9s7" (OuterVolumeSpecName: "kube-api-access-bp9s7") pod "3619ed82-06f8-42e6-b7c6-7b3ab832580b" (UID: "3619ed82-06f8-42e6-b7c6-7b3ab832580b"). InnerVolumeSpecName "kube-api-access-bp9s7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:32:38 crc kubenswrapper[4775]: I1002 04:32:38.882490 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp9s7\" (UniqueName: \"kubernetes.io/projected/3619ed82-06f8-42e6-b7c6-7b3ab832580b-kube-api-access-bp9s7\") on node \"crc\" DevicePath \"\"" Oct 02 04:32:39 crc kubenswrapper[4775]: I1002 04:32:39.513025 4775 scope.go:117] "RemoveContainer" containerID="cc1c03f2918b45138f5a72a8200ab07a247f40e7be5b6c11d045c5bc479f6c30" Oct 02 04:32:39 crc kubenswrapper[4775]: I1002 04:32:39.513072 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/crc-debug-78w7z" Oct 02 04:32:39 crc kubenswrapper[4775]: I1002 04:32:39.786239 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3619ed82-06f8-42e6-b7c6-7b3ab832580b" path="/var/lib/kubelet/pods/3619ed82-06f8-42e6-b7c6-7b3ab832580b/volumes" Oct 02 04:33:07 crc kubenswrapper[4775]: I1002 04:33:07.234473 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:33:07 crc kubenswrapper[4775]: I1002 04:33:07.235032 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:33:27 crc kubenswrapper[4775]: I1002 04:33:27.008058 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-5wrms_c2757728-9f90-4b12-8a40-3a7845c4f461/kube-rbac-proxy/0.log" Oct 02 04:33:27 crc kubenswrapper[4775]: I1002 04:33:27.106177 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-6ff8b75857-5wrms_c2757728-9f90-4b12-8a40-3a7845c4f461/manager/0.log" Oct 02 04:33:27 crc kubenswrapper[4775]: I1002 04:33:27.199821 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm_790fe7be-506d-4eb7-b754-8aa3e2856236/util/0.log" Oct 02 04:33:27 crc kubenswrapper[4775]: I1002 04:33:27.318356 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm_790fe7be-506d-4eb7-b754-8aa3e2856236/util/0.log" Oct 02 04:33:27 crc kubenswrapper[4775]: I1002 04:33:27.318810 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm_790fe7be-506d-4eb7-b754-8aa3e2856236/pull/0.log" Oct 02 04:33:27 crc kubenswrapper[4775]: I1002 04:33:27.352233 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm_790fe7be-506d-4eb7-b754-8aa3e2856236/pull/0.log" Oct 02 04:33:27 crc kubenswrapper[4775]: I1002 04:33:27.522906 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm_790fe7be-506d-4eb7-b754-8aa3e2856236/util/0.log" Oct 02 04:33:27 crc kubenswrapper[4775]: I1002 04:33:27.560659 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm_790fe7be-506d-4eb7-b754-8aa3e2856236/pull/0.log" Oct 02 04:33:27 crc kubenswrapper[4775]: I1002 04:33:27.581636 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cbbec88b66c61e3d10a0bbc868e336160e79c262c68f9747aec27f06595wrgm_790fe7be-506d-4eb7-b754-8aa3e2856236/extract/0.log" Oct 02 04:33:27 crc kubenswrapper[4775]: I1002 04:33:27.690827 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-mx8pj_63b4a9a2-62d3-48a8-a2ca-74496dcff908/kube-rbac-proxy/0.log" Oct 02 04:33:27 crc kubenswrapper[4775]: I1002 04:33:27.792217 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-644bddb6d8-mx8pj_63b4a9a2-62d3-48a8-a2ca-74496dcff908/manager/0.log" Oct 02 04:33:27 crc kubenswrapper[4775]: I1002 04:33:27.796771 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-4gjk2_7e0f47de-d77c-4840-a382-ed6a911f20b1/kube-rbac-proxy/0.log" Oct 02 04:33:27 crc kubenswrapper[4775]: I1002 04:33:27.898231 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-84f4f7b77b-4gjk2_7e0f47de-d77c-4840-a382-ed6a911f20b1/manager/0.log" Oct 02 04:33:27 crc kubenswrapper[4775]: I1002 04:33:27.967474 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-md78z_68306a2f-6d51-4de9-9645-d3f0e3bc6e61/kube-rbac-proxy/0.log" Oct 02 04:33:28 crc kubenswrapper[4775]: I1002 04:33:28.110016 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84958c4d49-md78z_68306a2f-6d51-4de9-9645-d3f0e3bc6e61/manager/0.log" Oct 02 04:33:28 crc kubenswrapper[4775]: I1002 04:33:28.163840 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-5nl49_c815738f-cb2a-4136-9320-f75ceea923c5/kube-rbac-proxy/0.log" Oct 02 04:33:28 crc kubenswrapper[4775]: I1002 04:33:28.234495 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5d889d78cf-5nl49_c815738f-cb2a-4136-9320-f75ceea923c5/manager/0.log" Oct 02 04:33:28 crc kubenswrapper[4775]: I1002 04:33:28.318441 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-2mxjr_03d2a01f-1564-46f8-9648-e4a826470b44/kube-rbac-proxy/0.log" Oct 02 04:33:28 crc kubenswrapper[4775]: I1002 04:33:28.324983 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-9f4696d94-2mxjr_03d2a01f-1564-46f8-9648-e4a826470b44/manager/0.log" Oct 02 04:33:28 crc kubenswrapper[4775]: I1002 04:33:28.501835 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-jt9hq_03a52fee-7a10-42b3-8316-c16a0dcc01aa/kube-rbac-proxy/0.log" Oct 02 04:33:28 crc kubenswrapper[4775]: I1002 04:33:28.693646 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-f2khq_f1ef22ed-d1cd-4bcc-b399-716552b889b8/kube-rbac-proxy/0.log" Oct 02 04:33:28 crc kubenswrapper[4775]: I1002 04:33:28.715924 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5cd4858477-f2khq_f1ef22ed-d1cd-4bcc-b399-716552b889b8/manager/0.log" Oct 02 04:33:28 crc kubenswrapper[4775]: I1002 04:33:28.732282 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-9d6c5db85-jt9hq_03a52fee-7a10-42b3-8316-c16a0dcc01aa/manager/0.log" Oct 02 04:33:28 crc kubenswrapper[4775]: I1002 04:33:28.871411 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-jvrxr_4dd0da8e-118c-4125-987a-a1122b592002/kube-rbac-proxy/0.log" Oct 02 04:33:28 crc kubenswrapper[4775]: I1002 04:33:28.950867 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-kvhkb_c066e712-5c8d-48f8-b2fe-daa26e9f53e6/kube-rbac-proxy/0.log" Oct 02 04:33:28 crc kubenswrapper[4775]: I1002 04:33:28.993000 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-5bd55b4bff-jvrxr_4dd0da8e-118c-4125-987a-a1122b592002/manager/0.log" Oct 02 04:33:29 crc kubenswrapper[4775]: I1002 04:33:29.127198 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-qkvrn_faa566e8-9b97-4766-a414-4ed0fccf9a81/kube-rbac-proxy/0.log" Oct 02 04:33:29 crc kubenswrapper[4775]: I1002 04:33:29.196906 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-88c7-qkvrn_faa566e8-9b97-4766-a414-4ed0fccf9a81/manager/0.log" Oct 02 04:33:29 crc kubenswrapper[4775]: I1002 04:33:29.201353 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-6d68dbc695-kvhkb_c066e712-5c8d-48f8-b2fe-daa26e9f53e6/manager/0.log" Oct 02 04:33:29 crc kubenswrapper[4775]: I1002 04:33:29.312029 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-2zft5_600073d0-68ea-45f5-87c2-0775377796b5/kube-rbac-proxy/0.log" Oct 02 04:33:29 crc kubenswrapper[4775]: I1002 04:33:29.405944 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-849d5b9b84-2zft5_600073d0-68ea-45f5-87c2-0775377796b5/manager/0.log" Oct 02 04:33:29 crc kubenswrapper[4775]: I1002 04:33:29.513176 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-5vt65_52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2/kube-rbac-proxy/0.log" Oct 02 04:33:29 crc kubenswrapper[4775]: I1002 04:33:29.608464 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-lqrg2_749bf5d6-b257-47e7-b632-e0edf0321adf/kube-rbac-proxy/0.log" Oct 02 04:33:29 crc kubenswrapper[4775]: I1002 04:33:29.746347 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-64cd67b5cb-5vt65_52bcb0eb-a3c9-42a0-9c7e-5bb2ef5fabb2/manager/0.log" Oct 02 04:33:29 crc kubenswrapper[4775]: I1002 04:33:29.801738 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-7b787867f4-lqrg2_749bf5d6-b257-47e7-b632-e0edf0321adf/manager/0.log" Oct 02 04:33:29 crc kubenswrapper[4775]: I1002 04:33:29.810034 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-9qscs_c9d4b689-6ed6-4784-a010-1cea2d49f469/kube-rbac-proxy/0.log" Oct 02 04:33:30 crc kubenswrapper[4775]: I1002 04:33:30.001517 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-5869cb545-9qscs_c9d4b689-6ed6-4784-a010-1cea2d49f469/manager/0.log" Oct 02 04:33:30 crc kubenswrapper[4775]: I1002 04:33:30.053142 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-557f5d867b-2m767_f0c5961b-3c87-4438-b862-e2d1a83b1d6f/kube-rbac-proxy/0.log" Oct 02 04:33:30 crc kubenswrapper[4775]: I1002 04:33:30.277337 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6bcc7488c-bprqn_f08455ed-dbee-4865-885e-c92d82da7457/kube-rbac-proxy/0.log" Oct 02 04:33:30 crc kubenswrapper[4775]: I1002 04:33:30.344561 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-f8jlc_b5b65461-e031-4f60-aa2b-a63a906053c4/registry-server/0.log" Oct 02 04:33:30 crc kubenswrapper[4775]: I1002 04:33:30.491287 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6bcc7488c-bprqn_f08455ed-dbee-4865-885e-c92d82da7457/operator/0.log" Oct 02 04:33:30 crc kubenswrapper[4775]: I1002 04:33:30.564764 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-xhpvs_6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f/kube-rbac-proxy/0.log" Oct 02 04:33:30 crc kubenswrapper[4775]: I1002 04:33:30.738053 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-kv2nm_35c85a80-497b-47f8-afdd-8ad771e4557d/kube-rbac-proxy/0.log" Oct 02 04:33:30 crc kubenswrapper[4775]: I1002 04:33:30.822429 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-589c58c6c-kv2nm_35c85a80-497b-47f8-afdd-8ad771e4557d/manager/0.log" Oct 02 04:33:30 crc kubenswrapper[4775]: I1002 04:33:30.869661 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-9976ff44c-xhpvs_6ac8ba4b-5cd5-41b8-92e1-fabd2543a33f/manager/0.log" Oct 02 04:33:31 crc kubenswrapper[4775]: I1002 04:33:31.064365 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-8fmc6_4175c381-6ab1-4e55-b786-c7b886d69b35/operator/0.log" Oct 02 04:33:31 crc kubenswrapper[4775]: I1002 04:33:31.159997 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-ptq59_4ea65180-a9ee-476e-a260-be2d9f276f8c/kube-rbac-proxy/0.log" Oct 02 04:33:31 crc kubenswrapper[4775]: I1002 04:33:31.312128 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-84d6b4b759-ptq59_4ea65180-a9ee-476e-a260-be2d9f276f8c/manager/0.log" Oct 02 04:33:31 crc kubenswrapper[4775]: I1002 04:33:31.384271 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-vbs7k_c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f/kube-rbac-proxy/0.log" Oct 02 04:33:31 crc kubenswrapper[4775]: I1002 04:33:31.676887 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-bpg5g_c058610a-487c-4f23-b906-b03b0a458534/kube-rbac-proxy/0.log" Oct 02 04:33:31 crc kubenswrapper[4775]: I1002 04:33:31.692492 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-85777745bb-bpg5g_c058610a-487c-4f23-b906-b03b0a458534/manager/0.log" Oct 02 04:33:31 crc kubenswrapper[4775]: I1002 04:33:31.845391 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-b8d54b5d7-vbs7k_c4e4f22c-1f43-4d7c-8dc4-b7dabc97425f/manager/0.log" Oct 02 04:33:31 crc kubenswrapper[4775]: I1002 04:33:31.980602 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-mlksw_7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62/manager/0.log" Oct 02 04:33:32 crc kubenswrapper[4775]: I1002 04:33:32.002806 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6b9957f54f-mlksw_7fe9e47b-ad8f-4ebf-ae98-6adaba8cae62/kube-rbac-proxy/0.log" Oct 02 04:33:32 crc kubenswrapper[4775]: I1002 04:33:32.009586 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-557f5d867b-2m767_f0c5961b-3c87-4438-b862-e2d1a83b1d6f/manager/0.log" Oct 02 04:33:37 crc kubenswrapper[4775]: I1002 04:33:37.233392 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:33:37 crc kubenswrapper[4775]: I1002 04:33:37.233870 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:33:37 crc kubenswrapper[4775]: I1002 04:33:37.233921 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 04:33:37 crc kubenswrapper[4775]: I1002 04:33:37.234838 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f16b9f11307f31ca81e98942662cc69d6baccc40acb87cced0a5c54c35b4a97c"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 04:33:37 crc kubenswrapper[4775]: I1002 04:33:37.234900 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://f16b9f11307f31ca81e98942662cc69d6baccc40acb87cced0a5c54c35b4a97c" gracePeriod=600 Oct 02 04:33:38 crc kubenswrapper[4775]: I1002 04:33:38.278802 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="f16b9f11307f31ca81e98942662cc69d6baccc40acb87cced0a5c54c35b4a97c" exitCode=0 Oct 02 04:33:38 crc kubenswrapper[4775]: I1002 04:33:38.278874 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"f16b9f11307f31ca81e98942662cc69d6baccc40acb87cced0a5c54c35b4a97c"} Oct 02 04:33:38 crc kubenswrapper[4775]: I1002 04:33:38.279460 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerStarted","Data":"55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e"} Oct 02 04:33:38 crc kubenswrapper[4775]: I1002 04:33:38.279487 4775 scope.go:117] "RemoveContainer" containerID="8e791e2ab58368c9c644e15792b6a68c4929f7e403e1b3c400b8910a0e1f182a" Oct 02 04:33:49 crc kubenswrapper[4775]: I1002 04:33:49.866611 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-wnfts_115575fa-2b4f-4b7e-bdd8-46c85da3855d/control-plane-machine-set-operator/0.log" Oct 02 04:33:49 crc kubenswrapper[4775]: I1002 04:33:49.965753 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5v4h8_958d5b03-aa2b-4068-b6fc-6efb125179a0/kube-rbac-proxy/0.log" Oct 02 04:33:50 crc kubenswrapper[4775]: I1002 04:33:50.044612 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5v4h8_958d5b03-aa2b-4068-b6fc-6efb125179a0/machine-api-operator/0.log" Oct 02 04:34:04 crc kubenswrapper[4775]: I1002 04:34:04.499314 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-7d4cc89fcb-6lsvs_8c968ad9-3556-49c9-859b-323223695d9e/cert-manager-controller/0.log" Oct 02 04:34:04 crc kubenswrapper[4775]: I1002 04:34:04.645845 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7d9f95dbf-ljzw5_0a404f98-18c9-437f-9da0-b3075699cc4a/cert-manager-cainjector/0.log" Oct 02 04:34:04 crc kubenswrapper[4775]: I1002 04:34:04.694292 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-d969966f-m456n_d130200c-29a8-45ca-8497-fec6226297a3/cert-manager-webhook/0.log" Oct 02 04:34:10 crc kubenswrapper[4775]: I1002 04:34:10.898902 4775 trace.go:236] Trace[2092880149]: "Calculate volume metrics of ovndbcluster-nb-etc-ovn for pod openstack/ovsdbserver-nb-0" (02-Oct-2025 04:34:09.882) (total time: 1016ms): Oct 02 04:34:10 crc kubenswrapper[4775]: Trace[2092880149]: [1.016191569s] [1.016191569s] END Oct 02 04:34:18 crc kubenswrapper[4775]: I1002 04:34:18.407920 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-q9jhv_336c609a-f0e8-42c1-8cd8-bafb09510123/nmstate-console-plugin/0.log" Oct 02 04:34:18 crc kubenswrapper[4775]: I1002 04:34:18.601007 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-qf7xb_69f99de2-12a0-4dbb-9d40-c44352e5859b/nmstate-handler/0.log" Oct 02 04:34:18 crc kubenswrapper[4775]: I1002 04:34:18.645696 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-xfm7h_9922a2fe-fb9f-4622-ba18-518eae27300f/kube-rbac-proxy/0.log" Oct 02 04:34:18 crc kubenswrapper[4775]: I1002 04:34:18.680561 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-xfm7h_9922a2fe-fb9f-4622-ba18-518eae27300f/nmstate-metrics/0.log" Oct 02 04:34:18 crc kubenswrapper[4775]: I1002 04:34:18.914867 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-g5pnj_b4c70ff6-dec5-4d9e-a7ac-b3f047e99c50/nmstate-webhook/0.log" Oct 02 04:34:18 crc kubenswrapper[4775]: I1002 04:34:18.945107 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-t84m4_bc7f4fa5-9d4f-4c61-9cd1-b91f57af5ca0/nmstate-operator/0.log" Oct 02 04:34:33 crc kubenswrapper[4775]: I1002 04:34:33.876415 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-42v88_8d6246a6-7cbf-4d65-9d84-ee9649d3fed6/kube-rbac-proxy/0.log" Oct 02 04:34:34 crc kubenswrapper[4775]: I1002 04:34:34.101442 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/cp-frr-files/0.log" Oct 02 04:34:34 crc kubenswrapper[4775]: I1002 04:34:34.309333 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/cp-reloader/0.log" Oct 02 04:34:34 crc kubenswrapper[4775]: I1002 04:34:34.309458 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/cp-frr-files/0.log" Oct 02 04:34:34 crc kubenswrapper[4775]: I1002 04:34:34.335225 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/cp-metrics/0.log" Oct 02 04:34:34 crc kubenswrapper[4775]: I1002 04:34:34.366982 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-42v88_8d6246a6-7cbf-4d65-9d84-ee9649d3fed6/controller/0.log" Oct 02 04:34:34 crc kubenswrapper[4775]: I1002 04:34:34.469820 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/cp-reloader/0.log" Oct 02 04:34:34 crc kubenswrapper[4775]: I1002 04:34:34.621297 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/cp-metrics/0.log" Oct 02 04:34:34 crc kubenswrapper[4775]: I1002 04:34:34.626133 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/cp-frr-files/0.log" Oct 02 04:34:34 crc kubenswrapper[4775]: I1002 04:34:34.646666 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/cp-reloader/0.log" Oct 02 04:34:34 crc kubenswrapper[4775]: I1002 04:34:34.682591 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/cp-metrics/0.log" Oct 02 04:34:34 crc kubenswrapper[4775]: I1002 04:34:34.844410 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/cp-frr-files/0.log" Oct 02 04:34:34 crc kubenswrapper[4775]: I1002 04:34:34.854240 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/cp-metrics/0.log" Oct 02 04:34:34 crc kubenswrapper[4775]: I1002 04:34:34.886318 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/cp-reloader/0.log" Oct 02 04:34:34 crc kubenswrapper[4775]: I1002 04:34:34.898162 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/controller/0.log" Oct 02 04:34:35 crc kubenswrapper[4775]: I1002 04:34:35.057319 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/frr-metrics/0.log" Oct 02 04:34:35 crc kubenswrapper[4775]: I1002 04:34:35.060517 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/kube-rbac-proxy/0.log" Oct 02 04:34:35 crc kubenswrapper[4775]: I1002 04:34:35.087922 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/kube-rbac-proxy-frr/0.log" Oct 02 04:34:35 crc kubenswrapper[4775]: I1002 04:34:35.308807 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/reloader/0.log" Oct 02 04:34:35 crc kubenswrapper[4775]: I1002 04:34:35.356129 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-7d29r_3a402158-a679-4d43-8e16-7b3ccfe91452/frr-k8s-webhook-server/0.log" Oct 02 04:34:35 crc kubenswrapper[4775]: I1002 04:34:35.538383 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-584458db67-zj7kv_8e0db7b3-7942-4d66-8ab1-5cfc51e5e355/manager/0.log" Oct 02 04:34:35 crc kubenswrapper[4775]: I1002 04:34:35.752290 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5c6cc6d6c-c2kfx_921014ce-ad65-4bb9-b673-0c2883ba98cd/webhook-server/0.log" Oct 02 04:34:35 crc kubenswrapper[4775]: I1002 04:34:35.812831 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tq2ld_a77c5c49-2e87-45ae-b2ab-5994102a0f38/kube-rbac-proxy/0.log" Oct 02 04:34:36 crc kubenswrapper[4775]: I1002 04:34:36.919717 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tq2ld_a77c5c49-2e87-45ae-b2ab-5994102a0f38/speaker/0.log" Oct 02 04:34:38 crc kubenswrapper[4775]: I1002 04:34:38.611665 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-g9pzp_15825320-b57d-4195-b410-65ca428465fa/frr/0.log" Oct 02 04:34:50 crc kubenswrapper[4775]: I1002 04:34:50.684632 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h_9802ba9a-2a80-44d8-b475-27bfa5044497/util/0.log" Oct 02 04:34:50 crc kubenswrapper[4775]: I1002 04:34:50.880464 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h_9802ba9a-2a80-44d8-b475-27bfa5044497/util/0.log" Oct 02 04:34:50 crc kubenswrapper[4775]: I1002 04:34:50.911182 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h_9802ba9a-2a80-44d8-b475-27bfa5044497/pull/0.log" Oct 02 04:34:50 crc kubenswrapper[4775]: I1002 04:34:50.952885 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h_9802ba9a-2a80-44d8-b475-27bfa5044497/pull/0.log" Oct 02 04:34:51 crc kubenswrapper[4775]: I1002 04:34:51.058243 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h_9802ba9a-2a80-44d8-b475-27bfa5044497/util/0.log" Oct 02 04:34:51 crc kubenswrapper[4775]: I1002 04:34:51.092037 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h_9802ba9a-2a80-44d8-b475-27bfa5044497/pull/0.log" Oct 02 04:34:51 crc kubenswrapper[4775]: I1002 04:34:51.137639 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_695e9552c02c72940c72621f824780f00ca58086c3badc308bf0a2eb69zxr5h_9802ba9a-2a80-44d8-b475-27bfa5044497/extract/0.log" Oct 02 04:34:51 crc kubenswrapper[4775]: I1002 04:34:51.222157 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l_bb58aabc-fee6-480d-8d28-d7ebf0506266/util/0.log" Oct 02 04:34:51 crc kubenswrapper[4775]: I1002 04:34:51.388900 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l_bb58aabc-fee6-480d-8d28-d7ebf0506266/util/0.log" Oct 02 04:34:51 crc kubenswrapper[4775]: I1002 04:34:51.423399 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l_bb58aabc-fee6-480d-8d28-d7ebf0506266/pull/0.log" Oct 02 04:34:51 crc kubenswrapper[4775]: I1002 04:34:51.437393 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l_bb58aabc-fee6-480d-8d28-d7ebf0506266/pull/0.log" Oct 02 04:34:51 crc kubenswrapper[4775]: I1002 04:34:51.582803 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l_bb58aabc-fee6-480d-8d28-d7ebf0506266/pull/0.log" Oct 02 04:34:51 crc kubenswrapper[4775]: I1002 04:34:51.589938 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l_bb58aabc-fee6-480d-8d28-d7ebf0506266/util/0.log" Oct 02 04:34:51 crc kubenswrapper[4775]: I1002 04:34:51.594178 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vp67l_bb58aabc-fee6-480d-8d28-d7ebf0506266/extract/0.log" Oct 02 04:34:51 crc kubenswrapper[4775]: I1002 04:34:51.772595 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm_4dbd62e7-a723-4c3b-945a-9014be98cb02/util/0.log" Oct 02 04:34:51 crc kubenswrapper[4775]: I1002 04:34:51.929131 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm_4dbd62e7-a723-4c3b-945a-9014be98cb02/pull/0.log" Oct 02 04:34:51 crc kubenswrapper[4775]: I1002 04:34:51.992811 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm_4dbd62e7-a723-4c3b-945a-9014be98cb02/util/0.log" Oct 02 04:34:51 crc kubenswrapper[4775]: I1002 04:34:51.993765 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm_4dbd62e7-a723-4c3b-945a-9014be98cb02/pull/0.log" Oct 02 04:34:52 crc kubenswrapper[4775]: I1002 04:34:52.107749 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm_4dbd62e7-a723-4c3b-945a-9014be98cb02/pull/0.log" Oct 02 04:34:52 crc kubenswrapper[4775]: I1002 04:34:52.111452 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm_4dbd62e7-a723-4c3b-945a-9014be98cb02/util/0.log" Oct 02 04:34:52 crc kubenswrapper[4775]: I1002 04:34:52.201030 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a6d815214afcb93f379916e45350d3de39072121f31a1d7eaaf6e22c2d2w7dm_4dbd62e7-a723-4c3b-945a-9014be98cb02/extract/0.log" Oct 02 04:34:52 crc kubenswrapper[4775]: I1002 04:34:52.277967 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ltrcf_a17388ea-bfd3-40b8-8b69-984e95eec731/extract-utilities/0.log" Oct 02 04:34:52 crc kubenswrapper[4775]: I1002 04:34:52.521378 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ltrcf_a17388ea-bfd3-40b8-8b69-984e95eec731/extract-content/0.log" Oct 02 04:34:52 crc kubenswrapper[4775]: I1002 04:34:52.526194 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ltrcf_a17388ea-bfd3-40b8-8b69-984e95eec731/extract-content/0.log" Oct 02 04:34:52 crc kubenswrapper[4775]: I1002 04:34:52.541407 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ltrcf_a17388ea-bfd3-40b8-8b69-984e95eec731/extract-utilities/0.log" Oct 02 04:34:52 crc kubenswrapper[4775]: I1002 04:34:52.704522 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ltrcf_a17388ea-bfd3-40b8-8b69-984e95eec731/extract-utilities/0.log" Oct 02 04:34:52 crc kubenswrapper[4775]: I1002 04:34:52.744428 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ltrcf_a17388ea-bfd3-40b8-8b69-984e95eec731/extract-content/0.log" Oct 02 04:34:52 crc kubenswrapper[4775]: I1002 04:34:52.935606 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-74hz2_02a708a6-f32e-45b7-8989-4983ee6ada6c/extract-utilities/0.log" Oct 02 04:34:53 crc kubenswrapper[4775]: I1002 04:34:53.191615 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-74hz2_02a708a6-f32e-45b7-8989-4983ee6ada6c/extract-content/0.log" Oct 02 04:34:53 crc kubenswrapper[4775]: I1002 04:34:53.202332 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-74hz2_02a708a6-f32e-45b7-8989-4983ee6ada6c/extract-utilities/0.log" Oct 02 04:34:53 crc kubenswrapper[4775]: I1002 04:34:53.205392 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-74hz2_02a708a6-f32e-45b7-8989-4983ee6ada6c/extract-content/0.log" Oct 02 04:34:53 crc kubenswrapper[4775]: I1002 04:34:53.398737 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-74hz2_02a708a6-f32e-45b7-8989-4983ee6ada6c/extract-content/0.log" Oct 02 04:34:53 crc kubenswrapper[4775]: I1002 04:34:53.469702 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-74hz2_02a708a6-f32e-45b7-8989-4983ee6ada6c/extract-utilities/0.log" Oct 02 04:34:53 crc kubenswrapper[4775]: I1002 04:34:53.701440 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp_7e933c35-a14e-4439-ae32-20f95df5eb7b/util/0.log" Oct 02 04:34:53 crc kubenswrapper[4775]: I1002 04:34:53.893736 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp_7e933c35-a14e-4439-ae32-20f95df5eb7b/util/0.log" Oct 02 04:34:53 crc kubenswrapper[4775]: I1002 04:34:53.903403 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp_7e933c35-a14e-4439-ae32-20f95df5eb7b/pull/0.log" Oct 02 04:34:54 crc kubenswrapper[4775]: I1002 04:34:54.032671 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp_7e933c35-a14e-4439-ae32-20f95df5eb7b/pull/0.log" Oct 02 04:34:54 crc kubenswrapper[4775]: I1002 04:34:54.092912 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ltrcf_a17388ea-bfd3-40b8-8b69-984e95eec731/registry-server/0.log" Oct 02 04:34:54 crc kubenswrapper[4775]: I1002 04:34:54.781617 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp_7e933c35-a14e-4439-ae32-20f95df5eb7b/util/0.log" Oct 02 04:34:54 crc kubenswrapper[4775]: I1002 04:34:54.987250 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp_7e933c35-a14e-4439-ae32-20f95df5eb7b/extract/0.log" Oct 02 04:34:55 crc kubenswrapper[4775]: I1002 04:34:55.012781 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cqpzcp_7e933c35-a14e-4439-ae32-20f95df5eb7b/pull/0.log" Oct 02 04:34:55 crc kubenswrapper[4775]: I1002 04:34:55.055814 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-n4prl_6f9799cb-9770-4dd2-87a3-55604cad82d2/marketplace-operator/0.log" Oct 02 04:34:55 crc kubenswrapper[4775]: I1002 04:34:55.262348 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-74hz2_02a708a6-f32e-45b7-8989-4983ee6ada6c/registry-server/0.log" Oct 02 04:34:55 crc kubenswrapper[4775]: I1002 04:34:55.286400 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mpj5m_83400c9e-1f98-434e-a937-9acb45d8a61b/extract-utilities/0.log" Oct 02 04:34:55 crc kubenswrapper[4775]: I1002 04:34:55.379994 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mpj5m_83400c9e-1f98-434e-a937-9acb45d8a61b/extract-utilities/0.log" Oct 02 04:34:55 crc kubenswrapper[4775]: I1002 04:34:55.423752 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mpj5m_83400c9e-1f98-434e-a937-9acb45d8a61b/extract-content/0.log" Oct 02 04:34:55 crc kubenswrapper[4775]: I1002 04:34:55.423924 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mpj5m_83400c9e-1f98-434e-a937-9acb45d8a61b/extract-content/0.log" Oct 02 04:34:55 crc kubenswrapper[4775]: I1002 04:34:55.564824 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mpj5m_83400c9e-1f98-434e-a937-9acb45d8a61b/extract-content/0.log" Oct 02 04:34:55 crc kubenswrapper[4775]: I1002 04:34:55.618211 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mpj5m_83400c9e-1f98-434e-a937-9acb45d8a61b/extract-utilities/0.log" Oct 02 04:34:55 crc kubenswrapper[4775]: I1002 04:34:55.651348 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sdnvn_763784f6-4d43-4db9-88cb-0ba8f213f78e/extract-utilities/0.log" Oct 02 04:34:55 crc kubenswrapper[4775]: I1002 04:34:55.888244 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sdnvn_763784f6-4d43-4db9-88cb-0ba8f213f78e/extract-utilities/0.log" Oct 02 04:34:55 crc kubenswrapper[4775]: I1002 04:34:55.917583 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sdnvn_763784f6-4d43-4db9-88cb-0ba8f213f78e/extract-content/0.log" Oct 02 04:34:55 crc kubenswrapper[4775]: I1002 04:34:55.946533 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sdnvn_763784f6-4d43-4db9-88cb-0ba8f213f78e/extract-content/0.log" Oct 02 04:34:55 crc kubenswrapper[4775]: I1002 04:34:55.949031 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-mpj5m_83400c9e-1f98-434e-a937-9acb45d8a61b/registry-server/0.log" Oct 02 04:34:56 crc kubenswrapper[4775]: I1002 04:34:56.070270 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sdnvn_763784f6-4d43-4db9-88cb-0ba8f213f78e/extract-content/0.log" Oct 02 04:34:56 crc kubenswrapper[4775]: I1002 04:34:56.105731 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sdnvn_763784f6-4d43-4db9-88cb-0ba8f213f78e/extract-utilities/0.log" Oct 02 04:34:57 crc kubenswrapper[4775]: I1002 04:34:57.391619 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-sdnvn_763784f6-4d43-4db9-88cb-0ba8f213f78e/registry-server/0.log" Oct 02 04:35:09 crc kubenswrapper[4775]: I1002 04:35:09.971703 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-7c8cf85677-twdtn_09b1503e-76b6-40c6-9494-b69347e2ba6b/prometheus-operator/0.log" Oct 02 04:35:10 crc kubenswrapper[4775]: I1002 04:35:10.201674 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-674cbc6567-qwjzq_b6c7d2d9-eef5-4f3a-b195-81cd70fd6c07/prometheus-operator-admission-webhook/0.log" Oct 02 04:35:10 crc kubenswrapper[4775]: I1002 04:35:10.232626 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-674cbc6567-xwfwt_902f36d9-43f8-451f-b83a-382bcc50e46c/prometheus-operator-admission-webhook/0.log" Oct 02 04:35:10 crc kubenswrapper[4775]: I1002 04:35:10.362775 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-cc5f78dfc-2gm7h_6705f52c-cdc8-4b21-a92d-88eadcdf8754/operator/0.log" Oct 02 04:35:10 crc kubenswrapper[4775]: I1002 04:35:10.409217 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-54bc95c9fb-lzh9v_4dac8c52-da57-4264-a2c1-43641d9a4429/perses-operator/0.log" Oct 02 04:35:34 crc kubenswrapper[4775]: E1002 04:35:34.601493 4775 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.47:52092->38.102.83.47:45167: write tcp 38.102.83.47:52092->38.102.83.47:45167: write: broken pipe Oct 02 04:35:37 crc kubenswrapper[4775]: I1002 04:35:37.233638 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:35:37 crc kubenswrapper[4775]: I1002 04:35:37.234278 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:35:57 crc kubenswrapper[4775]: I1002 04:35:57.864986 4775 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-294lk"] Oct 02 04:35:57 crc kubenswrapper[4775]: E1002 04:35:57.866153 4775 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3619ed82-06f8-42e6-b7c6-7b3ab832580b" containerName="container-00" Oct 02 04:35:57 crc kubenswrapper[4775]: I1002 04:35:57.866170 4775 state_mem.go:107] "Deleted CPUSet assignment" podUID="3619ed82-06f8-42e6-b7c6-7b3ab832580b" containerName="container-00" Oct 02 04:35:57 crc kubenswrapper[4775]: I1002 04:35:57.866465 4775 memory_manager.go:354] "RemoveStaleState removing state" podUID="3619ed82-06f8-42e6-b7c6-7b3ab832580b" containerName="container-00" Oct 02 04:35:57 crc kubenswrapper[4775]: I1002 04:35:57.868554 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:35:57 crc kubenswrapper[4775]: I1002 04:35:57.878931 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-294lk"] Oct 02 04:35:57 crc kubenswrapper[4775]: I1002 04:35:57.880897 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-catalog-content\") pod \"redhat-marketplace-294lk\" (UID: \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\") " pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:35:57 crc kubenswrapper[4775]: I1002 04:35:57.881033 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z96p7\" (UniqueName: \"kubernetes.io/projected/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-kube-api-access-z96p7\") pod \"redhat-marketplace-294lk\" (UID: \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\") " pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:35:57 crc kubenswrapper[4775]: I1002 04:35:57.881119 4775 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-utilities\") pod \"redhat-marketplace-294lk\" (UID: \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\") " pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:35:57 crc kubenswrapper[4775]: I1002 04:35:57.984793 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-catalog-content\") pod \"redhat-marketplace-294lk\" (UID: \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\") " pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:35:57 crc kubenswrapper[4775]: I1002 04:35:57.984922 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z96p7\" (UniqueName: \"kubernetes.io/projected/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-kube-api-access-z96p7\") pod \"redhat-marketplace-294lk\" (UID: \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\") " pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:35:57 crc kubenswrapper[4775]: I1002 04:35:57.985464 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-catalog-content\") pod \"redhat-marketplace-294lk\" (UID: \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\") " pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:35:57 crc kubenswrapper[4775]: I1002 04:35:57.985585 4775 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-utilities\") pod \"redhat-marketplace-294lk\" (UID: \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\") " pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:35:57 crc kubenswrapper[4775]: I1002 04:35:57.985885 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-utilities\") pod \"redhat-marketplace-294lk\" (UID: \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\") " pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:35:58 crc kubenswrapper[4775]: I1002 04:35:58.015080 4775 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z96p7\" (UniqueName: \"kubernetes.io/projected/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-kube-api-access-z96p7\") pod \"redhat-marketplace-294lk\" (UID: \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\") " pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:35:58 crc kubenswrapper[4775]: I1002 04:35:58.191462 4775 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:35:58 crc kubenswrapper[4775]: I1002 04:35:58.703926 4775 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-294lk"] Oct 02 04:35:58 crc kubenswrapper[4775]: W1002 04:35:58.712945 4775 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda986aeb9_c8c4_4fd3_ac4c_985164947bc4.slice/crio-07f9f5be1a09e5b51a668d67b994ef3210c2a656a1913306808a349cc1ef0b26 WatchSource:0}: Error finding container 07f9f5be1a09e5b51a668d67b994ef3210c2a656a1913306808a349cc1ef0b26: Status 404 returned error can't find the container with id 07f9f5be1a09e5b51a668d67b994ef3210c2a656a1913306808a349cc1ef0b26 Oct 02 04:35:59 crc kubenswrapper[4775]: I1002 04:35:59.011076 4775 generic.go:334] "Generic (PLEG): container finished" podID="a986aeb9-c8c4-4fd3-ac4c-985164947bc4" containerID="3884d770b30daaf17341608e7ba7cd34150c1badde759f4f3296902311463533" exitCode=0 Oct 02 04:35:59 crc kubenswrapper[4775]: I1002 04:35:59.011561 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-294lk" event={"ID":"a986aeb9-c8c4-4fd3-ac4c-985164947bc4","Type":"ContainerDied","Data":"3884d770b30daaf17341608e7ba7cd34150c1badde759f4f3296902311463533"} Oct 02 04:35:59 crc kubenswrapper[4775]: I1002 04:35:59.011601 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-294lk" event={"ID":"a986aeb9-c8c4-4fd3-ac4c-985164947bc4","Type":"ContainerStarted","Data":"07f9f5be1a09e5b51a668d67b994ef3210c2a656a1913306808a349cc1ef0b26"} Oct 02 04:35:59 crc kubenswrapper[4775]: I1002 04:35:59.015890 4775 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 02 04:36:00 crc kubenswrapper[4775]: I1002 04:36:00.026628 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-294lk" event={"ID":"a986aeb9-c8c4-4fd3-ac4c-985164947bc4","Type":"ContainerStarted","Data":"37e90cdca83d7903ad0b6d484153b55af08474d072d221c61cdec321b04a1266"} Oct 02 04:36:01 crc kubenswrapper[4775]: I1002 04:36:01.039197 4775 generic.go:334] "Generic (PLEG): container finished" podID="a986aeb9-c8c4-4fd3-ac4c-985164947bc4" containerID="37e90cdca83d7903ad0b6d484153b55af08474d072d221c61cdec321b04a1266" exitCode=0 Oct 02 04:36:01 crc kubenswrapper[4775]: I1002 04:36:01.039244 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-294lk" event={"ID":"a986aeb9-c8c4-4fd3-ac4c-985164947bc4","Type":"ContainerDied","Data":"37e90cdca83d7903ad0b6d484153b55af08474d072d221c61cdec321b04a1266"} Oct 02 04:36:02 crc kubenswrapper[4775]: I1002 04:36:02.056377 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-294lk" event={"ID":"a986aeb9-c8c4-4fd3-ac4c-985164947bc4","Type":"ContainerStarted","Data":"f1941778ca37b02e52062b2996d661f73242a5c2a01e371c382d2238475ac7fb"} Oct 02 04:36:02 crc kubenswrapper[4775]: I1002 04:36:02.076520 4775 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-294lk" podStartSLOduration=2.477538699 podStartE2EDuration="5.076501162s" podCreationTimestamp="2025-10-02 04:35:57 +0000 UTC" firstStartedPulling="2025-10-02 04:35:59.015516991 +0000 UTC m=+10496.182261031" lastFinishedPulling="2025-10-02 04:36:01.614479454 +0000 UTC m=+10498.781223494" observedRunningTime="2025-10-02 04:36:02.071561793 +0000 UTC m=+10499.238305864" watchObservedRunningTime="2025-10-02 04:36:02.076501162 +0000 UTC m=+10499.243245202" Oct 02 04:36:07 crc kubenswrapper[4775]: I1002 04:36:07.234048 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:36:07 crc kubenswrapper[4775]: I1002 04:36:07.234691 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:36:08 crc kubenswrapper[4775]: I1002 04:36:08.191580 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:36:08 crc kubenswrapper[4775]: I1002 04:36:08.192379 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:36:08 crc kubenswrapper[4775]: I1002 04:36:08.281257 4775 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:36:09 crc kubenswrapper[4775]: I1002 04:36:09.259423 4775 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:36:09 crc kubenswrapper[4775]: I1002 04:36:09.333705 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-294lk"] Oct 02 04:36:11 crc kubenswrapper[4775]: I1002 04:36:11.219809 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-294lk" podUID="a986aeb9-c8c4-4fd3-ac4c-985164947bc4" containerName="registry-server" containerID="cri-o://f1941778ca37b02e52062b2996d661f73242a5c2a01e371c382d2238475ac7fb" gracePeriod=2 Oct 02 04:36:11 crc kubenswrapper[4775]: I1002 04:36:11.782532 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:36:11 crc kubenswrapper[4775]: I1002 04:36:11.871723 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-utilities\") pod \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\" (UID: \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\") " Oct 02 04:36:11 crc kubenswrapper[4775]: I1002 04:36:11.872059 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z96p7\" (UniqueName: \"kubernetes.io/projected/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-kube-api-access-z96p7\") pod \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\" (UID: \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\") " Oct 02 04:36:11 crc kubenswrapper[4775]: I1002 04:36:11.872268 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-catalog-content\") pod \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\" (UID: \"a986aeb9-c8c4-4fd3-ac4c-985164947bc4\") " Oct 02 04:36:11 crc kubenswrapper[4775]: I1002 04:36:11.874263 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-utilities" (OuterVolumeSpecName: "utilities") pod "a986aeb9-c8c4-4fd3-ac4c-985164947bc4" (UID: "a986aeb9-c8c4-4fd3-ac4c-985164947bc4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:36:11 crc kubenswrapper[4775]: I1002 04:36:11.882373 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-kube-api-access-z96p7" (OuterVolumeSpecName: "kube-api-access-z96p7") pod "a986aeb9-c8c4-4fd3-ac4c-985164947bc4" (UID: "a986aeb9-c8c4-4fd3-ac4c-985164947bc4"). InnerVolumeSpecName "kube-api-access-z96p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:36:11 crc kubenswrapper[4775]: I1002 04:36:11.903370 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a986aeb9-c8c4-4fd3-ac4c-985164947bc4" (UID: "a986aeb9-c8c4-4fd3-ac4c-985164947bc4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:36:11 crc kubenswrapper[4775]: I1002 04:36:11.975126 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z96p7\" (UniqueName: \"kubernetes.io/projected/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-kube-api-access-z96p7\") on node \"crc\" DevicePath \"\"" Oct 02 04:36:11 crc kubenswrapper[4775]: I1002 04:36:11.975284 4775 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 02 04:36:11 crc kubenswrapper[4775]: I1002 04:36:11.975351 4775 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a986aeb9-c8c4-4fd3-ac4c-985164947bc4-utilities\") on node \"crc\" DevicePath \"\"" Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.232375 4775 generic.go:334] "Generic (PLEG): container finished" podID="a986aeb9-c8c4-4fd3-ac4c-985164947bc4" containerID="f1941778ca37b02e52062b2996d661f73242a5c2a01e371c382d2238475ac7fb" exitCode=0 Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.232420 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-294lk" event={"ID":"a986aeb9-c8c4-4fd3-ac4c-985164947bc4","Type":"ContainerDied","Data":"f1941778ca37b02e52062b2996d661f73242a5c2a01e371c382d2238475ac7fb"} Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.232447 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-294lk" event={"ID":"a986aeb9-c8c4-4fd3-ac4c-985164947bc4","Type":"ContainerDied","Data":"07f9f5be1a09e5b51a668d67b994ef3210c2a656a1913306808a349cc1ef0b26"} Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.232468 4775 scope.go:117] "RemoveContainer" containerID="f1941778ca37b02e52062b2996d661f73242a5c2a01e371c382d2238475ac7fb" Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.234436 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-294lk" Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.264285 4775 scope.go:117] "RemoveContainer" containerID="37e90cdca83d7903ad0b6d484153b55af08474d072d221c61cdec321b04a1266" Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.294510 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-294lk"] Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.304899 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-294lk"] Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.318230 4775 scope.go:117] "RemoveContainer" containerID="3884d770b30daaf17341608e7ba7cd34150c1badde759f4f3296902311463533" Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.385141 4775 scope.go:117] "RemoveContainer" containerID="f1941778ca37b02e52062b2996d661f73242a5c2a01e371c382d2238475ac7fb" Oct 02 04:36:12 crc kubenswrapper[4775]: E1002 04:36:12.385809 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1941778ca37b02e52062b2996d661f73242a5c2a01e371c382d2238475ac7fb\": container with ID starting with f1941778ca37b02e52062b2996d661f73242a5c2a01e371c382d2238475ac7fb not found: ID does not exist" containerID="f1941778ca37b02e52062b2996d661f73242a5c2a01e371c382d2238475ac7fb" Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.385875 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1941778ca37b02e52062b2996d661f73242a5c2a01e371c382d2238475ac7fb"} err="failed to get container status \"f1941778ca37b02e52062b2996d661f73242a5c2a01e371c382d2238475ac7fb\": rpc error: code = NotFound desc = could not find container \"f1941778ca37b02e52062b2996d661f73242a5c2a01e371c382d2238475ac7fb\": container with ID starting with f1941778ca37b02e52062b2996d661f73242a5c2a01e371c382d2238475ac7fb not found: ID does not exist" Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.385916 4775 scope.go:117] "RemoveContainer" containerID="37e90cdca83d7903ad0b6d484153b55af08474d072d221c61cdec321b04a1266" Oct 02 04:36:12 crc kubenswrapper[4775]: E1002 04:36:12.386486 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37e90cdca83d7903ad0b6d484153b55af08474d072d221c61cdec321b04a1266\": container with ID starting with 37e90cdca83d7903ad0b6d484153b55af08474d072d221c61cdec321b04a1266 not found: ID does not exist" containerID="37e90cdca83d7903ad0b6d484153b55af08474d072d221c61cdec321b04a1266" Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.386650 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37e90cdca83d7903ad0b6d484153b55af08474d072d221c61cdec321b04a1266"} err="failed to get container status \"37e90cdca83d7903ad0b6d484153b55af08474d072d221c61cdec321b04a1266\": rpc error: code = NotFound desc = could not find container \"37e90cdca83d7903ad0b6d484153b55af08474d072d221c61cdec321b04a1266\": container with ID starting with 37e90cdca83d7903ad0b6d484153b55af08474d072d221c61cdec321b04a1266 not found: ID does not exist" Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.386804 4775 scope.go:117] "RemoveContainer" containerID="3884d770b30daaf17341608e7ba7cd34150c1badde759f4f3296902311463533" Oct 02 04:36:12 crc kubenswrapper[4775]: E1002 04:36:12.387429 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3884d770b30daaf17341608e7ba7cd34150c1badde759f4f3296902311463533\": container with ID starting with 3884d770b30daaf17341608e7ba7cd34150c1badde759f4f3296902311463533 not found: ID does not exist" containerID="3884d770b30daaf17341608e7ba7cd34150c1badde759f4f3296902311463533" Oct 02 04:36:12 crc kubenswrapper[4775]: I1002 04:36:12.387481 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3884d770b30daaf17341608e7ba7cd34150c1badde759f4f3296902311463533"} err="failed to get container status \"3884d770b30daaf17341608e7ba7cd34150c1badde759f4f3296902311463533\": rpc error: code = NotFound desc = could not find container \"3884d770b30daaf17341608e7ba7cd34150c1badde759f4f3296902311463533\": container with ID starting with 3884d770b30daaf17341608e7ba7cd34150c1badde759f4f3296902311463533 not found: ID does not exist" Oct 02 04:36:13 crc kubenswrapper[4775]: I1002 04:36:13.797083 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a986aeb9-c8c4-4fd3-ac4c-985164947bc4" path="/var/lib/kubelet/pods/a986aeb9-c8c4-4fd3-ac4c-985164947bc4/volumes" Oct 02 04:36:37 crc kubenswrapper[4775]: I1002 04:36:37.233470 4775 patch_prober.go:28] interesting pod/machine-config-daemon-945lh container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 02 04:36:37 crc kubenswrapper[4775]: I1002 04:36:37.234269 4775 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 02 04:36:37 crc kubenswrapper[4775]: I1002 04:36:37.234328 4775 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-945lh" Oct 02 04:36:37 crc kubenswrapper[4775]: I1002 04:36:37.235220 4775 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e"} pod="openshift-machine-config-operator/machine-config-daemon-945lh" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 02 04:36:37 crc kubenswrapper[4775]: I1002 04:36:37.235309 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerName="machine-config-daemon" containerID="cri-o://55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" gracePeriod=600 Oct 02 04:36:37 crc kubenswrapper[4775]: E1002 04:36:37.368287 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:36:37 crc kubenswrapper[4775]: I1002 04:36:37.624363 4775 generic.go:334] "Generic (PLEG): container finished" podID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" exitCode=0 Oct 02 04:36:37 crc kubenswrapper[4775]: I1002 04:36:37.624616 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-945lh" event={"ID":"c390c5b3-f65b-4f2e-9d03-bb09cf613134","Type":"ContainerDied","Data":"55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e"} Oct 02 04:36:37 crc kubenswrapper[4775]: I1002 04:36:37.624833 4775 scope.go:117] "RemoveContainer" containerID="f16b9f11307f31ca81e98942662cc69d6baccc40acb87cced0a5c54c35b4a97c" Oct 02 04:36:37 crc kubenswrapper[4775]: I1002 04:36:37.625598 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:36:37 crc kubenswrapper[4775]: E1002 04:36:37.626111 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:36:49 crc kubenswrapper[4775]: I1002 04:36:49.778549 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:36:49 crc kubenswrapper[4775]: E1002 04:36:49.780092 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:37:04 crc kubenswrapper[4775]: I1002 04:37:04.765789 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:37:04 crc kubenswrapper[4775]: E1002 04:37:04.766676 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:37:18 crc kubenswrapper[4775]: I1002 04:37:18.766564 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:37:18 crc kubenswrapper[4775]: E1002 04:37:18.767830 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:37:33 crc kubenswrapper[4775]: I1002 04:37:33.784001 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:37:33 crc kubenswrapper[4775]: E1002 04:37:33.787817 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:37:45 crc kubenswrapper[4775]: I1002 04:37:45.546558 4775 generic.go:334] "Generic (PLEG): container finished" podID="1fd65775-68aa-4633-a43e-803f6b3ce9a2" containerID="0878ae6e198eed664c32026ddeb8a8c0b26826d38c05a377ae2bb85df0ff5182" exitCode=0 Oct 02 04:37:45 crc kubenswrapper[4775]: I1002 04:37:45.546655 4775 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-tbjl6/must-gather-7px42" event={"ID":"1fd65775-68aa-4633-a43e-803f6b3ce9a2","Type":"ContainerDied","Data":"0878ae6e198eed664c32026ddeb8a8c0b26826d38c05a377ae2bb85df0ff5182"} Oct 02 04:37:45 crc kubenswrapper[4775]: I1002 04:37:45.547869 4775 scope.go:117] "RemoveContainer" containerID="0878ae6e198eed664c32026ddeb8a8c0b26826d38c05a377ae2bb85df0ff5182" Oct 02 04:37:46 crc kubenswrapper[4775]: I1002 04:37:46.335817 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tbjl6_must-gather-7px42_1fd65775-68aa-4633-a43e-803f6b3ce9a2/gather/0.log" Oct 02 04:37:47 crc kubenswrapper[4775]: I1002 04:37:47.767002 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:37:47 crc kubenswrapper[4775]: E1002 04:37:47.768191 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.067109 4775 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-tbjl6/must-gather-7px42"] Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.067853 4775 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-tbjl6/must-gather-7px42" podUID="1fd65775-68aa-4633-a43e-803f6b3ce9a2" containerName="copy" containerID="cri-o://9cef27e990f26117f18d7d535484f057592574392b3d9425a31d9314fb5fb18f" gracePeriod=2 Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.095851 4775 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-tbjl6/must-gather-7px42"] Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.530249 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tbjl6_must-gather-7px42_1fd65775-68aa-4633-a43e-803f6b3ce9a2/copy/0.log" Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.531056 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/must-gather-7px42" Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.654665 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1fd65775-68aa-4633-a43e-803f6b3ce9a2-must-gather-output\") pod \"1fd65775-68aa-4633-a43e-803f6b3ce9a2\" (UID: \"1fd65775-68aa-4633-a43e-803f6b3ce9a2\") " Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.655154 4775 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c9sk\" (UniqueName: \"kubernetes.io/projected/1fd65775-68aa-4633-a43e-803f6b3ce9a2-kube-api-access-9c9sk\") pod \"1fd65775-68aa-4633-a43e-803f6b3ce9a2\" (UID: \"1fd65775-68aa-4633-a43e-803f6b3ce9a2\") " Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.664048 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fd65775-68aa-4633-a43e-803f6b3ce9a2-kube-api-access-9c9sk" (OuterVolumeSpecName: "kube-api-access-9c9sk") pod "1fd65775-68aa-4633-a43e-803f6b3ce9a2" (UID: "1fd65775-68aa-4633-a43e-803f6b3ce9a2"). InnerVolumeSpecName "kube-api-access-9c9sk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.683426 4775 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-tbjl6_must-gather-7px42_1fd65775-68aa-4633-a43e-803f6b3ce9a2/copy/0.log" Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.684001 4775 generic.go:334] "Generic (PLEG): container finished" podID="1fd65775-68aa-4633-a43e-803f6b3ce9a2" containerID="9cef27e990f26117f18d7d535484f057592574392b3d9425a31d9314fb5fb18f" exitCode=143 Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.684057 4775 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-tbjl6/must-gather-7px42" Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.684071 4775 scope.go:117] "RemoveContainer" containerID="9cef27e990f26117f18d7d535484f057592574392b3d9425a31d9314fb5fb18f" Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.745484 4775 scope.go:117] "RemoveContainer" containerID="0878ae6e198eed664c32026ddeb8a8c0b26826d38c05a377ae2bb85df0ff5182" Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.759486 4775 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c9sk\" (UniqueName: \"kubernetes.io/projected/1fd65775-68aa-4633-a43e-803f6b3ce9a2-kube-api-access-9c9sk\") on node \"crc\" DevicePath \"\"" Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.839149 4775 scope.go:117] "RemoveContainer" containerID="9cef27e990f26117f18d7d535484f057592574392b3d9425a31d9314fb5fb18f" Oct 02 04:37:56 crc kubenswrapper[4775]: E1002 04:37:56.839578 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cef27e990f26117f18d7d535484f057592574392b3d9425a31d9314fb5fb18f\": container with ID starting with 9cef27e990f26117f18d7d535484f057592574392b3d9425a31d9314fb5fb18f not found: ID does not exist" containerID="9cef27e990f26117f18d7d535484f057592574392b3d9425a31d9314fb5fb18f" Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.839614 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cef27e990f26117f18d7d535484f057592574392b3d9425a31d9314fb5fb18f"} err="failed to get container status \"9cef27e990f26117f18d7d535484f057592574392b3d9425a31d9314fb5fb18f\": rpc error: code = NotFound desc = could not find container \"9cef27e990f26117f18d7d535484f057592574392b3d9425a31d9314fb5fb18f\": container with ID starting with 9cef27e990f26117f18d7d535484f057592574392b3d9425a31d9314fb5fb18f not found: ID does not exist" Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.839641 4775 scope.go:117] "RemoveContainer" containerID="0878ae6e198eed664c32026ddeb8a8c0b26826d38c05a377ae2bb85df0ff5182" Oct 02 04:37:56 crc kubenswrapper[4775]: E1002 04:37:56.839893 4775 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0878ae6e198eed664c32026ddeb8a8c0b26826d38c05a377ae2bb85df0ff5182\": container with ID starting with 0878ae6e198eed664c32026ddeb8a8c0b26826d38c05a377ae2bb85df0ff5182 not found: ID does not exist" containerID="0878ae6e198eed664c32026ddeb8a8c0b26826d38c05a377ae2bb85df0ff5182" Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.839930 4775 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0878ae6e198eed664c32026ddeb8a8c0b26826d38c05a377ae2bb85df0ff5182"} err="failed to get container status \"0878ae6e198eed664c32026ddeb8a8c0b26826d38c05a377ae2bb85df0ff5182\": rpc error: code = NotFound desc = could not find container \"0878ae6e198eed664c32026ddeb8a8c0b26826d38c05a377ae2bb85df0ff5182\": container with ID starting with 0878ae6e198eed664c32026ddeb8a8c0b26826d38c05a377ae2bb85df0ff5182 not found: ID does not exist" Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.898385 4775 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1fd65775-68aa-4633-a43e-803f6b3ce9a2-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "1fd65775-68aa-4633-a43e-803f6b3ce9a2" (UID: "1fd65775-68aa-4633-a43e-803f6b3ce9a2"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 02 04:37:56 crc kubenswrapper[4775]: I1002 04:37:56.962931 4775 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/1fd65775-68aa-4633-a43e-803f6b3ce9a2-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 02 04:37:57 crc kubenswrapper[4775]: I1002 04:37:57.779007 4775 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fd65775-68aa-4633-a43e-803f6b3ce9a2" path="/var/lib/kubelet/pods/1fd65775-68aa-4633-a43e-803f6b3ce9a2/volumes" Oct 02 04:38:01 crc kubenswrapper[4775]: I1002 04:38:01.765171 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:38:01 crc kubenswrapper[4775]: E1002 04:38:01.765911 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:38:12 crc kubenswrapper[4775]: I1002 04:38:12.766402 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:38:12 crc kubenswrapper[4775]: E1002 04:38:12.769884 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:38:23 crc kubenswrapper[4775]: I1002 04:38:23.780080 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:38:23 crc kubenswrapper[4775]: E1002 04:38:23.781464 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:38:37 crc kubenswrapper[4775]: I1002 04:38:37.767228 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:38:37 crc kubenswrapper[4775]: E1002 04:38:37.768359 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:38:48 crc kubenswrapper[4775]: I1002 04:38:48.765835 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:38:48 crc kubenswrapper[4775]: E1002 04:38:48.766878 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:39:01 crc kubenswrapper[4775]: I1002 04:39:01.765695 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:39:01 crc kubenswrapper[4775]: E1002 04:39:01.766888 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:39:12 crc kubenswrapper[4775]: I1002 04:39:12.766328 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:39:12 crc kubenswrapper[4775]: E1002 04:39:12.767266 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:39:19 crc kubenswrapper[4775]: I1002 04:39:19.807814 4775 scope.go:117] "RemoveContainer" containerID="41a876660cb21d50b93e7ffd5d4cffc46f1f6e23e7605909f55164ab2ac0d2b4" Oct 02 04:39:27 crc kubenswrapper[4775]: I1002 04:39:27.765333 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:39:27 crc kubenswrapper[4775]: E1002 04:39:27.766268 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" Oct 02 04:39:39 crc kubenswrapper[4775]: I1002 04:39:39.765196 4775 scope.go:117] "RemoveContainer" containerID="55dadfe1a37dba1f3c3326171cb0c7ad387fb2adbdf571288dedcc5e663aac4e" Oct 02 04:39:39 crc kubenswrapper[4775]: E1002 04:39:39.765997 4775 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-945lh_openshift-machine-config-operator(c390c5b3-f65b-4f2e-9d03-bb09cf613134)\"" pod="openshift-machine-config-operator/machine-config-daemon-945lh" podUID="c390c5b3-f65b-4f2e-9d03-bb09cf613134" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515067401034024445 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015067401035017363 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015067353517016521 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015067353520015463 5ustar corecore